xref: /aosp_15_r20/prebuilts/clang-tools/linux-x86/clang-headers/arm_neon.h (revision bed243d3d9cd544cfb038bfa7be843dedc6e6bf7)
1 /*===---- arm_neon.h - ARM Neon intrinsics ---------------------------------===
2  *
3  * Permission is hereby granted, free of charge, to any person obtaining a copy
4  * of this software and associated documentation files (the "Software"), to deal
5  * in the Software without restriction, including without limitation the rights
6  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
7  * copies of the Software, and to permit persons to whom the Software is
8  * furnished to do so, subject to the following conditions:
9  *
10  * The above copyright notice and this permission notice shall be included in
11  * all copies or substantial portions of the Software.
12  *
13  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
18  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
19  * THE SOFTWARE.
20  *
21  *===-----------------------------------------------------------------------===
22  */
23 
24 #ifndef __ARM_NEON_H
25 #define __ARM_NEON_H
26 
27 #ifndef __ARM_FP
28 #error "NEON intrinsics not available with the soft-float ABI. Please use -mfloat-abi=softfp or -mfloat-abi=hard"
29 #else
30 
31 #if !defined(__ARM_NEON)
32 #error "NEON support not enabled"
33 #else
34 
35 #include <stdint.h>
36 
37 #include <arm_bf16.h>
38 #include <arm_vector_types.h>
39 #ifdef __aarch64__
40 typedef uint8_t poly8_t;
41 typedef uint16_t poly16_t;
42 typedef uint64_t poly64_t;
43 typedef __uint128_t poly128_t;
44 #else
45 typedef int8_t poly8_t;
46 typedef int16_t poly16_t;
47 typedef int64_t poly64_t;
48 #endif
49 typedef __attribute__((neon_polyvector_type(8))) poly8_t poly8x8_t;
50 typedef __attribute__((neon_polyvector_type(16))) poly8_t poly8x16_t;
51 typedef __attribute__((neon_polyvector_type(4))) poly16_t poly16x4_t;
52 typedef __attribute__((neon_polyvector_type(8))) poly16_t poly16x8_t;
53 typedef __attribute__((neon_polyvector_type(1))) poly64_t poly64x1_t;
54 typedef __attribute__((neon_polyvector_type(2))) poly64_t poly64x2_t;
55 
56 typedef struct poly8x8x2_t {
57   poly8x8_t val[2];
58 } poly8x8x2_t;
59 
60 typedef struct poly8x16x2_t {
61   poly8x16_t val[2];
62 } poly8x16x2_t;
63 
64 typedef struct poly16x4x2_t {
65   poly16x4_t val[2];
66 } poly16x4x2_t;
67 
68 typedef struct poly16x8x2_t {
69   poly16x8_t val[2];
70 } poly16x8x2_t;
71 
72 typedef struct poly64x1x2_t {
73   poly64x1_t val[2];
74 } poly64x1x2_t;
75 
76 typedef struct poly64x2x2_t {
77   poly64x2_t val[2];
78 } poly64x2x2_t;
79 
80 typedef struct poly8x8x3_t {
81   poly8x8_t val[3];
82 } poly8x8x3_t;
83 
84 typedef struct poly8x16x3_t {
85   poly8x16_t val[3];
86 } poly8x16x3_t;
87 
88 typedef struct poly16x4x3_t {
89   poly16x4_t val[3];
90 } poly16x4x3_t;
91 
92 typedef struct poly16x8x3_t {
93   poly16x8_t val[3];
94 } poly16x8x3_t;
95 
96 typedef struct poly64x1x3_t {
97   poly64x1_t val[3];
98 } poly64x1x3_t;
99 
100 typedef struct poly64x2x3_t {
101   poly64x2_t val[3];
102 } poly64x2x3_t;
103 
104 typedef struct poly8x8x4_t {
105   poly8x8_t val[4];
106 } poly8x8x4_t;
107 
108 typedef struct poly8x16x4_t {
109   poly8x16_t val[4];
110 } poly8x16x4_t;
111 
112 typedef struct poly16x4x4_t {
113   poly16x4_t val[4];
114 } poly16x4x4_t;
115 
116 typedef struct poly16x8x4_t {
117   poly16x8_t val[4];
118 } poly16x8x4_t;
119 
120 typedef struct poly64x1x4_t {
121   poly64x1_t val[4];
122 } poly64x1x4_t;
123 
124 typedef struct poly64x2x4_t {
125   poly64x2_t val[4];
126 } poly64x2x4_t;
127 
128 #define __ai static __inline__ __attribute__((__always_inline__, __nodebug__))
129 
130 #ifdef __LITTLE_ENDIAN__
131 #define splat_lane_p8(__p0, __p1) __extension__ ({ \
132   poly8x8_t __ret; \
133   poly8x8_t __s0 = __p0; \
134   __ret = (poly8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 4); \
135   __ret; \
136 })
137 #else
138 #define splat_lane_p8(__p0, __p1) __extension__ ({ \
139   poly8x8_t __ret; \
140   poly8x8_t __s0 = __p0; \
141   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
142   __ret = (poly8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 4); \
143   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
144   __ret; \
145 })
146 #define __noswap_splat_lane_p8(__p0, __p1) __extension__ ({ \
147   poly8x8_t __ret; \
148   poly8x8_t __s0 = __p0; \
149   __ret = (poly8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 4); \
150   __ret; \
151 })
152 #endif
153 
154 #define splat_lane_p64(__p0, __p1) __extension__ ({ \
155   poly64x1_t __ret; \
156   poly64x1_t __s0 = __p0; \
157   __ret = (poly64x1_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 6); \
158   __ret; \
159 })
160 #ifdef __LITTLE_ENDIAN__
161 #define splat_lane_p16(__p0, __p1) __extension__ ({ \
162   poly16x4_t __ret; \
163   poly16x4_t __s0 = __p0; \
164   __ret = (poly16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 5); \
165   __ret; \
166 })
167 #else
168 #define splat_lane_p16(__p0, __p1) __extension__ ({ \
169   poly16x4_t __ret; \
170   poly16x4_t __s0 = __p0; \
171   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
172   __ret = (poly16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 5); \
173   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
174   __ret; \
175 })
176 #define __noswap_splat_lane_p16(__p0, __p1) __extension__ ({ \
177   poly16x4_t __ret; \
178   poly16x4_t __s0 = __p0; \
179   __ret = (poly16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 5); \
180   __ret; \
181 })
182 #endif
183 
184 #ifdef __LITTLE_ENDIAN__
185 #define splatq_lane_p8(__p0, __p1) __extension__ ({ \
186   poly8x16_t __ret; \
187   poly8x8_t __s0 = __p0; \
188   __ret = (poly8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 4); \
189   __ret; \
190 })
191 #else
192 #define splatq_lane_p8(__p0, __p1) __extension__ ({ \
193   poly8x16_t __ret; \
194   poly8x8_t __s0 = __p0; \
195   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
196   __ret = (poly8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 4); \
197   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
198   __ret; \
199 })
200 #define __noswap_splatq_lane_p8(__p0, __p1) __extension__ ({ \
201   poly8x16_t __ret; \
202   poly8x8_t __s0 = __p0; \
203   __ret = (poly8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 4); \
204   __ret; \
205 })
206 #endif
207 
208 #ifdef __LITTLE_ENDIAN__
209 #define splatq_lane_p64(__p0, __p1) __extension__ ({ \
210   poly64x2_t __ret; \
211   poly64x1_t __s0 = __p0; \
212   __ret = (poly64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 6); \
213   __ret; \
214 })
215 #else
216 #define splatq_lane_p64(__p0, __p1) __extension__ ({ \
217   poly64x2_t __ret; \
218   poly64x1_t __s0 = __p0; \
219   __ret = (poly64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 6); \
220   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
221   __ret; \
222 })
223 #define __noswap_splatq_lane_p64(__p0, __p1) __extension__ ({ \
224   poly64x2_t __ret; \
225   poly64x1_t __s0 = __p0; \
226   __ret = (poly64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 6); \
227   __ret; \
228 })
229 #endif
230 
231 #ifdef __LITTLE_ENDIAN__
232 #define splatq_lane_p16(__p0, __p1) __extension__ ({ \
233   poly16x8_t __ret; \
234   poly16x4_t __s0 = __p0; \
235   __ret = (poly16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 5); \
236   __ret; \
237 })
238 #else
239 #define splatq_lane_p16(__p0, __p1) __extension__ ({ \
240   poly16x8_t __ret; \
241   poly16x4_t __s0 = __p0; \
242   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
243   __ret = (poly16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 5); \
244   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
245   __ret; \
246 })
247 #define __noswap_splatq_lane_p16(__p0, __p1) __extension__ ({ \
248   poly16x8_t __ret; \
249   poly16x4_t __s0 = __p0; \
250   __ret = (poly16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 5); \
251   __ret; \
252 })
253 #endif
254 
255 #ifdef __LITTLE_ENDIAN__
256 #define splatq_lane_u8(__p0, __p1) __extension__ ({ \
257   uint8x16_t __ret; \
258   uint8x8_t __s0 = __p0; \
259   __ret = (uint8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 16); \
260   __ret; \
261 })
262 #else
263 #define splatq_lane_u8(__p0, __p1) __extension__ ({ \
264   uint8x16_t __ret; \
265   uint8x8_t __s0 = __p0; \
266   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
267   __ret = (uint8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 16); \
268   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
269   __ret; \
270 })
271 #define __noswap_splatq_lane_u8(__p0, __p1) __extension__ ({ \
272   uint8x16_t __ret; \
273   uint8x8_t __s0 = __p0; \
274   __ret = (uint8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 16); \
275   __ret; \
276 })
277 #endif
278 
279 #ifdef __LITTLE_ENDIAN__
280 #define splatq_lane_u32(__p0, __p1) __extension__ ({ \
281   uint32x4_t __ret; \
282   uint32x2_t __s0 = __p0; \
283   __ret = (uint32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 18); \
284   __ret; \
285 })
286 #else
287 #define splatq_lane_u32(__p0, __p1) __extension__ ({ \
288   uint32x4_t __ret; \
289   uint32x2_t __s0 = __p0; \
290   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
291   __ret = (uint32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 18); \
292   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
293   __ret; \
294 })
295 #define __noswap_splatq_lane_u32(__p0, __p1) __extension__ ({ \
296   uint32x4_t __ret; \
297   uint32x2_t __s0 = __p0; \
298   __ret = (uint32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 18); \
299   __ret; \
300 })
301 #endif
302 
303 #ifdef __LITTLE_ENDIAN__
304 #define splatq_lane_u64(__p0, __p1) __extension__ ({ \
305   uint64x2_t __ret; \
306   uint64x1_t __s0 = __p0; \
307   __ret = (uint64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 19); \
308   __ret; \
309 })
310 #else
311 #define splatq_lane_u64(__p0, __p1) __extension__ ({ \
312   uint64x2_t __ret; \
313   uint64x1_t __s0 = __p0; \
314   __ret = (uint64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 19); \
315   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
316   __ret; \
317 })
318 #define __noswap_splatq_lane_u64(__p0, __p1) __extension__ ({ \
319   uint64x2_t __ret; \
320   uint64x1_t __s0 = __p0; \
321   __ret = (uint64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 19); \
322   __ret; \
323 })
324 #endif
325 
326 #ifdef __LITTLE_ENDIAN__
327 #define splatq_lane_u16(__p0, __p1) __extension__ ({ \
328   uint16x8_t __ret; \
329   uint16x4_t __s0 = __p0; \
330   __ret = (uint16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 17); \
331   __ret; \
332 })
333 #else
334 #define splatq_lane_u16(__p0, __p1) __extension__ ({ \
335   uint16x8_t __ret; \
336   uint16x4_t __s0 = __p0; \
337   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
338   __ret = (uint16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 17); \
339   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
340   __ret; \
341 })
342 #define __noswap_splatq_lane_u16(__p0, __p1) __extension__ ({ \
343   uint16x8_t __ret; \
344   uint16x4_t __s0 = __p0; \
345   __ret = (uint16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 17); \
346   __ret; \
347 })
348 #endif
349 
350 #ifdef __LITTLE_ENDIAN__
351 #define splatq_lane_s8(__p0, __p1) __extension__ ({ \
352   int8x16_t __ret; \
353   int8x8_t __s0 = __p0; \
354   __ret = (int8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 0); \
355   __ret; \
356 })
357 #else
358 #define splatq_lane_s8(__p0, __p1) __extension__ ({ \
359   int8x16_t __ret; \
360   int8x8_t __s0 = __p0; \
361   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
362   __ret = (int8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 0); \
363   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
364   __ret; \
365 })
366 #define __noswap_splatq_lane_s8(__p0, __p1) __extension__ ({ \
367   int8x16_t __ret; \
368   int8x8_t __s0 = __p0; \
369   __ret = (int8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 0); \
370   __ret; \
371 })
372 #endif
373 
374 #ifdef __LITTLE_ENDIAN__
375 #define splatq_lane_f64(__p0, __p1) __extension__ ({ \
376   float64x2_t __ret; \
377   float64x1_t __s0 = __p0; \
378   __ret = (float64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 10); \
379   __ret; \
380 })
381 #else
382 #define splatq_lane_f64(__p0, __p1) __extension__ ({ \
383   float64x2_t __ret; \
384   float64x1_t __s0 = __p0; \
385   __ret = (float64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 10); \
386   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
387   __ret; \
388 })
389 #define __noswap_splatq_lane_f64(__p0, __p1) __extension__ ({ \
390   float64x2_t __ret; \
391   float64x1_t __s0 = __p0; \
392   __ret = (float64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 10); \
393   __ret; \
394 })
395 #endif
396 
397 #ifdef __LITTLE_ENDIAN__
398 #define splatq_lane_f32(__p0, __p1) __extension__ ({ \
399   float32x4_t __ret; \
400   float32x2_t __s0 = __p0; \
401   __ret = (float32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 9); \
402   __ret; \
403 })
404 #else
405 #define splatq_lane_f32(__p0, __p1) __extension__ ({ \
406   float32x4_t __ret; \
407   float32x2_t __s0 = __p0; \
408   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
409   __ret = (float32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 9); \
410   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
411   __ret; \
412 })
413 #define __noswap_splatq_lane_f32(__p0, __p1) __extension__ ({ \
414   float32x4_t __ret; \
415   float32x2_t __s0 = __p0; \
416   __ret = (float32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 9); \
417   __ret; \
418 })
419 #endif
420 
421 #ifdef __LITTLE_ENDIAN__
422 #define splatq_lane_f16(__p0, __p1) __extension__ ({ \
423   float16x8_t __ret; \
424   float16x4_t __s0 = __p0; \
425   __ret = (float16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 8); \
426   __ret; \
427 })
428 #else
429 #define splatq_lane_f16(__p0, __p1) __extension__ ({ \
430   float16x8_t __ret; \
431   float16x4_t __s0 = __p0; \
432   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
433   __ret = (float16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 8); \
434   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
435   __ret; \
436 })
437 #define __noswap_splatq_lane_f16(__p0, __p1) __extension__ ({ \
438   float16x8_t __ret; \
439   float16x4_t __s0 = __p0; \
440   __ret = (float16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 8); \
441   __ret; \
442 })
443 #endif
444 
445 #ifdef __LITTLE_ENDIAN__
446 #define splatq_lane_s32(__p0, __p1) __extension__ ({ \
447   int32x4_t __ret; \
448   int32x2_t __s0 = __p0; \
449   __ret = (int32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 2); \
450   __ret; \
451 })
452 #else
453 #define splatq_lane_s32(__p0, __p1) __extension__ ({ \
454   int32x4_t __ret; \
455   int32x2_t __s0 = __p0; \
456   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
457   __ret = (int32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 2); \
458   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
459   __ret; \
460 })
461 #define __noswap_splatq_lane_s32(__p0, __p1) __extension__ ({ \
462   int32x4_t __ret; \
463   int32x2_t __s0 = __p0; \
464   __ret = (int32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 2); \
465   __ret; \
466 })
467 #endif
468 
469 #ifdef __LITTLE_ENDIAN__
470 #define splatq_lane_s64(__p0, __p1) __extension__ ({ \
471   int64x2_t __ret; \
472   int64x1_t __s0 = __p0; \
473   __ret = (int64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 3); \
474   __ret; \
475 })
476 #else
477 #define splatq_lane_s64(__p0, __p1) __extension__ ({ \
478   int64x2_t __ret; \
479   int64x1_t __s0 = __p0; \
480   __ret = (int64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 3); \
481   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
482   __ret; \
483 })
484 #define __noswap_splatq_lane_s64(__p0, __p1) __extension__ ({ \
485   int64x2_t __ret; \
486   int64x1_t __s0 = __p0; \
487   __ret = (int64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 3); \
488   __ret; \
489 })
490 #endif
491 
492 #ifdef __LITTLE_ENDIAN__
493 #define splatq_lane_s16(__p0, __p1) __extension__ ({ \
494   int16x8_t __ret; \
495   int16x4_t __s0 = __p0; \
496   __ret = (int16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 1); \
497   __ret; \
498 })
499 #else
500 #define splatq_lane_s16(__p0, __p1) __extension__ ({ \
501   int16x8_t __ret; \
502   int16x4_t __s0 = __p0; \
503   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
504   __ret = (int16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 1); \
505   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
506   __ret; \
507 })
508 #define __noswap_splatq_lane_s16(__p0, __p1) __extension__ ({ \
509   int16x8_t __ret; \
510   int16x4_t __s0 = __p0; \
511   __ret = (int16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 1); \
512   __ret; \
513 })
514 #endif
515 
516 #ifdef __LITTLE_ENDIAN__
517 #define splat_lane_u8(__p0, __p1) __extension__ ({ \
518   uint8x8_t __ret; \
519   uint8x8_t __s0 = __p0; \
520   __ret = (uint8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 16); \
521   __ret; \
522 })
523 #else
524 #define splat_lane_u8(__p0, __p1) __extension__ ({ \
525   uint8x8_t __ret; \
526   uint8x8_t __s0 = __p0; \
527   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
528   __ret = (uint8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 16); \
529   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
530   __ret; \
531 })
532 #define __noswap_splat_lane_u8(__p0, __p1) __extension__ ({ \
533   uint8x8_t __ret; \
534   uint8x8_t __s0 = __p0; \
535   __ret = (uint8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 16); \
536   __ret; \
537 })
538 #endif
539 
540 #ifdef __LITTLE_ENDIAN__
541 #define splat_lane_u32(__p0, __p1) __extension__ ({ \
542   uint32x2_t __ret; \
543   uint32x2_t __s0 = __p0; \
544   __ret = (uint32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 18); \
545   __ret; \
546 })
547 #else
548 #define splat_lane_u32(__p0, __p1) __extension__ ({ \
549   uint32x2_t __ret; \
550   uint32x2_t __s0 = __p0; \
551   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
552   __ret = (uint32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 18); \
553   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
554   __ret; \
555 })
556 #define __noswap_splat_lane_u32(__p0, __p1) __extension__ ({ \
557   uint32x2_t __ret; \
558   uint32x2_t __s0 = __p0; \
559   __ret = (uint32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 18); \
560   __ret; \
561 })
562 #endif
563 
564 #define splat_lane_u64(__p0, __p1) __extension__ ({ \
565   uint64x1_t __ret; \
566   uint64x1_t __s0 = __p0; \
567   __ret = (uint64x1_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 19); \
568   __ret; \
569 })
570 #ifdef __LITTLE_ENDIAN__
571 #define splat_lane_u16(__p0, __p1) __extension__ ({ \
572   uint16x4_t __ret; \
573   uint16x4_t __s0 = __p0; \
574   __ret = (uint16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 17); \
575   __ret; \
576 })
577 #else
578 #define splat_lane_u16(__p0, __p1) __extension__ ({ \
579   uint16x4_t __ret; \
580   uint16x4_t __s0 = __p0; \
581   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
582   __ret = (uint16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 17); \
583   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
584   __ret; \
585 })
586 #define __noswap_splat_lane_u16(__p0, __p1) __extension__ ({ \
587   uint16x4_t __ret; \
588   uint16x4_t __s0 = __p0; \
589   __ret = (uint16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 17); \
590   __ret; \
591 })
592 #endif
593 
594 #ifdef __LITTLE_ENDIAN__
595 #define splat_lane_s8(__p0, __p1) __extension__ ({ \
596   int8x8_t __ret; \
597   int8x8_t __s0 = __p0; \
598   __ret = (int8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 0); \
599   __ret; \
600 })
601 #else
602 #define splat_lane_s8(__p0, __p1) __extension__ ({ \
603   int8x8_t __ret; \
604   int8x8_t __s0 = __p0; \
605   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
606   __ret = (int8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 0); \
607   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
608   __ret; \
609 })
610 #define __noswap_splat_lane_s8(__p0, __p1) __extension__ ({ \
611   int8x8_t __ret; \
612   int8x8_t __s0 = __p0; \
613   __ret = (int8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 0); \
614   __ret; \
615 })
616 #endif
617 
618 #define splat_lane_f64(__p0, __p1) __extension__ ({ \
619   float64x1_t __ret; \
620   float64x1_t __s0 = __p0; \
621   __ret = (float64x1_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 10); \
622   __ret; \
623 })
624 #ifdef __LITTLE_ENDIAN__
625 #define splat_lane_f32(__p0, __p1) __extension__ ({ \
626   float32x2_t __ret; \
627   float32x2_t __s0 = __p0; \
628   __ret = (float32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 9); \
629   __ret; \
630 })
631 #else
632 #define splat_lane_f32(__p0, __p1) __extension__ ({ \
633   float32x2_t __ret; \
634   float32x2_t __s0 = __p0; \
635   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
636   __ret = (float32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 9); \
637   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
638   __ret; \
639 })
640 #define __noswap_splat_lane_f32(__p0, __p1) __extension__ ({ \
641   float32x2_t __ret; \
642   float32x2_t __s0 = __p0; \
643   __ret = (float32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 9); \
644   __ret; \
645 })
646 #endif
647 
648 #ifdef __LITTLE_ENDIAN__
649 #define splat_lane_f16(__p0, __p1) __extension__ ({ \
650   float16x4_t __ret; \
651   float16x4_t __s0 = __p0; \
652   __ret = (float16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 8); \
653   __ret; \
654 })
655 #else
656 #define splat_lane_f16(__p0, __p1) __extension__ ({ \
657   float16x4_t __ret; \
658   float16x4_t __s0 = __p0; \
659   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
660   __ret = (float16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 8); \
661   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
662   __ret; \
663 })
664 #define __noswap_splat_lane_f16(__p0, __p1) __extension__ ({ \
665   float16x4_t __ret; \
666   float16x4_t __s0 = __p0; \
667   __ret = (float16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 8); \
668   __ret; \
669 })
670 #endif
671 
672 #ifdef __LITTLE_ENDIAN__
673 #define splat_lane_s32(__p0, __p1) __extension__ ({ \
674   int32x2_t __ret; \
675   int32x2_t __s0 = __p0; \
676   __ret = (int32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 2); \
677   __ret; \
678 })
679 #else
680 #define splat_lane_s32(__p0, __p1) __extension__ ({ \
681   int32x2_t __ret; \
682   int32x2_t __s0 = __p0; \
683   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
684   __ret = (int32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 2); \
685   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
686   __ret; \
687 })
688 #define __noswap_splat_lane_s32(__p0, __p1) __extension__ ({ \
689   int32x2_t __ret; \
690   int32x2_t __s0 = __p0; \
691   __ret = (int32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 2); \
692   __ret; \
693 })
694 #endif
695 
696 #define splat_lane_s64(__p0, __p1) __extension__ ({ \
697   int64x1_t __ret; \
698   int64x1_t __s0 = __p0; \
699   __ret = (int64x1_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 3); \
700   __ret; \
701 })
702 #ifdef __LITTLE_ENDIAN__
703 #define splat_lane_s16(__p0, __p1) __extension__ ({ \
704   int16x4_t __ret; \
705   int16x4_t __s0 = __p0; \
706   __ret = (int16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 1); \
707   __ret; \
708 })
709 #else
710 #define splat_lane_s16(__p0, __p1) __extension__ ({ \
711   int16x4_t __ret; \
712   int16x4_t __s0 = __p0; \
713   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
714   __ret = (int16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 1); \
715   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
716   __ret; \
717 })
718 #define __noswap_splat_lane_s16(__p0, __p1) __extension__ ({ \
719   int16x4_t __ret; \
720   int16x4_t __s0 = __p0; \
721   __ret = (int16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 1); \
722   __ret; \
723 })
724 #endif
725 
726 #ifdef __LITTLE_ENDIAN__
727 #define splat_laneq_p8(__p0, __p1) __extension__ ({ \
728   poly8x8_t __ret; \
729   poly8x16_t __s0 = __p0; \
730   __ret = (poly8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 36); \
731   __ret; \
732 })
733 #else
734 #define splat_laneq_p8(__p0, __p1) __extension__ ({ \
735   poly8x8_t __ret; \
736   poly8x16_t __s0 = __p0; \
737   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
738   __ret = (poly8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 36); \
739   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
740   __ret; \
741 })
742 #define __noswap_splat_laneq_p8(__p0, __p1) __extension__ ({ \
743   poly8x8_t __ret; \
744   poly8x16_t __s0 = __p0; \
745   __ret = (poly8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 36); \
746   __ret; \
747 })
748 #endif
749 
750 #ifdef __LITTLE_ENDIAN__
751 #define splat_laneq_p64(__p0, __p1) __extension__ ({ \
752   poly64x1_t __ret; \
753   poly64x2_t __s0 = __p0; \
754   __ret = (poly64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 38); \
755   __ret; \
756 })
757 #else
758 #define splat_laneq_p64(__p0, __p1) __extension__ ({ \
759   poly64x1_t __ret; \
760   poly64x2_t __s0 = __p0; \
761   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
762   __ret = (poly64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 38); \
763   __ret; \
764 })
765 #define __noswap_splat_laneq_p64(__p0, __p1) __extension__ ({ \
766   poly64x1_t __ret; \
767   poly64x2_t __s0 = __p0; \
768   __ret = (poly64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 38); \
769   __ret; \
770 })
771 #endif
772 
773 #ifdef __LITTLE_ENDIAN__
774 #define splat_laneq_p16(__p0, __p1) __extension__ ({ \
775   poly16x4_t __ret; \
776   poly16x8_t __s0 = __p0; \
777   __ret = (poly16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 37); \
778   __ret; \
779 })
780 #else
781 #define splat_laneq_p16(__p0, __p1) __extension__ ({ \
782   poly16x4_t __ret; \
783   poly16x8_t __s0 = __p0; \
784   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
785   __ret = (poly16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 37); \
786   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
787   __ret; \
788 })
789 #define __noswap_splat_laneq_p16(__p0, __p1) __extension__ ({ \
790   poly16x4_t __ret; \
791   poly16x8_t __s0 = __p0; \
792   __ret = (poly16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 37); \
793   __ret; \
794 })
795 #endif
796 
797 #ifdef __LITTLE_ENDIAN__
798 #define splatq_laneq_p8(__p0, __p1) __extension__ ({ \
799   poly8x16_t __ret; \
800   poly8x16_t __s0 = __p0; \
801   __ret = (poly8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 36); \
802   __ret; \
803 })
804 #else
805 #define splatq_laneq_p8(__p0, __p1) __extension__ ({ \
806   poly8x16_t __ret; \
807   poly8x16_t __s0 = __p0; \
808   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
809   __ret = (poly8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 36); \
810   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
811   __ret; \
812 })
813 #define __noswap_splatq_laneq_p8(__p0, __p1) __extension__ ({ \
814   poly8x16_t __ret; \
815   poly8x16_t __s0 = __p0; \
816   __ret = (poly8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 36); \
817   __ret; \
818 })
819 #endif
820 
821 #ifdef __LITTLE_ENDIAN__
822 #define splatq_laneq_p64(__p0, __p1) __extension__ ({ \
823   poly64x2_t __ret; \
824   poly64x2_t __s0 = __p0; \
825   __ret = (poly64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 38); \
826   __ret; \
827 })
828 #else
829 #define splatq_laneq_p64(__p0, __p1) __extension__ ({ \
830   poly64x2_t __ret; \
831   poly64x2_t __s0 = __p0; \
832   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
833   __ret = (poly64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 38); \
834   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
835   __ret; \
836 })
837 #define __noswap_splatq_laneq_p64(__p0, __p1) __extension__ ({ \
838   poly64x2_t __ret; \
839   poly64x2_t __s0 = __p0; \
840   __ret = (poly64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 38); \
841   __ret; \
842 })
843 #endif
844 
845 #ifdef __LITTLE_ENDIAN__
846 #define splatq_laneq_p16(__p0, __p1) __extension__ ({ \
847   poly16x8_t __ret; \
848   poly16x8_t __s0 = __p0; \
849   __ret = (poly16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 37); \
850   __ret; \
851 })
852 #else
853 #define splatq_laneq_p16(__p0, __p1) __extension__ ({ \
854   poly16x8_t __ret; \
855   poly16x8_t __s0 = __p0; \
856   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
857   __ret = (poly16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 37); \
858   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
859   __ret; \
860 })
861 #define __noswap_splatq_laneq_p16(__p0, __p1) __extension__ ({ \
862   poly16x8_t __ret; \
863   poly16x8_t __s0 = __p0; \
864   __ret = (poly16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 37); \
865   __ret; \
866 })
867 #endif
868 
869 #ifdef __LITTLE_ENDIAN__
870 #define splatq_laneq_u8(__p0, __p1) __extension__ ({ \
871   uint8x16_t __ret; \
872   uint8x16_t __s0 = __p0; \
873   __ret = (uint8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 48); \
874   __ret; \
875 })
876 #else
877 #define splatq_laneq_u8(__p0, __p1) __extension__ ({ \
878   uint8x16_t __ret; \
879   uint8x16_t __s0 = __p0; \
880   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
881   __ret = (uint8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 48); \
882   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
883   __ret; \
884 })
885 #define __noswap_splatq_laneq_u8(__p0, __p1) __extension__ ({ \
886   uint8x16_t __ret; \
887   uint8x16_t __s0 = __p0; \
888   __ret = (uint8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 48); \
889   __ret; \
890 })
891 #endif
892 
893 #ifdef __LITTLE_ENDIAN__
894 #define splatq_laneq_u32(__p0, __p1) __extension__ ({ \
895   uint32x4_t __ret; \
896   uint32x4_t __s0 = __p0; \
897   __ret = (uint32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 50); \
898   __ret; \
899 })
900 #else
901 #define splatq_laneq_u32(__p0, __p1) __extension__ ({ \
902   uint32x4_t __ret; \
903   uint32x4_t __s0 = __p0; \
904   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
905   __ret = (uint32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 50); \
906   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
907   __ret; \
908 })
909 #define __noswap_splatq_laneq_u32(__p0, __p1) __extension__ ({ \
910   uint32x4_t __ret; \
911   uint32x4_t __s0 = __p0; \
912   __ret = (uint32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 50); \
913   __ret; \
914 })
915 #endif
916 
917 #ifdef __LITTLE_ENDIAN__
918 #define splatq_laneq_u64(__p0, __p1) __extension__ ({ \
919   uint64x2_t __ret; \
920   uint64x2_t __s0 = __p0; \
921   __ret = (uint64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 51); \
922   __ret; \
923 })
924 #else
925 #define splatq_laneq_u64(__p0, __p1) __extension__ ({ \
926   uint64x2_t __ret; \
927   uint64x2_t __s0 = __p0; \
928   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
929   __ret = (uint64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 51); \
930   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
931   __ret; \
932 })
933 #define __noswap_splatq_laneq_u64(__p0, __p1) __extension__ ({ \
934   uint64x2_t __ret; \
935   uint64x2_t __s0 = __p0; \
936   __ret = (uint64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 51); \
937   __ret; \
938 })
939 #endif
940 
941 #ifdef __LITTLE_ENDIAN__
942 #define splatq_laneq_u16(__p0, __p1) __extension__ ({ \
943   uint16x8_t __ret; \
944   uint16x8_t __s0 = __p0; \
945   __ret = (uint16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 49); \
946   __ret; \
947 })
948 #else
949 #define splatq_laneq_u16(__p0, __p1) __extension__ ({ \
950   uint16x8_t __ret; \
951   uint16x8_t __s0 = __p0; \
952   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
953   __ret = (uint16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 49); \
954   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
955   __ret; \
956 })
957 #define __noswap_splatq_laneq_u16(__p0, __p1) __extension__ ({ \
958   uint16x8_t __ret; \
959   uint16x8_t __s0 = __p0; \
960   __ret = (uint16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 49); \
961   __ret; \
962 })
963 #endif
964 
965 #ifdef __LITTLE_ENDIAN__
966 #define splatq_laneq_s8(__p0, __p1) __extension__ ({ \
967   int8x16_t __ret; \
968   int8x16_t __s0 = __p0; \
969   __ret = (int8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 32); \
970   __ret; \
971 })
972 #else
973 #define splatq_laneq_s8(__p0, __p1) __extension__ ({ \
974   int8x16_t __ret; \
975   int8x16_t __s0 = __p0; \
976   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
977   __ret = (int8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 32); \
978   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
979   __ret; \
980 })
981 #define __noswap_splatq_laneq_s8(__p0, __p1) __extension__ ({ \
982   int8x16_t __ret; \
983   int8x16_t __s0 = __p0; \
984   __ret = (int8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 32); \
985   __ret; \
986 })
987 #endif
988 
989 #ifdef __LITTLE_ENDIAN__
990 #define splatq_laneq_f64(__p0, __p1) __extension__ ({ \
991   float64x2_t __ret; \
992   float64x2_t __s0 = __p0; \
993   __ret = (float64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 42); \
994   __ret; \
995 })
996 #else
997 #define splatq_laneq_f64(__p0, __p1) __extension__ ({ \
998   float64x2_t __ret; \
999   float64x2_t __s0 = __p0; \
1000   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1001   __ret = (float64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 42); \
1002   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1003   __ret; \
1004 })
1005 #define __noswap_splatq_laneq_f64(__p0, __p1) __extension__ ({ \
1006   float64x2_t __ret; \
1007   float64x2_t __s0 = __p0; \
1008   __ret = (float64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 42); \
1009   __ret; \
1010 })
1011 #endif
1012 
1013 #ifdef __LITTLE_ENDIAN__
1014 #define splatq_laneq_f32(__p0, __p1) __extension__ ({ \
1015   float32x4_t __ret; \
1016   float32x4_t __s0 = __p0; \
1017   __ret = (float32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 41); \
1018   __ret; \
1019 })
1020 #else
1021 #define splatq_laneq_f32(__p0, __p1) __extension__ ({ \
1022   float32x4_t __ret; \
1023   float32x4_t __s0 = __p0; \
1024   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1025   __ret = (float32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 41); \
1026   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1027   __ret; \
1028 })
1029 #define __noswap_splatq_laneq_f32(__p0, __p1) __extension__ ({ \
1030   float32x4_t __ret; \
1031   float32x4_t __s0 = __p0; \
1032   __ret = (float32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 41); \
1033   __ret; \
1034 })
1035 #endif
1036 
1037 #ifdef __LITTLE_ENDIAN__
1038 #define splatq_laneq_f16(__p0, __p1) __extension__ ({ \
1039   float16x8_t __ret; \
1040   float16x8_t __s0 = __p0; \
1041   __ret = (float16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 40); \
1042   __ret; \
1043 })
1044 #else
1045 #define splatq_laneq_f16(__p0, __p1) __extension__ ({ \
1046   float16x8_t __ret; \
1047   float16x8_t __s0 = __p0; \
1048   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1049   __ret = (float16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 40); \
1050   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1051   __ret; \
1052 })
1053 #define __noswap_splatq_laneq_f16(__p0, __p1) __extension__ ({ \
1054   float16x8_t __ret; \
1055   float16x8_t __s0 = __p0; \
1056   __ret = (float16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 40); \
1057   __ret; \
1058 })
1059 #endif
1060 
1061 #ifdef __LITTLE_ENDIAN__
1062 #define splatq_laneq_s32(__p0, __p1) __extension__ ({ \
1063   int32x4_t __ret; \
1064   int32x4_t __s0 = __p0; \
1065   __ret = (int32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 34); \
1066   __ret; \
1067 })
1068 #else
1069 #define splatq_laneq_s32(__p0, __p1) __extension__ ({ \
1070   int32x4_t __ret; \
1071   int32x4_t __s0 = __p0; \
1072   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1073   __ret = (int32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 34); \
1074   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1075   __ret; \
1076 })
1077 #define __noswap_splatq_laneq_s32(__p0, __p1) __extension__ ({ \
1078   int32x4_t __ret; \
1079   int32x4_t __s0 = __p0; \
1080   __ret = (int32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 34); \
1081   __ret; \
1082 })
1083 #endif
1084 
1085 #ifdef __LITTLE_ENDIAN__
1086 #define splatq_laneq_s64(__p0, __p1) __extension__ ({ \
1087   int64x2_t __ret; \
1088   int64x2_t __s0 = __p0; \
1089   __ret = (int64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 35); \
1090   __ret; \
1091 })
1092 #else
1093 #define splatq_laneq_s64(__p0, __p1) __extension__ ({ \
1094   int64x2_t __ret; \
1095   int64x2_t __s0 = __p0; \
1096   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1097   __ret = (int64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 35); \
1098   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1099   __ret; \
1100 })
1101 #define __noswap_splatq_laneq_s64(__p0, __p1) __extension__ ({ \
1102   int64x2_t __ret; \
1103   int64x2_t __s0 = __p0; \
1104   __ret = (int64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 35); \
1105   __ret; \
1106 })
1107 #endif
1108 
1109 #ifdef __LITTLE_ENDIAN__
1110 #define splatq_laneq_s16(__p0, __p1) __extension__ ({ \
1111   int16x8_t __ret; \
1112   int16x8_t __s0 = __p0; \
1113   __ret = (int16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 33); \
1114   __ret; \
1115 })
1116 #else
1117 #define splatq_laneq_s16(__p0, __p1) __extension__ ({ \
1118   int16x8_t __ret; \
1119   int16x8_t __s0 = __p0; \
1120   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1121   __ret = (int16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 33); \
1122   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1123   __ret; \
1124 })
1125 #define __noswap_splatq_laneq_s16(__p0, __p1) __extension__ ({ \
1126   int16x8_t __ret; \
1127   int16x8_t __s0 = __p0; \
1128   __ret = (int16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 33); \
1129   __ret; \
1130 })
1131 #endif
1132 
1133 #ifdef __LITTLE_ENDIAN__
1134 #define splat_laneq_u8(__p0, __p1) __extension__ ({ \
1135   uint8x8_t __ret; \
1136   uint8x16_t __s0 = __p0; \
1137   __ret = (uint8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 48); \
1138   __ret; \
1139 })
1140 #else
1141 #define splat_laneq_u8(__p0, __p1) __extension__ ({ \
1142   uint8x8_t __ret; \
1143   uint8x16_t __s0 = __p0; \
1144   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1145   __ret = (uint8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 48); \
1146   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1147   __ret; \
1148 })
1149 #define __noswap_splat_laneq_u8(__p0, __p1) __extension__ ({ \
1150   uint8x8_t __ret; \
1151   uint8x16_t __s0 = __p0; \
1152   __ret = (uint8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 48); \
1153   __ret; \
1154 })
1155 #endif
1156 
1157 #ifdef __LITTLE_ENDIAN__
1158 #define splat_laneq_u32(__p0, __p1) __extension__ ({ \
1159   uint32x2_t __ret; \
1160   uint32x4_t __s0 = __p0; \
1161   __ret = (uint32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 50); \
1162   __ret; \
1163 })
1164 #else
1165 #define splat_laneq_u32(__p0, __p1) __extension__ ({ \
1166   uint32x2_t __ret; \
1167   uint32x4_t __s0 = __p0; \
1168   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1169   __ret = (uint32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 50); \
1170   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1171   __ret; \
1172 })
1173 #define __noswap_splat_laneq_u32(__p0, __p1) __extension__ ({ \
1174   uint32x2_t __ret; \
1175   uint32x4_t __s0 = __p0; \
1176   __ret = (uint32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 50); \
1177   __ret; \
1178 })
1179 #endif
1180 
1181 #ifdef __LITTLE_ENDIAN__
1182 #define splat_laneq_u64(__p0, __p1) __extension__ ({ \
1183   uint64x1_t __ret; \
1184   uint64x2_t __s0 = __p0; \
1185   __ret = (uint64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 51); \
1186   __ret; \
1187 })
1188 #else
1189 #define splat_laneq_u64(__p0, __p1) __extension__ ({ \
1190   uint64x1_t __ret; \
1191   uint64x2_t __s0 = __p0; \
1192   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1193   __ret = (uint64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 51); \
1194   __ret; \
1195 })
1196 #define __noswap_splat_laneq_u64(__p0, __p1) __extension__ ({ \
1197   uint64x1_t __ret; \
1198   uint64x2_t __s0 = __p0; \
1199   __ret = (uint64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 51); \
1200   __ret; \
1201 })
1202 #endif
1203 
1204 #ifdef __LITTLE_ENDIAN__
1205 #define splat_laneq_u16(__p0, __p1) __extension__ ({ \
1206   uint16x4_t __ret; \
1207   uint16x8_t __s0 = __p0; \
1208   __ret = (uint16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 49); \
1209   __ret; \
1210 })
1211 #else
1212 #define splat_laneq_u16(__p0, __p1) __extension__ ({ \
1213   uint16x4_t __ret; \
1214   uint16x8_t __s0 = __p0; \
1215   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1216   __ret = (uint16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 49); \
1217   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1218   __ret; \
1219 })
1220 #define __noswap_splat_laneq_u16(__p0, __p1) __extension__ ({ \
1221   uint16x4_t __ret; \
1222   uint16x8_t __s0 = __p0; \
1223   __ret = (uint16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 49); \
1224   __ret; \
1225 })
1226 #endif
1227 
1228 #ifdef __LITTLE_ENDIAN__
1229 #define splat_laneq_s8(__p0, __p1) __extension__ ({ \
1230   int8x8_t __ret; \
1231   int8x16_t __s0 = __p0; \
1232   __ret = (int8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 32); \
1233   __ret; \
1234 })
1235 #else
1236 #define splat_laneq_s8(__p0, __p1) __extension__ ({ \
1237   int8x8_t __ret; \
1238   int8x16_t __s0 = __p0; \
1239   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1240   __ret = (int8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 32); \
1241   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1242   __ret; \
1243 })
1244 #define __noswap_splat_laneq_s8(__p0, __p1) __extension__ ({ \
1245   int8x8_t __ret; \
1246   int8x16_t __s0 = __p0; \
1247   __ret = (int8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 32); \
1248   __ret; \
1249 })
1250 #endif
1251 
1252 #ifdef __LITTLE_ENDIAN__
1253 #define splat_laneq_f64(__p0, __p1) __extension__ ({ \
1254   float64x1_t __ret; \
1255   float64x2_t __s0 = __p0; \
1256   __ret = (float64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 42); \
1257   __ret; \
1258 })
1259 #else
1260 #define splat_laneq_f64(__p0, __p1) __extension__ ({ \
1261   float64x1_t __ret; \
1262   float64x2_t __s0 = __p0; \
1263   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1264   __ret = (float64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 42); \
1265   __ret; \
1266 })
1267 #define __noswap_splat_laneq_f64(__p0, __p1) __extension__ ({ \
1268   float64x1_t __ret; \
1269   float64x2_t __s0 = __p0; \
1270   __ret = (float64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 42); \
1271   __ret; \
1272 })
1273 #endif
1274 
1275 #ifdef __LITTLE_ENDIAN__
1276 #define splat_laneq_f32(__p0, __p1) __extension__ ({ \
1277   float32x2_t __ret; \
1278   float32x4_t __s0 = __p0; \
1279   __ret = (float32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 41); \
1280   __ret; \
1281 })
1282 #else
1283 #define splat_laneq_f32(__p0, __p1) __extension__ ({ \
1284   float32x2_t __ret; \
1285   float32x4_t __s0 = __p0; \
1286   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1287   __ret = (float32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 41); \
1288   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1289   __ret; \
1290 })
1291 #define __noswap_splat_laneq_f32(__p0, __p1) __extension__ ({ \
1292   float32x2_t __ret; \
1293   float32x4_t __s0 = __p0; \
1294   __ret = (float32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 41); \
1295   __ret; \
1296 })
1297 #endif
1298 
1299 #ifdef __LITTLE_ENDIAN__
1300 #define splat_laneq_f16(__p0, __p1) __extension__ ({ \
1301   float16x4_t __ret; \
1302   float16x8_t __s0 = __p0; \
1303   __ret = (float16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 40); \
1304   __ret; \
1305 })
1306 #else
1307 #define splat_laneq_f16(__p0, __p1) __extension__ ({ \
1308   float16x4_t __ret; \
1309   float16x8_t __s0 = __p0; \
1310   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1311   __ret = (float16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 40); \
1312   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1313   __ret; \
1314 })
1315 #define __noswap_splat_laneq_f16(__p0, __p1) __extension__ ({ \
1316   float16x4_t __ret; \
1317   float16x8_t __s0 = __p0; \
1318   __ret = (float16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 40); \
1319   __ret; \
1320 })
1321 #endif
1322 
1323 #ifdef __LITTLE_ENDIAN__
1324 #define splat_laneq_s32(__p0, __p1) __extension__ ({ \
1325   int32x2_t __ret; \
1326   int32x4_t __s0 = __p0; \
1327   __ret = (int32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 34); \
1328   __ret; \
1329 })
1330 #else
1331 #define splat_laneq_s32(__p0, __p1) __extension__ ({ \
1332   int32x2_t __ret; \
1333   int32x4_t __s0 = __p0; \
1334   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1335   __ret = (int32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 34); \
1336   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1337   __ret; \
1338 })
1339 #define __noswap_splat_laneq_s32(__p0, __p1) __extension__ ({ \
1340   int32x2_t __ret; \
1341   int32x4_t __s0 = __p0; \
1342   __ret = (int32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 34); \
1343   __ret; \
1344 })
1345 #endif
1346 
1347 #ifdef __LITTLE_ENDIAN__
1348 #define splat_laneq_s64(__p0, __p1) __extension__ ({ \
1349   int64x1_t __ret; \
1350   int64x2_t __s0 = __p0; \
1351   __ret = (int64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 35); \
1352   __ret; \
1353 })
1354 #else
1355 #define splat_laneq_s64(__p0, __p1) __extension__ ({ \
1356   int64x1_t __ret; \
1357   int64x2_t __s0 = __p0; \
1358   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1359   __ret = (int64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 35); \
1360   __ret; \
1361 })
1362 #define __noswap_splat_laneq_s64(__p0, __p1) __extension__ ({ \
1363   int64x1_t __ret; \
1364   int64x2_t __s0 = __p0; \
1365   __ret = (int64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 35); \
1366   __ret; \
1367 })
1368 #endif
1369 
1370 #ifdef __LITTLE_ENDIAN__
1371 #define splat_laneq_s16(__p0, __p1) __extension__ ({ \
1372   int16x4_t __ret; \
1373   int16x8_t __s0 = __p0; \
1374   __ret = (int16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 33); \
1375   __ret; \
1376 })
1377 #else
1378 #define splat_laneq_s16(__p0, __p1) __extension__ ({ \
1379   int16x4_t __ret; \
1380   int16x8_t __s0 = __p0; \
1381   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1382   __ret = (int16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 33); \
1383   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1384   __ret; \
1385 })
1386 #define __noswap_splat_laneq_s16(__p0, __p1) __extension__ ({ \
1387   int16x4_t __ret; \
1388   int16x8_t __s0 = __p0; \
1389   __ret = (int16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 33); \
1390   __ret; \
1391 })
1392 #endif
1393 
1394 #ifdef __LITTLE_ENDIAN__
vabdq_u8(uint8x16_t __p0,uint8x16_t __p1)1395 __ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1396   uint8x16_t __ret;
1397   __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
1398   return __ret;
1399 }
1400 #else
vabdq_u8(uint8x16_t __p0,uint8x16_t __p1)1401 __ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1402   uint8x16_t __ret;
1403   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1404   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1405   __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
1406   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1407   return __ret;
1408 }
__noswap_vabdq_u8(uint8x16_t __p0,uint8x16_t __p1)1409 __ai uint8x16_t __noswap_vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1410   uint8x16_t __ret;
1411   __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
1412   return __ret;
1413 }
1414 #endif
1415 
1416 #ifdef __LITTLE_ENDIAN__
vabdq_u32(uint32x4_t __p0,uint32x4_t __p1)1417 __ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1418   uint32x4_t __ret;
1419   __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
1420   return __ret;
1421 }
1422 #else
vabdq_u32(uint32x4_t __p0,uint32x4_t __p1)1423 __ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1424   uint32x4_t __ret;
1425   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1426   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1427   __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
1428   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1429   return __ret;
1430 }
__noswap_vabdq_u32(uint32x4_t __p0,uint32x4_t __p1)1431 __ai uint32x4_t __noswap_vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1432   uint32x4_t __ret;
1433   __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
1434   return __ret;
1435 }
1436 #endif
1437 
1438 #ifdef __LITTLE_ENDIAN__
vabdq_u16(uint16x8_t __p0,uint16x8_t __p1)1439 __ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1440   uint16x8_t __ret;
1441   __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
1442   return __ret;
1443 }
1444 #else
vabdq_u16(uint16x8_t __p0,uint16x8_t __p1)1445 __ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1446   uint16x8_t __ret;
1447   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1448   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1449   __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
1450   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1451   return __ret;
1452 }
__noswap_vabdq_u16(uint16x8_t __p0,uint16x8_t __p1)1453 __ai uint16x8_t __noswap_vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1454   uint16x8_t __ret;
1455   __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
1456   return __ret;
1457 }
1458 #endif
1459 
1460 #ifdef __LITTLE_ENDIAN__
vabdq_s8(int8x16_t __p0,int8x16_t __p1)1461 __ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
1462   int8x16_t __ret;
1463   __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
1464   return __ret;
1465 }
1466 #else
vabdq_s8(int8x16_t __p0,int8x16_t __p1)1467 __ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
1468   int8x16_t __ret;
1469   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1470   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1471   __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
1472   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1473   return __ret;
1474 }
__noswap_vabdq_s8(int8x16_t __p0,int8x16_t __p1)1475 __ai int8x16_t __noswap_vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
1476   int8x16_t __ret;
1477   __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
1478   return __ret;
1479 }
1480 #endif
1481 
1482 #ifdef __LITTLE_ENDIAN__
vabdq_f32(float32x4_t __p0,float32x4_t __p1)1483 __ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
1484   float32x4_t __ret;
1485   __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
1486   return __ret;
1487 }
1488 #else
vabdq_f32(float32x4_t __p0,float32x4_t __p1)1489 __ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
1490   float32x4_t __ret;
1491   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1492   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1493   __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
1494   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1495   return __ret;
1496 }
1497 #endif
1498 
1499 #ifdef __LITTLE_ENDIAN__
vabdq_s32(int32x4_t __p0,int32x4_t __p1)1500 __ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
1501   int32x4_t __ret;
1502   __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
1503   return __ret;
1504 }
1505 #else
vabdq_s32(int32x4_t __p0,int32x4_t __p1)1506 __ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
1507   int32x4_t __ret;
1508   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1509   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1510   __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
1511   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1512   return __ret;
1513 }
__noswap_vabdq_s32(int32x4_t __p0,int32x4_t __p1)1514 __ai int32x4_t __noswap_vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
1515   int32x4_t __ret;
1516   __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
1517   return __ret;
1518 }
1519 #endif
1520 
1521 #ifdef __LITTLE_ENDIAN__
vabdq_s16(int16x8_t __p0,int16x8_t __p1)1522 __ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
1523   int16x8_t __ret;
1524   __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
1525   return __ret;
1526 }
1527 #else
vabdq_s16(int16x8_t __p0,int16x8_t __p1)1528 __ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
1529   int16x8_t __ret;
1530   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1531   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1532   __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
1533   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1534   return __ret;
1535 }
__noswap_vabdq_s16(int16x8_t __p0,int16x8_t __p1)1536 __ai int16x8_t __noswap_vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
1537   int16x8_t __ret;
1538   __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
1539   return __ret;
1540 }
1541 #endif
1542 
1543 #ifdef __LITTLE_ENDIAN__
vabd_u8(uint8x8_t __p0,uint8x8_t __p1)1544 __ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1545   uint8x8_t __ret;
1546   __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
1547   return __ret;
1548 }
1549 #else
vabd_u8(uint8x8_t __p0,uint8x8_t __p1)1550 __ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1551   uint8x8_t __ret;
1552   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1553   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1554   __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
1555   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1556   return __ret;
1557 }
__noswap_vabd_u8(uint8x8_t __p0,uint8x8_t __p1)1558 __ai uint8x8_t __noswap_vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1559   uint8x8_t __ret;
1560   __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
1561   return __ret;
1562 }
1563 #endif
1564 
1565 #ifdef __LITTLE_ENDIAN__
vabd_u32(uint32x2_t __p0,uint32x2_t __p1)1566 __ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1567   uint32x2_t __ret;
1568   __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
1569   return __ret;
1570 }
1571 #else
vabd_u32(uint32x2_t __p0,uint32x2_t __p1)1572 __ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1573   uint32x2_t __ret;
1574   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1575   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1576   __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
1577   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1578   return __ret;
1579 }
__noswap_vabd_u32(uint32x2_t __p0,uint32x2_t __p1)1580 __ai uint32x2_t __noswap_vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1581   uint32x2_t __ret;
1582   __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
1583   return __ret;
1584 }
1585 #endif
1586 
1587 #ifdef __LITTLE_ENDIAN__
vabd_u16(uint16x4_t __p0,uint16x4_t __p1)1588 __ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1589   uint16x4_t __ret;
1590   __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
1591   return __ret;
1592 }
1593 #else
vabd_u16(uint16x4_t __p0,uint16x4_t __p1)1594 __ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1595   uint16x4_t __ret;
1596   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1597   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1598   __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
1599   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1600   return __ret;
1601 }
__noswap_vabd_u16(uint16x4_t __p0,uint16x4_t __p1)1602 __ai uint16x4_t __noswap_vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1603   uint16x4_t __ret;
1604   __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
1605   return __ret;
1606 }
1607 #endif
1608 
1609 #ifdef __LITTLE_ENDIAN__
vabd_s8(int8x8_t __p0,int8x8_t __p1)1610 __ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
1611   int8x8_t __ret;
1612   __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
1613   return __ret;
1614 }
1615 #else
vabd_s8(int8x8_t __p0,int8x8_t __p1)1616 __ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
1617   int8x8_t __ret;
1618   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1619   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1620   __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
1621   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1622   return __ret;
1623 }
__noswap_vabd_s8(int8x8_t __p0,int8x8_t __p1)1624 __ai int8x8_t __noswap_vabd_s8(int8x8_t __p0, int8x8_t __p1) {
1625   int8x8_t __ret;
1626   __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
1627   return __ret;
1628 }
1629 #endif
1630 
1631 #ifdef __LITTLE_ENDIAN__
vabd_f32(float32x2_t __p0,float32x2_t __p1)1632 __ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
1633   float32x2_t __ret;
1634   __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
1635   return __ret;
1636 }
1637 #else
vabd_f32(float32x2_t __p0,float32x2_t __p1)1638 __ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
1639   float32x2_t __ret;
1640   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1641   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1642   __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
1643   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1644   return __ret;
1645 }
1646 #endif
1647 
1648 #ifdef __LITTLE_ENDIAN__
vabd_s32(int32x2_t __p0,int32x2_t __p1)1649 __ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
1650   int32x2_t __ret;
1651   __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
1652   return __ret;
1653 }
1654 #else
vabd_s32(int32x2_t __p0,int32x2_t __p1)1655 __ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
1656   int32x2_t __ret;
1657   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1658   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1659   __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
1660   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1661   return __ret;
1662 }
__noswap_vabd_s32(int32x2_t __p0,int32x2_t __p1)1663 __ai int32x2_t __noswap_vabd_s32(int32x2_t __p0, int32x2_t __p1) {
1664   int32x2_t __ret;
1665   __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
1666   return __ret;
1667 }
1668 #endif
1669 
1670 #ifdef __LITTLE_ENDIAN__
vabd_s16(int16x4_t __p0,int16x4_t __p1)1671 __ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
1672   int16x4_t __ret;
1673   __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
1674   return __ret;
1675 }
1676 #else
vabd_s16(int16x4_t __p0,int16x4_t __p1)1677 __ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
1678   int16x4_t __ret;
1679   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1680   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1681   __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
1682   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1683   return __ret;
1684 }
__noswap_vabd_s16(int16x4_t __p0,int16x4_t __p1)1685 __ai int16x4_t __noswap_vabd_s16(int16x4_t __p0, int16x4_t __p1) {
1686   int16x4_t __ret;
1687   __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
1688   return __ret;
1689 }
1690 #endif
1691 
1692 #ifdef __LITTLE_ENDIAN__
vabsq_s8(int8x16_t __p0)1693 __ai int8x16_t vabsq_s8(int8x16_t __p0) {
1694   int8x16_t __ret;
1695   __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 32);
1696   return __ret;
1697 }
1698 #else
vabsq_s8(int8x16_t __p0)1699 __ai int8x16_t vabsq_s8(int8x16_t __p0) {
1700   int8x16_t __ret;
1701   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1702   __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 32);
1703   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1704   return __ret;
1705 }
1706 #endif
1707 
1708 #ifdef __LITTLE_ENDIAN__
vabsq_f32(float32x4_t __p0)1709 __ai float32x4_t vabsq_f32(float32x4_t __p0) {
1710   float32x4_t __ret;
1711   __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 41);
1712   return __ret;
1713 }
1714 #else
vabsq_f32(float32x4_t __p0)1715 __ai float32x4_t vabsq_f32(float32x4_t __p0) {
1716   float32x4_t __ret;
1717   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1718   __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 41);
1719   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1720   return __ret;
1721 }
1722 #endif
1723 
1724 #ifdef __LITTLE_ENDIAN__
vabsq_s32(int32x4_t __p0)1725 __ai int32x4_t vabsq_s32(int32x4_t __p0) {
1726   int32x4_t __ret;
1727   __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 34);
1728   return __ret;
1729 }
1730 #else
vabsq_s32(int32x4_t __p0)1731 __ai int32x4_t vabsq_s32(int32x4_t __p0) {
1732   int32x4_t __ret;
1733   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1734   __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 34);
1735   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1736   return __ret;
1737 }
1738 #endif
1739 
1740 #ifdef __LITTLE_ENDIAN__
vabsq_s16(int16x8_t __p0)1741 __ai int16x8_t vabsq_s16(int16x8_t __p0) {
1742   int16x8_t __ret;
1743   __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 33);
1744   return __ret;
1745 }
1746 #else
vabsq_s16(int16x8_t __p0)1747 __ai int16x8_t vabsq_s16(int16x8_t __p0) {
1748   int16x8_t __ret;
1749   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1750   __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 33);
1751   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1752   return __ret;
1753 }
1754 #endif
1755 
1756 #ifdef __LITTLE_ENDIAN__
vabs_s8(int8x8_t __p0)1757 __ai int8x8_t vabs_s8(int8x8_t __p0) {
1758   int8x8_t __ret;
1759   __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__p0, 0);
1760   return __ret;
1761 }
1762 #else
vabs_s8(int8x8_t __p0)1763 __ai int8x8_t vabs_s8(int8x8_t __p0) {
1764   int8x8_t __ret;
1765   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1766   __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 0);
1767   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1768   return __ret;
1769 }
1770 #endif
1771 
1772 #ifdef __LITTLE_ENDIAN__
vabs_f32(float32x2_t __p0)1773 __ai float32x2_t vabs_f32(float32x2_t __p0) {
1774   float32x2_t __ret;
1775   __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 9);
1776   return __ret;
1777 }
1778 #else
vabs_f32(float32x2_t __p0)1779 __ai float32x2_t vabs_f32(float32x2_t __p0) {
1780   float32x2_t __ret;
1781   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1782   __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 9);
1783   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1784   return __ret;
1785 }
1786 #endif
1787 
1788 #ifdef __LITTLE_ENDIAN__
vabs_s32(int32x2_t __p0)1789 __ai int32x2_t vabs_s32(int32x2_t __p0) {
1790   int32x2_t __ret;
1791   __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 2);
1792   return __ret;
1793 }
1794 #else
vabs_s32(int32x2_t __p0)1795 __ai int32x2_t vabs_s32(int32x2_t __p0) {
1796   int32x2_t __ret;
1797   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1798   __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 2);
1799   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1800   return __ret;
1801 }
1802 #endif
1803 
1804 #ifdef __LITTLE_ENDIAN__
vabs_s16(int16x4_t __p0)1805 __ai int16x4_t vabs_s16(int16x4_t __p0) {
1806   int16x4_t __ret;
1807   __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__p0, 1);
1808   return __ret;
1809 }
1810 #else
vabs_s16(int16x4_t __p0)1811 __ai int16x4_t vabs_s16(int16x4_t __p0) {
1812   int16x4_t __ret;
1813   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1814   __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 1);
1815   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1816   return __ret;
1817 }
1818 #endif
1819 
1820 #ifdef __LITTLE_ENDIAN__
vaddq_u8(uint8x16_t __p0,uint8x16_t __p1)1821 __ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1822   uint8x16_t __ret;
1823   __ret = __p0 + __p1;
1824   return __ret;
1825 }
1826 #else
vaddq_u8(uint8x16_t __p0,uint8x16_t __p1)1827 __ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1828   uint8x16_t __ret;
1829   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1830   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1831   __ret = __rev0 + __rev1;
1832   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1833   return __ret;
1834 }
1835 #endif
1836 
1837 #ifdef __LITTLE_ENDIAN__
vaddq_u32(uint32x4_t __p0,uint32x4_t __p1)1838 __ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1839   uint32x4_t __ret;
1840   __ret = __p0 + __p1;
1841   return __ret;
1842 }
1843 #else
vaddq_u32(uint32x4_t __p0,uint32x4_t __p1)1844 __ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1845   uint32x4_t __ret;
1846   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1847   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1848   __ret = __rev0 + __rev1;
1849   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1850   return __ret;
1851 }
1852 #endif
1853 
1854 #ifdef __LITTLE_ENDIAN__
vaddq_u64(uint64x2_t __p0,uint64x2_t __p1)1855 __ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1856   uint64x2_t __ret;
1857   __ret = __p0 + __p1;
1858   return __ret;
1859 }
1860 #else
vaddq_u64(uint64x2_t __p0,uint64x2_t __p1)1861 __ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1862   uint64x2_t __ret;
1863   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1864   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1865   __ret = __rev0 + __rev1;
1866   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1867   return __ret;
1868 }
1869 #endif
1870 
1871 #ifdef __LITTLE_ENDIAN__
vaddq_u16(uint16x8_t __p0,uint16x8_t __p1)1872 __ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1873   uint16x8_t __ret;
1874   __ret = __p0 + __p1;
1875   return __ret;
1876 }
1877 #else
vaddq_u16(uint16x8_t __p0,uint16x8_t __p1)1878 __ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1879   uint16x8_t __ret;
1880   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1881   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1882   __ret = __rev0 + __rev1;
1883   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1884   return __ret;
1885 }
1886 #endif
1887 
1888 #ifdef __LITTLE_ENDIAN__
vaddq_s8(int8x16_t __p0,int8x16_t __p1)1889 __ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
1890   int8x16_t __ret;
1891   __ret = __p0 + __p1;
1892   return __ret;
1893 }
1894 #else
vaddq_s8(int8x16_t __p0,int8x16_t __p1)1895 __ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
1896   int8x16_t __ret;
1897   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1898   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1899   __ret = __rev0 + __rev1;
1900   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1901   return __ret;
1902 }
1903 #endif
1904 
1905 #ifdef __LITTLE_ENDIAN__
vaddq_f32(float32x4_t __p0,float32x4_t __p1)1906 __ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
1907   float32x4_t __ret;
1908   __ret = __p0 + __p1;
1909   return __ret;
1910 }
1911 #else
vaddq_f32(float32x4_t __p0,float32x4_t __p1)1912 __ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
1913   float32x4_t __ret;
1914   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1915   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1916   __ret = __rev0 + __rev1;
1917   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1918   return __ret;
1919 }
1920 #endif
1921 
1922 #ifdef __LITTLE_ENDIAN__
vaddq_s32(int32x4_t __p0,int32x4_t __p1)1923 __ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
1924   int32x4_t __ret;
1925   __ret = __p0 + __p1;
1926   return __ret;
1927 }
1928 #else
vaddq_s32(int32x4_t __p0,int32x4_t __p1)1929 __ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
1930   int32x4_t __ret;
1931   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1932   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1933   __ret = __rev0 + __rev1;
1934   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1935   return __ret;
1936 }
1937 #endif
1938 
1939 #ifdef __LITTLE_ENDIAN__
vaddq_s64(int64x2_t __p0,int64x2_t __p1)1940 __ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
1941   int64x2_t __ret;
1942   __ret = __p0 + __p1;
1943   return __ret;
1944 }
1945 #else
vaddq_s64(int64x2_t __p0,int64x2_t __p1)1946 __ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
1947   int64x2_t __ret;
1948   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1949   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1950   __ret = __rev0 + __rev1;
1951   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1952   return __ret;
1953 }
1954 #endif
1955 
1956 #ifdef __LITTLE_ENDIAN__
vaddq_s16(int16x8_t __p0,int16x8_t __p1)1957 __ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
1958   int16x8_t __ret;
1959   __ret = __p0 + __p1;
1960   return __ret;
1961 }
1962 #else
vaddq_s16(int16x8_t __p0,int16x8_t __p1)1963 __ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
1964   int16x8_t __ret;
1965   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1966   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1967   __ret = __rev0 + __rev1;
1968   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1969   return __ret;
1970 }
1971 #endif
1972 
1973 #ifdef __LITTLE_ENDIAN__
vadd_u8(uint8x8_t __p0,uint8x8_t __p1)1974 __ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1975   uint8x8_t __ret;
1976   __ret = __p0 + __p1;
1977   return __ret;
1978 }
1979 #else
vadd_u8(uint8x8_t __p0,uint8x8_t __p1)1980 __ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1981   uint8x8_t __ret;
1982   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1983   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1984   __ret = __rev0 + __rev1;
1985   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1986   return __ret;
1987 }
1988 #endif
1989 
1990 #ifdef __LITTLE_ENDIAN__
vadd_u32(uint32x2_t __p0,uint32x2_t __p1)1991 __ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1992   uint32x2_t __ret;
1993   __ret = __p0 + __p1;
1994   return __ret;
1995 }
1996 #else
vadd_u32(uint32x2_t __p0,uint32x2_t __p1)1997 __ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1998   uint32x2_t __ret;
1999   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2000   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2001   __ret = __rev0 + __rev1;
2002   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2003   return __ret;
2004 }
2005 #endif
2006 
vadd_u64(uint64x1_t __p0,uint64x1_t __p1)2007 __ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
2008   uint64x1_t __ret;
2009   __ret = __p0 + __p1;
2010   return __ret;
2011 }
2012 #ifdef __LITTLE_ENDIAN__
vadd_u16(uint16x4_t __p0,uint16x4_t __p1)2013 __ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
2014   uint16x4_t __ret;
2015   __ret = __p0 + __p1;
2016   return __ret;
2017 }
2018 #else
vadd_u16(uint16x4_t __p0,uint16x4_t __p1)2019 __ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
2020   uint16x4_t __ret;
2021   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2022   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2023   __ret = __rev0 + __rev1;
2024   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2025   return __ret;
2026 }
2027 #endif
2028 
2029 #ifdef __LITTLE_ENDIAN__
vadd_s8(int8x8_t __p0,int8x8_t __p1)2030 __ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
2031   int8x8_t __ret;
2032   __ret = __p0 + __p1;
2033   return __ret;
2034 }
2035 #else
vadd_s8(int8x8_t __p0,int8x8_t __p1)2036 __ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
2037   int8x8_t __ret;
2038   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2039   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2040   __ret = __rev0 + __rev1;
2041   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2042   return __ret;
2043 }
2044 #endif
2045 
2046 #ifdef __LITTLE_ENDIAN__
vadd_f32(float32x2_t __p0,float32x2_t __p1)2047 __ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
2048   float32x2_t __ret;
2049   __ret = __p0 + __p1;
2050   return __ret;
2051 }
2052 #else
vadd_f32(float32x2_t __p0,float32x2_t __p1)2053 __ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
2054   float32x2_t __ret;
2055   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2056   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2057   __ret = __rev0 + __rev1;
2058   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2059   return __ret;
2060 }
2061 #endif
2062 
2063 #ifdef __LITTLE_ENDIAN__
vadd_s32(int32x2_t __p0,int32x2_t __p1)2064 __ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
2065   int32x2_t __ret;
2066   __ret = __p0 + __p1;
2067   return __ret;
2068 }
2069 #else
vadd_s32(int32x2_t __p0,int32x2_t __p1)2070 __ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
2071   int32x2_t __ret;
2072   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2073   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2074   __ret = __rev0 + __rev1;
2075   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2076   return __ret;
2077 }
2078 #endif
2079 
vadd_s64(int64x1_t __p0,int64x1_t __p1)2080 __ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
2081   int64x1_t __ret;
2082   __ret = __p0 + __p1;
2083   return __ret;
2084 }
2085 #ifdef __LITTLE_ENDIAN__
vadd_s16(int16x4_t __p0,int16x4_t __p1)2086 __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
2087   int16x4_t __ret;
2088   __ret = __p0 + __p1;
2089   return __ret;
2090 }
2091 #else
vadd_s16(int16x4_t __p0,int16x4_t __p1)2092 __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
2093   int16x4_t __ret;
2094   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2095   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2096   __ret = __rev0 + __rev1;
2097   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2098   return __ret;
2099 }
2100 #endif
2101 
2102 #ifdef __LITTLE_ENDIAN__
vadd_p8(poly8x8_t __p0,poly8x8_t __p1)2103 __ai poly8x8_t vadd_p8(poly8x8_t __p0, poly8x8_t __p1) {
2104   poly8x8_t __ret;
2105   __ret = (poly8x8_t) __builtin_neon_vadd_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
2106   return __ret;
2107 }
2108 #else
vadd_p8(poly8x8_t __p0,poly8x8_t __p1)2109 __ai poly8x8_t vadd_p8(poly8x8_t __p0, poly8x8_t __p1) {
2110   poly8x8_t __ret;
2111   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2112   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2113   __ret = (poly8x8_t) __builtin_neon_vadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
2114   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2115   return __ret;
2116 }
2117 #endif
2118 
vadd_p64(poly64x1_t __p0,poly64x1_t __p1)2119 __ai poly64x1_t vadd_p64(poly64x1_t __p0, poly64x1_t __p1) {
2120   poly64x1_t __ret;
2121   __ret = (poly64x1_t) __builtin_neon_vadd_v((int8x8_t)__p0, (int8x8_t)__p1, 6);
2122   return __ret;
2123 }
2124 #ifdef __LITTLE_ENDIAN__
vadd_p16(poly16x4_t __p0,poly16x4_t __p1)2125 __ai poly16x4_t vadd_p16(poly16x4_t __p0, poly16x4_t __p1) {
2126   poly16x4_t __ret;
2127   __ret = (poly16x4_t) __builtin_neon_vadd_v((int8x8_t)__p0, (int8x8_t)__p1, 5);
2128   return __ret;
2129 }
2130 #else
vadd_p16(poly16x4_t __p0,poly16x4_t __p1)2131 __ai poly16x4_t vadd_p16(poly16x4_t __p0, poly16x4_t __p1) {
2132   poly16x4_t __ret;
2133   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2134   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2135   __ret = (poly16x4_t) __builtin_neon_vadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 5);
2136   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2137   return __ret;
2138 }
2139 #endif
2140 
2141 #ifdef __LITTLE_ENDIAN__
vaddq_p8(poly8x16_t __p0,poly8x16_t __p1)2142 __ai poly8x16_t vaddq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2143   poly8x16_t __ret;
2144   __ret = (poly8x16_t) __builtin_neon_vaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
2145   return __ret;
2146 }
2147 #else
vaddq_p8(poly8x16_t __p0,poly8x16_t __p1)2148 __ai poly8x16_t vaddq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2149   poly8x16_t __ret;
2150   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2151   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2152   __ret = (poly8x16_t) __builtin_neon_vaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
2153   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2154   return __ret;
2155 }
2156 #endif
2157 
2158 #ifdef __LITTLE_ENDIAN__
vaddq_p64(poly64x2_t __p0,poly64x2_t __p1)2159 __ai poly64x2_t vaddq_p64(poly64x2_t __p0, poly64x2_t __p1) {
2160   poly64x2_t __ret;
2161   __ret = (poly64x2_t) __builtin_neon_vaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 38);
2162   return __ret;
2163 }
2164 #else
vaddq_p64(poly64x2_t __p0,poly64x2_t __p1)2165 __ai poly64x2_t vaddq_p64(poly64x2_t __p0, poly64x2_t __p1) {
2166   poly64x2_t __ret;
2167   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2168   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2169   __ret = (poly64x2_t) __builtin_neon_vaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 38);
2170   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2171   return __ret;
2172 }
2173 #endif
2174 
2175 #ifdef __LITTLE_ENDIAN__
vaddq_p16(poly16x8_t __p0,poly16x8_t __p1)2176 __ai poly16x8_t vaddq_p16(poly16x8_t __p0, poly16x8_t __p1) {
2177   poly16x8_t __ret;
2178   __ret = (poly16x8_t) __builtin_neon_vaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 37);
2179   return __ret;
2180 }
2181 #else
vaddq_p16(poly16x8_t __p0,poly16x8_t __p1)2182 __ai poly16x8_t vaddq_p16(poly16x8_t __p0, poly16x8_t __p1) {
2183   poly16x8_t __ret;
2184   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2185   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2186   __ret = (poly16x8_t) __builtin_neon_vaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 37);
2187   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2188   return __ret;
2189 }
2190 #endif
2191 
2192 #ifdef __LITTLE_ENDIAN__
vaddhn_u32(uint32x4_t __p0,uint32x4_t __p1)2193 __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
2194   uint16x4_t __ret;
2195   __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
2196   return __ret;
2197 }
2198 #else
vaddhn_u32(uint32x4_t __p0,uint32x4_t __p1)2199 __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
2200   uint16x4_t __ret;
2201   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2202   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2203   __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
2204   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2205   return __ret;
2206 }
__noswap_vaddhn_u32(uint32x4_t __p0,uint32x4_t __p1)2207 __ai uint16x4_t __noswap_vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
2208   uint16x4_t __ret;
2209   __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
2210   return __ret;
2211 }
2212 #endif
2213 
2214 #ifdef __LITTLE_ENDIAN__
vaddhn_u64(uint64x2_t __p0,uint64x2_t __p1)2215 __ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
2216   uint32x2_t __ret;
2217   __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
2218   return __ret;
2219 }
2220 #else
vaddhn_u64(uint64x2_t __p0,uint64x2_t __p1)2221 __ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
2222   uint32x2_t __ret;
2223   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2224   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2225   __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
2226   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2227   return __ret;
2228 }
__noswap_vaddhn_u64(uint64x2_t __p0,uint64x2_t __p1)2229 __ai uint32x2_t __noswap_vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
2230   uint32x2_t __ret;
2231   __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
2232   return __ret;
2233 }
2234 #endif
2235 
2236 #ifdef __LITTLE_ENDIAN__
vaddhn_u16(uint16x8_t __p0,uint16x8_t __p1)2237 __ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
2238   uint8x8_t __ret;
2239   __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
2240   return __ret;
2241 }
2242 #else
vaddhn_u16(uint16x8_t __p0,uint16x8_t __p1)2243 __ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
2244   uint8x8_t __ret;
2245   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2246   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2247   __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
2248   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2249   return __ret;
2250 }
__noswap_vaddhn_u16(uint16x8_t __p0,uint16x8_t __p1)2251 __ai uint8x8_t __noswap_vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
2252   uint8x8_t __ret;
2253   __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
2254   return __ret;
2255 }
2256 #endif
2257 
2258 #ifdef __LITTLE_ENDIAN__
vaddhn_s32(int32x4_t __p0,int32x4_t __p1)2259 __ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
2260   int16x4_t __ret;
2261   __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
2262   return __ret;
2263 }
2264 #else
vaddhn_s32(int32x4_t __p0,int32x4_t __p1)2265 __ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
2266   int16x4_t __ret;
2267   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2268   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2269   __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
2270   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2271   return __ret;
2272 }
__noswap_vaddhn_s32(int32x4_t __p0,int32x4_t __p1)2273 __ai int16x4_t __noswap_vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
2274   int16x4_t __ret;
2275   __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
2276   return __ret;
2277 }
2278 #endif
2279 
2280 #ifdef __LITTLE_ENDIAN__
vaddhn_s64(int64x2_t __p0,int64x2_t __p1)2281 __ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
2282   int32x2_t __ret;
2283   __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
2284   return __ret;
2285 }
2286 #else
vaddhn_s64(int64x2_t __p0,int64x2_t __p1)2287 __ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
2288   int32x2_t __ret;
2289   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2290   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2291   __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
2292   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2293   return __ret;
2294 }
__noswap_vaddhn_s64(int64x2_t __p0,int64x2_t __p1)2295 __ai int32x2_t __noswap_vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
2296   int32x2_t __ret;
2297   __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
2298   return __ret;
2299 }
2300 #endif
2301 
2302 #ifdef __LITTLE_ENDIAN__
vaddhn_s16(int16x8_t __p0,int16x8_t __p1)2303 __ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
2304   int8x8_t __ret;
2305   __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
2306   return __ret;
2307 }
2308 #else
vaddhn_s16(int16x8_t __p0,int16x8_t __p1)2309 __ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
2310   int8x8_t __ret;
2311   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2312   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2313   __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
2314   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2315   return __ret;
2316 }
__noswap_vaddhn_s16(int16x8_t __p0,int16x8_t __p1)2317 __ai int8x8_t __noswap_vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
2318   int8x8_t __ret;
2319   __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
2320   return __ret;
2321 }
2322 #endif
2323 
2324 #ifdef __LITTLE_ENDIAN__
vandq_u8(uint8x16_t __p0,uint8x16_t __p1)2325 __ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2326   uint8x16_t __ret;
2327   __ret = __p0 & __p1;
2328   return __ret;
2329 }
2330 #else
vandq_u8(uint8x16_t __p0,uint8x16_t __p1)2331 __ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2332   uint8x16_t __ret;
2333   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2334   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2335   __ret = __rev0 & __rev1;
2336   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2337   return __ret;
2338 }
2339 #endif
2340 
2341 #ifdef __LITTLE_ENDIAN__
vandq_u32(uint32x4_t __p0,uint32x4_t __p1)2342 __ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2343   uint32x4_t __ret;
2344   __ret = __p0 & __p1;
2345   return __ret;
2346 }
2347 #else
vandq_u32(uint32x4_t __p0,uint32x4_t __p1)2348 __ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2349   uint32x4_t __ret;
2350   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2351   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2352   __ret = __rev0 & __rev1;
2353   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2354   return __ret;
2355 }
2356 #endif
2357 
2358 #ifdef __LITTLE_ENDIAN__
vandq_u64(uint64x2_t __p0,uint64x2_t __p1)2359 __ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2360   uint64x2_t __ret;
2361   __ret = __p0 & __p1;
2362   return __ret;
2363 }
2364 #else
vandq_u64(uint64x2_t __p0,uint64x2_t __p1)2365 __ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2366   uint64x2_t __ret;
2367   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2368   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2369   __ret = __rev0 & __rev1;
2370   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2371   return __ret;
2372 }
2373 #endif
2374 
2375 #ifdef __LITTLE_ENDIAN__
vandq_u16(uint16x8_t __p0,uint16x8_t __p1)2376 __ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2377   uint16x8_t __ret;
2378   __ret = __p0 & __p1;
2379   return __ret;
2380 }
2381 #else
vandq_u16(uint16x8_t __p0,uint16x8_t __p1)2382 __ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2383   uint16x8_t __ret;
2384   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2385   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2386   __ret = __rev0 & __rev1;
2387   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2388   return __ret;
2389 }
2390 #endif
2391 
2392 #ifdef __LITTLE_ENDIAN__
vandq_s8(int8x16_t __p0,int8x16_t __p1)2393 __ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
2394   int8x16_t __ret;
2395   __ret = __p0 & __p1;
2396   return __ret;
2397 }
2398 #else
vandq_s8(int8x16_t __p0,int8x16_t __p1)2399 __ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
2400   int8x16_t __ret;
2401   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2402   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2403   __ret = __rev0 & __rev1;
2404   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2405   return __ret;
2406 }
2407 #endif
2408 
2409 #ifdef __LITTLE_ENDIAN__
vandq_s32(int32x4_t __p0,int32x4_t __p1)2410 __ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
2411   int32x4_t __ret;
2412   __ret = __p0 & __p1;
2413   return __ret;
2414 }
2415 #else
vandq_s32(int32x4_t __p0,int32x4_t __p1)2416 __ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
2417   int32x4_t __ret;
2418   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2419   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2420   __ret = __rev0 & __rev1;
2421   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2422   return __ret;
2423 }
2424 #endif
2425 
2426 #ifdef __LITTLE_ENDIAN__
vandq_s64(int64x2_t __p0,int64x2_t __p1)2427 __ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
2428   int64x2_t __ret;
2429   __ret = __p0 & __p1;
2430   return __ret;
2431 }
2432 #else
vandq_s64(int64x2_t __p0,int64x2_t __p1)2433 __ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
2434   int64x2_t __ret;
2435   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2436   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2437   __ret = __rev0 & __rev1;
2438   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2439   return __ret;
2440 }
2441 #endif
2442 
2443 #ifdef __LITTLE_ENDIAN__
vandq_s16(int16x8_t __p0,int16x8_t __p1)2444 __ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
2445   int16x8_t __ret;
2446   __ret = __p0 & __p1;
2447   return __ret;
2448 }
2449 #else
vandq_s16(int16x8_t __p0,int16x8_t __p1)2450 __ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
2451   int16x8_t __ret;
2452   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2453   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2454   __ret = __rev0 & __rev1;
2455   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2456   return __ret;
2457 }
2458 #endif
2459 
2460 #ifdef __LITTLE_ENDIAN__
vand_u8(uint8x8_t __p0,uint8x8_t __p1)2461 __ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
2462   uint8x8_t __ret;
2463   __ret = __p0 & __p1;
2464   return __ret;
2465 }
2466 #else
vand_u8(uint8x8_t __p0,uint8x8_t __p1)2467 __ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
2468   uint8x8_t __ret;
2469   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2470   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2471   __ret = __rev0 & __rev1;
2472   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2473   return __ret;
2474 }
2475 #endif
2476 
2477 #ifdef __LITTLE_ENDIAN__
vand_u32(uint32x2_t __p0,uint32x2_t __p1)2478 __ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
2479   uint32x2_t __ret;
2480   __ret = __p0 & __p1;
2481   return __ret;
2482 }
2483 #else
vand_u32(uint32x2_t __p0,uint32x2_t __p1)2484 __ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
2485   uint32x2_t __ret;
2486   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2487   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2488   __ret = __rev0 & __rev1;
2489   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2490   return __ret;
2491 }
2492 #endif
2493 
vand_u64(uint64x1_t __p0,uint64x1_t __p1)2494 __ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
2495   uint64x1_t __ret;
2496   __ret = __p0 & __p1;
2497   return __ret;
2498 }
2499 #ifdef __LITTLE_ENDIAN__
vand_u16(uint16x4_t __p0,uint16x4_t __p1)2500 __ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
2501   uint16x4_t __ret;
2502   __ret = __p0 & __p1;
2503   return __ret;
2504 }
2505 #else
vand_u16(uint16x4_t __p0,uint16x4_t __p1)2506 __ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
2507   uint16x4_t __ret;
2508   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2509   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2510   __ret = __rev0 & __rev1;
2511   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2512   return __ret;
2513 }
2514 #endif
2515 
2516 #ifdef __LITTLE_ENDIAN__
vand_s8(int8x8_t __p0,int8x8_t __p1)2517 __ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
2518   int8x8_t __ret;
2519   __ret = __p0 & __p1;
2520   return __ret;
2521 }
2522 #else
vand_s8(int8x8_t __p0,int8x8_t __p1)2523 __ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
2524   int8x8_t __ret;
2525   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2526   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2527   __ret = __rev0 & __rev1;
2528   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2529   return __ret;
2530 }
2531 #endif
2532 
2533 #ifdef __LITTLE_ENDIAN__
vand_s32(int32x2_t __p0,int32x2_t __p1)2534 __ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
2535   int32x2_t __ret;
2536   __ret = __p0 & __p1;
2537   return __ret;
2538 }
2539 #else
vand_s32(int32x2_t __p0,int32x2_t __p1)2540 __ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
2541   int32x2_t __ret;
2542   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2543   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2544   __ret = __rev0 & __rev1;
2545   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2546   return __ret;
2547 }
2548 #endif
2549 
vand_s64(int64x1_t __p0,int64x1_t __p1)2550 __ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
2551   int64x1_t __ret;
2552   __ret = __p0 & __p1;
2553   return __ret;
2554 }
2555 #ifdef __LITTLE_ENDIAN__
vand_s16(int16x4_t __p0,int16x4_t __p1)2556 __ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
2557   int16x4_t __ret;
2558   __ret = __p0 & __p1;
2559   return __ret;
2560 }
2561 #else
vand_s16(int16x4_t __p0,int16x4_t __p1)2562 __ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
2563   int16x4_t __ret;
2564   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2565   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2566   __ret = __rev0 & __rev1;
2567   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2568   return __ret;
2569 }
2570 #endif
2571 
2572 #ifdef __LITTLE_ENDIAN__
vbicq_u8(uint8x16_t __p0,uint8x16_t __p1)2573 __ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2574   uint8x16_t __ret;
2575   __ret = __p0 & ~__p1;
2576   return __ret;
2577 }
2578 #else
vbicq_u8(uint8x16_t __p0,uint8x16_t __p1)2579 __ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2580   uint8x16_t __ret;
2581   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2582   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2583   __ret = __rev0 & ~__rev1;
2584   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2585   return __ret;
2586 }
2587 #endif
2588 
2589 #ifdef __LITTLE_ENDIAN__
vbicq_u32(uint32x4_t __p0,uint32x4_t __p1)2590 __ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2591   uint32x4_t __ret;
2592   __ret = __p0 & ~__p1;
2593   return __ret;
2594 }
2595 #else
vbicq_u32(uint32x4_t __p0,uint32x4_t __p1)2596 __ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2597   uint32x4_t __ret;
2598   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2599   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2600   __ret = __rev0 & ~__rev1;
2601   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2602   return __ret;
2603 }
2604 #endif
2605 
2606 #ifdef __LITTLE_ENDIAN__
vbicq_u64(uint64x2_t __p0,uint64x2_t __p1)2607 __ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2608   uint64x2_t __ret;
2609   __ret = __p0 & ~__p1;
2610   return __ret;
2611 }
2612 #else
vbicq_u64(uint64x2_t __p0,uint64x2_t __p1)2613 __ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2614   uint64x2_t __ret;
2615   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2616   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2617   __ret = __rev0 & ~__rev1;
2618   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2619   return __ret;
2620 }
2621 #endif
2622 
2623 #ifdef __LITTLE_ENDIAN__
vbicq_u16(uint16x8_t __p0,uint16x8_t __p1)2624 __ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2625   uint16x8_t __ret;
2626   __ret = __p0 & ~__p1;
2627   return __ret;
2628 }
2629 #else
vbicq_u16(uint16x8_t __p0,uint16x8_t __p1)2630 __ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2631   uint16x8_t __ret;
2632   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2633   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2634   __ret = __rev0 & ~__rev1;
2635   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2636   return __ret;
2637 }
2638 #endif
2639 
2640 #ifdef __LITTLE_ENDIAN__
vbicq_s8(int8x16_t __p0,int8x16_t __p1)2641 __ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
2642   int8x16_t __ret;
2643   __ret = __p0 & ~__p1;
2644   return __ret;
2645 }
2646 #else
vbicq_s8(int8x16_t __p0,int8x16_t __p1)2647 __ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
2648   int8x16_t __ret;
2649   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2650   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2651   __ret = __rev0 & ~__rev1;
2652   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2653   return __ret;
2654 }
2655 #endif
2656 
2657 #ifdef __LITTLE_ENDIAN__
vbicq_s32(int32x4_t __p0,int32x4_t __p1)2658 __ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
2659   int32x4_t __ret;
2660   __ret = __p0 & ~__p1;
2661   return __ret;
2662 }
2663 #else
vbicq_s32(int32x4_t __p0,int32x4_t __p1)2664 __ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
2665   int32x4_t __ret;
2666   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2667   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2668   __ret = __rev0 & ~__rev1;
2669   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2670   return __ret;
2671 }
2672 #endif
2673 
2674 #ifdef __LITTLE_ENDIAN__
vbicq_s64(int64x2_t __p0,int64x2_t __p1)2675 __ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
2676   int64x2_t __ret;
2677   __ret = __p0 & ~__p1;
2678   return __ret;
2679 }
2680 #else
vbicq_s64(int64x2_t __p0,int64x2_t __p1)2681 __ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
2682   int64x2_t __ret;
2683   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2684   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2685   __ret = __rev0 & ~__rev1;
2686   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2687   return __ret;
2688 }
2689 #endif
2690 
2691 #ifdef __LITTLE_ENDIAN__
vbicq_s16(int16x8_t __p0,int16x8_t __p1)2692 __ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
2693   int16x8_t __ret;
2694   __ret = __p0 & ~__p1;
2695   return __ret;
2696 }
2697 #else
vbicq_s16(int16x8_t __p0,int16x8_t __p1)2698 __ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
2699   int16x8_t __ret;
2700   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2701   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2702   __ret = __rev0 & ~__rev1;
2703   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2704   return __ret;
2705 }
2706 #endif
2707 
2708 #ifdef __LITTLE_ENDIAN__
vbic_u8(uint8x8_t __p0,uint8x8_t __p1)2709 __ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
2710   uint8x8_t __ret;
2711   __ret = __p0 & ~__p1;
2712   return __ret;
2713 }
2714 #else
vbic_u8(uint8x8_t __p0,uint8x8_t __p1)2715 __ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
2716   uint8x8_t __ret;
2717   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2718   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2719   __ret = __rev0 & ~__rev1;
2720   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2721   return __ret;
2722 }
2723 #endif
2724 
2725 #ifdef __LITTLE_ENDIAN__
vbic_u32(uint32x2_t __p0,uint32x2_t __p1)2726 __ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
2727   uint32x2_t __ret;
2728   __ret = __p0 & ~__p1;
2729   return __ret;
2730 }
2731 #else
vbic_u32(uint32x2_t __p0,uint32x2_t __p1)2732 __ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
2733   uint32x2_t __ret;
2734   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2735   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2736   __ret = __rev0 & ~__rev1;
2737   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2738   return __ret;
2739 }
2740 #endif
2741 
vbic_u64(uint64x1_t __p0,uint64x1_t __p1)2742 __ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
2743   uint64x1_t __ret;
2744   __ret = __p0 & ~__p1;
2745   return __ret;
2746 }
2747 #ifdef __LITTLE_ENDIAN__
vbic_u16(uint16x4_t __p0,uint16x4_t __p1)2748 __ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
2749   uint16x4_t __ret;
2750   __ret = __p0 & ~__p1;
2751   return __ret;
2752 }
2753 #else
vbic_u16(uint16x4_t __p0,uint16x4_t __p1)2754 __ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
2755   uint16x4_t __ret;
2756   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2757   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2758   __ret = __rev0 & ~__rev1;
2759   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2760   return __ret;
2761 }
2762 #endif
2763 
2764 #ifdef __LITTLE_ENDIAN__
vbic_s8(int8x8_t __p0,int8x8_t __p1)2765 __ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
2766   int8x8_t __ret;
2767   __ret = __p0 & ~__p1;
2768   return __ret;
2769 }
2770 #else
vbic_s8(int8x8_t __p0,int8x8_t __p1)2771 __ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
2772   int8x8_t __ret;
2773   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2774   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2775   __ret = __rev0 & ~__rev1;
2776   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2777   return __ret;
2778 }
2779 #endif
2780 
2781 #ifdef __LITTLE_ENDIAN__
vbic_s32(int32x2_t __p0,int32x2_t __p1)2782 __ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
2783   int32x2_t __ret;
2784   __ret = __p0 & ~__p1;
2785   return __ret;
2786 }
2787 #else
vbic_s32(int32x2_t __p0,int32x2_t __p1)2788 __ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
2789   int32x2_t __ret;
2790   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2791   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2792   __ret = __rev0 & ~__rev1;
2793   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2794   return __ret;
2795 }
2796 #endif
2797 
vbic_s64(int64x1_t __p0,int64x1_t __p1)2798 __ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
2799   int64x1_t __ret;
2800   __ret = __p0 & ~__p1;
2801   return __ret;
2802 }
2803 #ifdef __LITTLE_ENDIAN__
vbic_s16(int16x4_t __p0,int16x4_t __p1)2804 __ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
2805   int16x4_t __ret;
2806   __ret = __p0 & ~__p1;
2807   return __ret;
2808 }
2809 #else
vbic_s16(int16x4_t __p0,int16x4_t __p1)2810 __ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
2811   int16x4_t __ret;
2812   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2813   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2814   __ret = __rev0 & ~__rev1;
2815   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2816   return __ret;
2817 }
2818 #endif
2819 
2820 #ifdef __LITTLE_ENDIAN__
vbsl_p8(uint8x8_t __p0,poly8x8_t __p1,poly8x8_t __p2)2821 __ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
2822   poly8x8_t __ret;
2823   __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
2824   return __ret;
2825 }
2826 #else
vbsl_p8(uint8x8_t __p0,poly8x8_t __p1,poly8x8_t __p2)2827 __ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
2828   poly8x8_t __ret;
2829   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2830   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2831   poly8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2832   __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
2833   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2834   return __ret;
2835 }
2836 #endif
2837 
2838 #ifdef __LITTLE_ENDIAN__
vbsl_p16(uint16x4_t __p0,poly16x4_t __p1,poly16x4_t __p2)2839 __ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
2840   poly16x4_t __ret;
2841   __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 5);
2842   return __ret;
2843 }
2844 #else
vbsl_p16(uint16x4_t __p0,poly16x4_t __p1,poly16x4_t __p2)2845 __ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
2846   poly16x4_t __ret;
2847   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2848   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2849   poly16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2850   __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 5);
2851   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2852   return __ret;
2853 }
2854 #endif
2855 
2856 #ifdef __LITTLE_ENDIAN__
vbslq_p8(uint8x16_t __p0,poly8x16_t __p1,poly8x16_t __p2)2857 __ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
2858   poly8x16_t __ret;
2859   __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
2860   return __ret;
2861 }
2862 #else
vbslq_p8(uint8x16_t __p0,poly8x16_t __p1,poly8x16_t __p2)2863 __ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
2864   poly8x16_t __ret;
2865   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2866   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2867   poly8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2868   __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
2869   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2870   return __ret;
2871 }
2872 #endif
2873 
2874 #ifdef __LITTLE_ENDIAN__
vbslq_p16(uint16x8_t __p0,poly16x8_t __p1,poly16x8_t __p2)2875 __ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
2876   poly16x8_t __ret;
2877   __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 37);
2878   return __ret;
2879 }
2880 #else
vbslq_p16(uint16x8_t __p0,poly16x8_t __p1,poly16x8_t __p2)2881 __ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
2882   poly16x8_t __ret;
2883   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2884   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2885   poly16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2886   __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 37);
2887   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2888   return __ret;
2889 }
2890 #endif
2891 
2892 #ifdef __LITTLE_ENDIAN__
vbslq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)2893 __ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
2894   uint8x16_t __ret;
2895   __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
2896   return __ret;
2897 }
2898 #else
vbslq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)2899 __ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
2900   uint8x16_t __ret;
2901   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2902   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2903   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2904   __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
2905   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2906   return __ret;
2907 }
2908 #endif
2909 
2910 #ifdef __LITTLE_ENDIAN__
vbslq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)2911 __ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
2912   uint32x4_t __ret;
2913   __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
2914   return __ret;
2915 }
2916 #else
vbslq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)2917 __ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
2918   uint32x4_t __ret;
2919   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2920   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2921   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2922   __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
2923   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2924   return __ret;
2925 }
2926 #endif
2927 
2928 #ifdef __LITTLE_ENDIAN__
vbslq_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)2929 __ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
2930   uint64x2_t __ret;
2931   __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
2932   return __ret;
2933 }
2934 #else
vbslq_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)2935 __ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
2936   uint64x2_t __ret;
2937   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2938   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2939   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2940   __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
2941   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2942   return __ret;
2943 }
2944 #endif
2945 
2946 #ifdef __LITTLE_ENDIAN__
vbslq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)2947 __ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
2948   uint16x8_t __ret;
2949   __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49);
2950   return __ret;
2951 }
2952 #else
vbslq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)2953 __ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
2954   uint16x8_t __ret;
2955   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2956   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2957   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2958   __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49);
2959   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2960   return __ret;
2961 }
2962 #endif
2963 
2964 #ifdef __LITTLE_ENDIAN__
vbslq_s8(uint8x16_t __p0,int8x16_t __p1,int8x16_t __p2)2965 __ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
2966   int8x16_t __ret;
2967   __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
2968   return __ret;
2969 }
2970 #else
vbslq_s8(uint8x16_t __p0,int8x16_t __p1,int8x16_t __p2)2971 __ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
2972   int8x16_t __ret;
2973   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2974   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2975   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2976   __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
2977   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2978   return __ret;
2979 }
2980 #endif
2981 
2982 #ifdef __LITTLE_ENDIAN__
vbslq_f32(uint32x4_t __p0,float32x4_t __p1,float32x4_t __p2)2983 __ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
2984   float32x4_t __ret;
2985   __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
2986   return __ret;
2987 }
2988 #else
vbslq_f32(uint32x4_t __p0,float32x4_t __p1,float32x4_t __p2)2989 __ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
2990   float32x4_t __ret;
2991   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2992   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2993   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2994   __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
2995   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2996   return __ret;
2997 }
2998 #endif
2999 
3000 #ifdef __LITTLE_ENDIAN__
vbslq_s32(uint32x4_t __p0,int32x4_t __p1,int32x4_t __p2)3001 __ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
3002   int32x4_t __ret;
3003   __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
3004   return __ret;
3005 }
3006 #else
vbslq_s32(uint32x4_t __p0,int32x4_t __p1,int32x4_t __p2)3007 __ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
3008   int32x4_t __ret;
3009   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3010   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3011   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
3012   __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
3013   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3014   return __ret;
3015 }
3016 #endif
3017 
3018 #ifdef __LITTLE_ENDIAN__
vbslq_s64(uint64x2_t __p0,int64x2_t __p1,int64x2_t __p2)3019 __ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
3020   int64x2_t __ret;
3021   __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35);
3022   return __ret;
3023 }
3024 #else
vbslq_s64(uint64x2_t __p0,int64x2_t __p1,int64x2_t __p2)3025 __ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
3026   int64x2_t __ret;
3027   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3028   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3029   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
3030   __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35);
3031   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3032   return __ret;
3033 }
3034 #endif
3035 
3036 #ifdef __LITTLE_ENDIAN__
vbslq_s16(uint16x8_t __p0,int16x8_t __p1,int16x8_t __p2)3037 __ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
3038   int16x8_t __ret;
3039   __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
3040   return __ret;
3041 }
3042 #else
vbslq_s16(uint16x8_t __p0,int16x8_t __p1,int16x8_t __p2)3043 __ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
3044   int16x8_t __ret;
3045   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3046   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3047   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
3048   __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
3049   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3050   return __ret;
3051 }
3052 #endif
3053 
3054 #ifdef __LITTLE_ENDIAN__
vbsl_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)3055 __ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
3056   uint8x8_t __ret;
3057   __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
3058   return __ret;
3059 }
3060 #else
vbsl_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)3061 __ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
3062   uint8x8_t __ret;
3063   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3064   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3065   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
3066   __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
3067   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3068   return __ret;
3069 }
3070 #endif
3071 
3072 #ifdef __LITTLE_ENDIAN__
vbsl_u32(uint32x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)3073 __ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
3074   uint32x2_t __ret;
3075   __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
3076   return __ret;
3077 }
3078 #else
vbsl_u32(uint32x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)3079 __ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
3080   uint32x2_t __ret;
3081   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3082   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3083   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
3084   __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 18);
3085   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3086   return __ret;
3087 }
3088 #endif
3089 
vbsl_u64(uint64x1_t __p0,uint64x1_t __p1,uint64x1_t __p2)3090 __ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
3091   uint64x1_t __ret;
3092   __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
3093   return __ret;
3094 }
3095 #ifdef __LITTLE_ENDIAN__
vbsl_u16(uint16x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)3096 __ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
3097   uint16x4_t __ret;
3098   __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 17);
3099   return __ret;
3100 }
3101 #else
vbsl_u16(uint16x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)3102 __ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
3103   uint16x4_t __ret;
3104   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3105   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3106   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
3107   __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 17);
3108   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3109   return __ret;
3110 }
3111 #endif
3112 
3113 #ifdef __LITTLE_ENDIAN__
vbsl_s8(uint8x8_t __p0,int8x8_t __p1,int8x8_t __p2)3114 __ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
3115   int8x8_t __ret;
3116   __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
3117   return __ret;
3118 }
3119 #else
vbsl_s8(uint8x8_t __p0,int8x8_t __p1,int8x8_t __p2)3120 __ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
3121   int8x8_t __ret;
3122   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3123   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3124   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
3125   __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
3126   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3127   return __ret;
3128 }
3129 #endif
3130 
3131 #ifdef __LITTLE_ENDIAN__
vbsl_f32(uint32x2_t __p0,float32x2_t __p1,float32x2_t __p2)3132 __ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
3133   float32x2_t __ret;
3134   __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
3135   return __ret;
3136 }
3137 #else
vbsl_f32(uint32x2_t __p0,float32x2_t __p1,float32x2_t __p2)3138 __ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
3139   float32x2_t __ret;
3140   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3141   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3142   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
3143   __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
3144   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3145   return __ret;
3146 }
3147 #endif
3148 
3149 #ifdef __LITTLE_ENDIAN__
vbsl_s32(uint32x2_t __p0,int32x2_t __p1,int32x2_t __p2)3150 __ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
3151   int32x2_t __ret;
3152   __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
3153   return __ret;
3154 }
3155 #else
vbsl_s32(uint32x2_t __p0,int32x2_t __p1,int32x2_t __p2)3156 __ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
3157   int32x2_t __ret;
3158   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3159   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3160   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
3161   __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
3162   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3163   return __ret;
3164 }
3165 #endif
3166 
vbsl_s64(uint64x1_t __p0,int64x1_t __p1,int64x1_t __p2)3167 __ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
3168   int64x1_t __ret;
3169   __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
3170   return __ret;
3171 }
3172 #ifdef __LITTLE_ENDIAN__
vbsl_s16(uint16x4_t __p0,int16x4_t __p1,int16x4_t __p2)3173 __ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
3174   int16x4_t __ret;
3175   __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
3176   return __ret;
3177 }
3178 #else
vbsl_s16(uint16x4_t __p0,int16x4_t __p1,int16x4_t __p2)3179 __ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
3180   int16x4_t __ret;
3181   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3182   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3183   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
3184   __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 1);
3185   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3186   return __ret;
3187 }
3188 #endif
3189 
3190 #ifdef __LITTLE_ENDIAN__
vcageq_f32(float32x4_t __p0,float32x4_t __p1)3191 __ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
3192   uint32x4_t __ret;
3193   __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
3194   return __ret;
3195 }
3196 #else
vcageq_f32(float32x4_t __p0,float32x4_t __p1)3197 __ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
3198   uint32x4_t __ret;
3199   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3200   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3201   __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
3202   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3203   return __ret;
3204 }
3205 #endif
3206 
3207 #ifdef __LITTLE_ENDIAN__
vcage_f32(float32x2_t __p0,float32x2_t __p1)3208 __ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
3209   uint32x2_t __ret;
3210   __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
3211   return __ret;
3212 }
3213 #else
vcage_f32(float32x2_t __p0,float32x2_t __p1)3214 __ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
3215   uint32x2_t __ret;
3216   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3217   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3218   __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
3219   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3220   return __ret;
3221 }
3222 #endif
3223 
3224 #ifdef __LITTLE_ENDIAN__
vcagtq_f32(float32x4_t __p0,float32x4_t __p1)3225 __ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
3226   uint32x4_t __ret;
3227   __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
3228   return __ret;
3229 }
3230 #else
vcagtq_f32(float32x4_t __p0,float32x4_t __p1)3231 __ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
3232   uint32x4_t __ret;
3233   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3234   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3235   __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
3236   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3237   return __ret;
3238 }
3239 #endif
3240 
3241 #ifdef __LITTLE_ENDIAN__
vcagt_f32(float32x2_t __p0,float32x2_t __p1)3242 __ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
3243   uint32x2_t __ret;
3244   __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
3245   return __ret;
3246 }
3247 #else
vcagt_f32(float32x2_t __p0,float32x2_t __p1)3248 __ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
3249   uint32x2_t __ret;
3250   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3251   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3252   __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
3253   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3254   return __ret;
3255 }
3256 #endif
3257 
3258 #ifdef __LITTLE_ENDIAN__
vcaleq_f32(float32x4_t __p0,float32x4_t __p1)3259 __ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
3260   uint32x4_t __ret;
3261   __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
3262   return __ret;
3263 }
3264 #else
vcaleq_f32(float32x4_t __p0,float32x4_t __p1)3265 __ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
3266   uint32x4_t __ret;
3267   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3268   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3269   __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
3270   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3271   return __ret;
3272 }
3273 #endif
3274 
3275 #ifdef __LITTLE_ENDIAN__
vcale_f32(float32x2_t __p0,float32x2_t __p1)3276 __ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
3277   uint32x2_t __ret;
3278   __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
3279   return __ret;
3280 }
3281 #else
vcale_f32(float32x2_t __p0,float32x2_t __p1)3282 __ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
3283   uint32x2_t __ret;
3284   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3285   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3286   __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
3287   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3288   return __ret;
3289 }
3290 #endif
3291 
3292 #ifdef __LITTLE_ENDIAN__
vcaltq_f32(float32x4_t __p0,float32x4_t __p1)3293 __ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
3294   uint32x4_t __ret;
3295   __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
3296   return __ret;
3297 }
3298 #else
vcaltq_f32(float32x4_t __p0,float32x4_t __p1)3299 __ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
3300   uint32x4_t __ret;
3301   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3302   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3303   __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
3304   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3305   return __ret;
3306 }
3307 #endif
3308 
3309 #ifdef __LITTLE_ENDIAN__
vcalt_f32(float32x2_t __p0,float32x2_t __p1)3310 __ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
3311   uint32x2_t __ret;
3312   __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
3313   return __ret;
3314 }
3315 #else
vcalt_f32(float32x2_t __p0,float32x2_t __p1)3316 __ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
3317   uint32x2_t __ret;
3318   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3319   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3320   __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
3321   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3322   return __ret;
3323 }
3324 #endif
3325 
3326 #ifdef __LITTLE_ENDIAN__
vceq_p8(poly8x8_t __p0,poly8x8_t __p1)3327 __ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
3328   uint8x8_t __ret;
3329   __ret = (uint8x8_t)(__p0 == __p1);
3330   return __ret;
3331 }
3332 #else
vceq_p8(poly8x8_t __p0,poly8x8_t __p1)3333 __ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
3334   uint8x8_t __ret;
3335   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3336   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3337   __ret = (uint8x8_t)(__rev0 == __rev1);
3338   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3339   return __ret;
3340 }
3341 #endif
3342 
3343 #ifdef __LITTLE_ENDIAN__
vceqq_p8(poly8x16_t __p0,poly8x16_t __p1)3344 __ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
3345   uint8x16_t __ret;
3346   __ret = (uint8x16_t)(__p0 == __p1);
3347   return __ret;
3348 }
3349 #else
vceqq_p8(poly8x16_t __p0,poly8x16_t __p1)3350 __ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
3351   uint8x16_t __ret;
3352   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3353   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3354   __ret = (uint8x16_t)(__rev0 == __rev1);
3355   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3356   return __ret;
3357 }
3358 #endif
3359 
3360 #ifdef __LITTLE_ENDIAN__
vceqq_u8(uint8x16_t __p0,uint8x16_t __p1)3361 __ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3362   uint8x16_t __ret;
3363   __ret = (uint8x16_t)(__p0 == __p1);
3364   return __ret;
3365 }
3366 #else
vceqq_u8(uint8x16_t __p0,uint8x16_t __p1)3367 __ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3368   uint8x16_t __ret;
3369   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3370   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3371   __ret = (uint8x16_t)(__rev0 == __rev1);
3372   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3373   return __ret;
3374 }
3375 #endif
3376 
3377 #ifdef __LITTLE_ENDIAN__
vceqq_u32(uint32x4_t __p0,uint32x4_t __p1)3378 __ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3379   uint32x4_t __ret;
3380   __ret = (uint32x4_t)(__p0 == __p1);
3381   return __ret;
3382 }
3383 #else
vceqq_u32(uint32x4_t __p0,uint32x4_t __p1)3384 __ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3385   uint32x4_t __ret;
3386   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3387   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3388   __ret = (uint32x4_t)(__rev0 == __rev1);
3389   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3390   return __ret;
3391 }
3392 #endif
3393 
3394 #ifdef __LITTLE_ENDIAN__
vceqq_u16(uint16x8_t __p0,uint16x8_t __p1)3395 __ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3396   uint16x8_t __ret;
3397   __ret = (uint16x8_t)(__p0 == __p1);
3398   return __ret;
3399 }
3400 #else
vceqq_u16(uint16x8_t __p0,uint16x8_t __p1)3401 __ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3402   uint16x8_t __ret;
3403   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3404   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3405   __ret = (uint16x8_t)(__rev0 == __rev1);
3406   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3407   return __ret;
3408 }
3409 #endif
3410 
3411 #ifdef __LITTLE_ENDIAN__
vceqq_s8(int8x16_t __p0,int8x16_t __p1)3412 __ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
3413   uint8x16_t __ret;
3414   __ret = (uint8x16_t)(__p0 == __p1);
3415   return __ret;
3416 }
3417 #else
vceqq_s8(int8x16_t __p0,int8x16_t __p1)3418 __ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
3419   uint8x16_t __ret;
3420   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3421   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3422   __ret = (uint8x16_t)(__rev0 == __rev1);
3423   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3424   return __ret;
3425 }
3426 #endif
3427 
3428 #ifdef __LITTLE_ENDIAN__
vceqq_f32(float32x4_t __p0,float32x4_t __p1)3429 __ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
3430   uint32x4_t __ret;
3431   __ret = (uint32x4_t)(__p0 == __p1);
3432   return __ret;
3433 }
3434 #else
vceqq_f32(float32x4_t __p0,float32x4_t __p1)3435 __ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
3436   uint32x4_t __ret;
3437   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3438   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3439   __ret = (uint32x4_t)(__rev0 == __rev1);
3440   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3441   return __ret;
3442 }
3443 #endif
3444 
3445 #ifdef __LITTLE_ENDIAN__
vceqq_s32(int32x4_t __p0,int32x4_t __p1)3446 __ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
3447   uint32x4_t __ret;
3448   __ret = (uint32x4_t)(__p0 == __p1);
3449   return __ret;
3450 }
3451 #else
vceqq_s32(int32x4_t __p0,int32x4_t __p1)3452 __ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
3453   uint32x4_t __ret;
3454   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3455   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3456   __ret = (uint32x4_t)(__rev0 == __rev1);
3457   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3458   return __ret;
3459 }
3460 #endif
3461 
3462 #ifdef __LITTLE_ENDIAN__
vceqq_s16(int16x8_t __p0,int16x8_t __p1)3463 __ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
3464   uint16x8_t __ret;
3465   __ret = (uint16x8_t)(__p0 == __p1);
3466   return __ret;
3467 }
3468 #else
vceqq_s16(int16x8_t __p0,int16x8_t __p1)3469 __ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
3470   uint16x8_t __ret;
3471   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3472   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3473   __ret = (uint16x8_t)(__rev0 == __rev1);
3474   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3475   return __ret;
3476 }
3477 #endif
3478 
3479 #ifdef __LITTLE_ENDIAN__
vceq_u8(uint8x8_t __p0,uint8x8_t __p1)3480 __ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
3481   uint8x8_t __ret;
3482   __ret = (uint8x8_t)(__p0 == __p1);
3483   return __ret;
3484 }
3485 #else
vceq_u8(uint8x8_t __p0,uint8x8_t __p1)3486 __ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
3487   uint8x8_t __ret;
3488   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3489   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3490   __ret = (uint8x8_t)(__rev0 == __rev1);
3491   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3492   return __ret;
3493 }
3494 #endif
3495 
3496 #ifdef __LITTLE_ENDIAN__
vceq_u32(uint32x2_t __p0,uint32x2_t __p1)3497 __ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
3498   uint32x2_t __ret;
3499   __ret = (uint32x2_t)(__p0 == __p1);
3500   return __ret;
3501 }
3502 #else
vceq_u32(uint32x2_t __p0,uint32x2_t __p1)3503 __ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
3504   uint32x2_t __ret;
3505   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3506   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3507   __ret = (uint32x2_t)(__rev0 == __rev1);
3508   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3509   return __ret;
3510 }
3511 #endif
3512 
3513 #ifdef __LITTLE_ENDIAN__
vceq_u16(uint16x4_t __p0,uint16x4_t __p1)3514 __ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
3515   uint16x4_t __ret;
3516   __ret = (uint16x4_t)(__p0 == __p1);
3517   return __ret;
3518 }
3519 #else
vceq_u16(uint16x4_t __p0,uint16x4_t __p1)3520 __ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
3521   uint16x4_t __ret;
3522   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3523   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3524   __ret = (uint16x4_t)(__rev0 == __rev1);
3525   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3526   return __ret;
3527 }
3528 #endif
3529 
3530 #ifdef __LITTLE_ENDIAN__
vceq_s8(int8x8_t __p0,int8x8_t __p1)3531 __ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
3532   uint8x8_t __ret;
3533   __ret = (uint8x8_t)(__p0 == __p1);
3534   return __ret;
3535 }
3536 #else
vceq_s8(int8x8_t __p0,int8x8_t __p1)3537 __ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
3538   uint8x8_t __ret;
3539   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3540   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3541   __ret = (uint8x8_t)(__rev0 == __rev1);
3542   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3543   return __ret;
3544 }
3545 #endif
3546 
3547 #ifdef __LITTLE_ENDIAN__
vceq_f32(float32x2_t __p0,float32x2_t __p1)3548 __ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
3549   uint32x2_t __ret;
3550   __ret = (uint32x2_t)(__p0 == __p1);
3551   return __ret;
3552 }
3553 #else
vceq_f32(float32x2_t __p0,float32x2_t __p1)3554 __ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
3555   uint32x2_t __ret;
3556   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3557   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3558   __ret = (uint32x2_t)(__rev0 == __rev1);
3559   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3560   return __ret;
3561 }
3562 #endif
3563 
3564 #ifdef __LITTLE_ENDIAN__
vceq_s32(int32x2_t __p0,int32x2_t __p1)3565 __ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
3566   uint32x2_t __ret;
3567   __ret = (uint32x2_t)(__p0 == __p1);
3568   return __ret;
3569 }
3570 #else
vceq_s32(int32x2_t __p0,int32x2_t __p1)3571 __ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
3572   uint32x2_t __ret;
3573   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3574   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3575   __ret = (uint32x2_t)(__rev0 == __rev1);
3576   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3577   return __ret;
3578 }
3579 #endif
3580 
3581 #ifdef __LITTLE_ENDIAN__
vceq_s16(int16x4_t __p0,int16x4_t __p1)3582 __ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
3583   uint16x4_t __ret;
3584   __ret = (uint16x4_t)(__p0 == __p1);
3585   return __ret;
3586 }
3587 #else
vceq_s16(int16x4_t __p0,int16x4_t __p1)3588 __ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
3589   uint16x4_t __ret;
3590   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3591   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3592   __ret = (uint16x4_t)(__rev0 == __rev1);
3593   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3594   return __ret;
3595 }
3596 #endif
3597 
3598 #ifdef __LITTLE_ENDIAN__
vcgeq_u8(uint8x16_t __p0,uint8x16_t __p1)3599 __ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3600   uint8x16_t __ret;
3601   __ret = (uint8x16_t)(__p0 >= __p1);
3602   return __ret;
3603 }
3604 #else
vcgeq_u8(uint8x16_t __p0,uint8x16_t __p1)3605 __ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3606   uint8x16_t __ret;
3607   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3608   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3609   __ret = (uint8x16_t)(__rev0 >= __rev1);
3610   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3611   return __ret;
3612 }
3613 #endif
3614 
3615 #ifdef __LITTLE_ENDIAN__
vcgeq_u32(uint32x4_t __p0,uint32x4_t __p1)3616 __ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3617   uint32x4_t __ret;
3618   __ret = (uint32x4_t)(__p0 >= __p1);
3619   return __ret;
3620 }
3621 #else
vcgeq_u32(uint32x4_t __p0,uint32x4_t __p1)3622 __ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3623   uint32x4_t __ret;
3624   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3625   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3626   __ret = (uint32x4_t)(__rev0 >= __rev1);
3627   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3628   return __ret;
3629 }
3630 #endif
3631 
3632 #ifdef __LITTLE_ENDIAN__
vcgeq_u16(uint16x8_t __p0,uint16x8_t __p1)3633 __ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3634   uint16x8_t __ret;
3635   __ret = (uint16x8_t)(__p0 >= __p1);
3636   return __ret;
3637 }
3638 #else
vcgeq_u16(uint16x8_t __p0,uint16x8_t __p1)3639 __ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3640   uint16x8_t __ret;
3641   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3642   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3643   __ret = (uint16x8_t)(__rev0 >= __rev1);
3644   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3645   return __ret;
3646 }
3647 #endif
3648 
3649 #ifdef __LITTLE_ENDIAN__
vcgeq_s8(int8x16_t __p0,int8x16_t __p1)3650 __ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
3651   uint8x16_t __ret;
3652   __ret = (uint8x16_t)(__p0 >= __p1);
3653   return __ret;
3654 }
3655 #else
vcgeq_s8(int8x16_t __p0,int8x16_t __p1)3656 __ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
3657   uint8x16_t __ret;
3658   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3659   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3660   __ret = (uint8x16_t)(__rev0 >= __rev1);
3661   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3662   return __ret;
3663 }
3664 #endif
3665 
3666 #ifdef __LITTLE_ENDIAN__
vcgeq_f32(float32x4_t __p0,float32x4_t __p1)3667 __ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
3668   uint32x4_t __ret;
3669   __ret = (uint32x4_t)(__p0 >= __p1);
3670   return __ret;
3671 }
3672 #else
vcgeq_f32(float32x4_t __p0,float32x4_t __p1)3673 __ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
3674   uint32x4_t __ret;
3675   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3676   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3677   __ret = (uint32x4_t)(__rev0 >= __rev1);
3678   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3679   return __ret;
3680 }
3681 #endif
3682 
3683 #ifdef __LITTLE_ENDIAN__
vcgeq_s32(int32x4_t __p0,int32x4_t __p1)3684 __ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
3685   uint32x4_t __ret;
3686   __ret = (uint32x4_t)(__p0 >= __p1);
3687   return __ret;
3688 }
3689 #else
vcgeq_s32(int32x4_t __p0,int32x4_t __p1)3690 __ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
3691   uint32x4_t __ret;
3692   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3693   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3694   __ret = (uint32x4_t)(__rev0 >= __rev1);
3695   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3696   return __ret;
3697 }
3698 #endif
3699 
3700 #ifdef __LITTLE_ENDIAN__
vcgeq_s16(int16x8_t __p0,int16x8_t __p1)3701 __ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
3702   uint16x8_t __ret;
3703   __ret = (uint16x8_t)(__p0 >= __p1);
3704   return __ret;
3705 }
3706 #else
vcgeq_s16(int16x8_t __p0,int16x8_t __p1)3707 __ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
3708   uint16x8_t __ret;
3709   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3710   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3711   __ret = (uint16x8_t)(__rev0 >= __rev1);
3712   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3713   return __ret;
3714 }
3715 #endif
3716 
3717 #ifdef __LITTLE_ENDIAN__
vcge_u8(uint8x8_t __p0,uint8x8_t __p1)3718 __ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
3719   uint8x8_t __ret;
3720   __ret = (uint8x8_t)(__p0 >= __p1);
3721   return __ret;
3722 }
3723 #else
vcge_u8(uint8x8_t __p0,uint8x8_t __p1)3724 __ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
3725   uint8x8_t __ret;
3726   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3727   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3728   __ret = (uint8x8_t)(__rev0 >= __rev1);
3729   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3730   return __ret;
3731 }
3732 #endif
3733 
3734 #ifdef __LITTLE_ENDIAN__
vcge_u32(uint32x2_t __p0,uint32x2_t __p1)3735 __ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
3736   uint32x2_t __ret;
3737   __ret = (uint32x2_t)(__p0 >= __p1);
3738   return __ret;
3739 }
3740 #else
vcge_u32(uint32x2_t __p0,uint32x2_t __p1)3741 __ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
3742   uint32x2_t __ret;
3743   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3744   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3745   __ret = (uint32x2_t)(__rev0 >= __rev1);
3746   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3747   return __ret;
3748 }
3749 #endif
3750 
3751 #ifdef __LITTLE_ENDIAN__
vcge_u16(uint16x4_t __p0,uint16x4_t __p1)3752 __ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
3753   uint16x4_t __ret;
3754   __ret = (uint16x4_t)(__p0 >= __p1);
3755   return __ret;
3756 }
3757 #else
vcge_u16(uint16x4_t __p0,uint16x4_t __p1)3758 __ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
3759   uint16x4_t __ret;
3760   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3761   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3762   __ret = (uint16x4_t)(__rev0 >= __rev1);
3763   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3764   return __ret;
3765 }
3766 #endif
3767 
3768 #ifdef __LITTLE_ENDIAN__
vcge_s8(int8x8_t __p0,int8x8_t __p1)3769 __ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
3770   uint8x8_t __ret;
3771   __ret = (uint8x8_t)(__p0 >= __p1);
3772   return __ret;
3773 }
3774 #else
vcge_s8(int8x8_t __p0,int8x8_t __p1)3775 __ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
3776   uint8x8_t __ret;
3777   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3778   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3779   __ret = (uint8x8_t)(__rev0 >= __rev1);
3780   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3781   return __ret;
3782 }
3783 #endif
3784 
3785 #ifdef __LITTLE_ENDIAN__
vcge_f32(float32x2_t __p0,float32x2_t __p1)3786 __ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
3787   uint32x2_t __ret;
3788   __ret = (uint32x2_t)(__p0 >= __p1);
3789   return __ret;
3790 }
3791 #else
vcge_f32(float32x2_t __p0,float32x2_t __p1)3792 __ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
3793   uint32x2_t __ret;
3794   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3795   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3796   __ret = (uint32x2_t)(__rev0 >= __rev1);
3797   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3798   return __ret;
3799 }
3800 #endif
3801 
3802 #ifdef __LITTLE_ENDIAN__
vcge_s32(int32x2_t __p0,int32x2_t __p1)3803 __ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
3804   uint32x2_t __ret;
3805   __ret = (uint32x2_t)(__p0 >= __p1);
3806   return __ret;
3807 }
3808 #else
vcge_s32(int32x2_t __p0,int32x2_t __p1)3809 __ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
3810   uint32x2_t __ret;
3811   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3812   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3813   __ret = (uint32x2_t)(__rev0 >= __rev1);
3814   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3815   return __ret;
3816 }
3817 #endif
3818 
3819 #ifdef __LITTLE_ENDIAN__
vcge_s16(int16x4_t __p0,int16x4_t __p1)3820 __ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
3821   uint16x4_t __ret;
3822   __ret = (uint16x4_t)(__p0 >= __p1);
3823   return __ret;
3824 }
3825 #else
vcge_s16(int16x4_t __p0,int16x4_t __p1)3826 __ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
3827   uint16x4_t __ret;
3828   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3829   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3830   __ret = (uint16x4_t)(__rev0 >= __rev1);
3831   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3832   return __ret;
3833 }
3834 #endif
3835 
3836 #ifdef __LITTLE_ENDIAN__
vcgtq_u8(uint8x16_t __p0,uint8x16_t __p1)3837 __ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3838   uint8x16_t __ret;
3839   __ret = (uint8x16_t)(__p0 > __p1);
3840   return __ret;
3841 }
3842 #else
vcgtq_u8(uint8x16_t __p0,uint8x16_t __p1)3843 __ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3844   uint8x16_t __ret;
3845   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3846   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3847   __ret = (uint8x16_t)(__rev0 > __rev1);
3848   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3849   return __ret;
3850 }
3851 #endif
3852 
3853 #ifdef __LITTLE_ENDIAN__
vcgtq_u32(uint32x4_t __p0,uint32x4_t __p1)3854 __ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3855   uint32x4_t __ret;
3856   __ret = (uint32x4_t)(__p0 > __p1);
3857   return __ret;
3858 }
3859 #else
vcgtq_u32(uint32x4_t __p0,uint32x4_t __p1)3860 __ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3861   uint32x4_t __ret;
3862   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3863   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3864   __ret = (uint32x4_t)(__rev0 > __rev1);
3865   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3866   return __ret;
3867 }
3868 #endif
3869 
3870 #ifdef __LITTLE_ENDIAN__
vcgtq_u16(uint16x8_t __p0,uint16x8_t __p1)3871 __ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3872   uint16x8_t __ret;
3873   __ret = (uint16x8_t)(__p0 > __p1);
3874   return __ret;
3875 }
3876 #else
vcgtq_u16(uint16x8_t __p0,uint16x8_t __p1)3877 __ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3878   uint16x8_t __ret;
3879   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3880   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3881   __ret = (uint16x8_t)(__rev0 > __rev1);
3882   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3883   return __ret;
3884 }
3885 #endif
3886 
3887 #ifdef __LITTLE_ENDIAN__
vcgtq_s8(int8x16_t __p0,int8x16_t __p1)3888 __ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
3889   uint8x16_t __ret;
3890   __ret = (uint8x16_t)(__p0 > __p1);
3891   return __ret;
3892 }
3893 #else
vcgtq_s8(int8x16_t __p0,int8x16_t __p1)3894 __ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
3895   uint8x16_t __ret;
3896   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3897   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3898   __ret = (uint8x16_t)(__rev0 > __rev1);
3899   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3900   return __ret;
3901 }
3902 #endif
3903 
3904 #ifdef __LITTLE_ENDIAN__
vcgtq_f32(float32x4_t __p0,float32x4_t __p1)3905 __ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
3906   uint32x4_t __ret;
3907   __ret = (uint32x4_t)(__p0 > __p1);
3908   return __ret;
3909 }
3910 #else
vcgtq_f32(float32x4_t __p0,float32x4_t __p1)3911 __ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
3912   uint32x4_t __ret;
3913   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3914   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3915   __ret = (uint32x4_t)(__rev0 > __rev1);
3916   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3917   return __ret;
3918 }
3919 #endif
3920 
3921 #ifdef __LITTLE_ENDIAN__
vcgtq_s32(int32x4_t __p0,int32x4_t __p1)3922 __ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
3923   uint32x4_t __ret;
3924   __ret = (uint32x4_t)(__p0 > __p1);
3925   return __ret;
3926 }
3927 #else
vcgtq_s32(int32x4_t __p0,int32x4_t __p1)3928 __ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
3929   uint32x4_t __ret;
3930   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3931   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3932   __ret = (uint32x4_t)(__rev0 > __rev1);
3933   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3934   return __ret;
3935 }
3936 #endif
3937 
3938 #ifdef __LITTLE_ENDIAN__
vcgtq_s16(int16x8_t __p0,int16x8_t __p1)3939 __ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
3940   uint16x8_t __ret;
3941   __ret = (uint16x8_t)(__p0 > __p1);
3942   return __ret;
3943 }
3944 #else
vcgtq_s16(int16x8_t __p0,int16x8_t __p1)3945 __ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
3946   uint16x8_t __ret;
3947   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3948   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3949   __ret = (uint16x8_t)(__rev0 > __rev1);
3950   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3951   return __ret;
3952 }
3953 #endif
3954 
3955 #ifdef __LITTLE_ENDIAN__
vcgt_u8(uint8x8_t __p0,uint8x8_t __p1)3956 __ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3957   uint8x8_t __ret;
3958   __ret = (uint8x8_t)(__p0 > __p1);
3959   return __ret;
3960 }
3961 #else
vcgt_u8(uint8x8_t __p0,uint8x8_t __p1)3962 __ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3963   uint8x8_t __ret;
3964   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3965   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3966   __ret = (uint8x8_t)(__rev0 > __rev1);
3967   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3968   return __ret;
3969 }
3970 #endif
3971 
3972 #ifdef __LITTLE_ENDIAN__
vcgt_u32(uint32x2_t __p0,uint32x2_t __p1)3973 __ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3974   uint32x2_t __ret;
3975   __ret = (uint32x2_t)(__p0 > __p1);
3976   return __ret;
3977 }
3978 #else
vcgt_u32(uint32x2_t __p0,uint32x2_t __p1)3979 __ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3980   uint32x2_t __ret;
3981   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3982   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3983   __ret = (uint32x2_t)(__rev0 > __rev1);
3984   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3985   return __ret;
3986 }
3987 #endif
3988 
3989 #ifdef __LITTLE_ENDIAN__
vcgt_u16(uint16x4_t __p0,uint16x4_t __p1)3990 __ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3991   uint16x4_t __ret;
3992   __ret = (uint16x4_t)(__p0 > __p1);
3993   return __ret;
3994 }
3995 #else
vcgt_u16(uint16x4_t __p0,uint16x4_t __p1)3996 __ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3997   uint16x4_t __ret;
3998   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3999   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4000   __ret = (uint16x4_t)(__rev0 > __rev1);
4001   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4002   return __ret;
4003 }
4004 #endif
4005 
4006 #ifdef __LITTLE_ENDIAN__
vcgt_s8(int8x8_t __p0,int8x8_t __p1)4007 __ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
4008   uint8x8_t __ret;
4009   __ret = (uint8x8_t)(__p0 > __p1);
4010   return __ret;
4011 }
4012 #else
vcgt_s8(int8x8_t __p0,int8x8_t __p1)4013 __ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
4014   uint8x8_t __ret;
4015   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4016   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4017   __ret = (uint8x8_t)(__rev0 > __rev1);
4018   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4019   return __ret;
4020 }
4021 #endif
4022 
4023 #ifdef __LITTLE_ENDIAN__
vcgt_f32(float32x2_t __p0,float32x2_t __p1)4024 __ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
4025   uint32x2_t __ret;
4026   __ret = (uint32x2_t)(__p0 > __p1);
4027   return __ret;
4028 }
4029 #else
vcgt_f32(float32x2_t __p0,float32x2_t __p1)4030 __ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
4031   uint32x2_t __ret;
4032   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4033   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4034   __ret = (uint32x2_t)(__rev0 > __rev1);
4035   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4036   return __ret;
4037 }
4038 #endif
4039 
4040 #ifdef __LITTLE_ENDIAN__
vcgt_s32(int32x2_t __p0,int32x2_t __p1)4041 __ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
4042   uint32x2_t __ret;
4043   __ret = (uint32x2_t)(__p0 > __p1);
4044   return __ret;
4045 }
4046 #else
vcgt_s32(int32x2_t __p0,int32x2_t __p1)4047 __ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
4048   uint32x2_t __ret;
4049   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4050   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4051   __ret = (uint32x2_t)(__rev0 > __rev1);
4052   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4053   return __ret;
4054 }
4055 #endif
4056 
4057 #ifdef __LITTLE_ENDIAN__
vcgt_s16(int16x4_t __p0,int16x4_t __p1)4058 __ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
4059   uint16x4_t __ret;
4060   __ret = (uint16x4_t)(__p0 > __p1);
4061   return __ret;
4062 }
4063 #else
vcgt_s16(int16x4_t __p0,int16x4_t __p1)4064 __ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
4065   uint16x4_t __ret;
4066   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4067   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4068   __ret = (uint16x4_t)(__rev0 > __rev1);
4069   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4070   return __ret;
4071 }
4072 #endif
4073 
4074 #ifdef __LITTLE_ENDIAN__
vcleq_u8(uint8x16_t __p0,uint8x16_t __p1)4075 __ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4076   uint8x16_t __ret;
4077   __ret = (uint8x16_t)(__p0 <= __p1);
4078   return __ret;
4079 }
4080 #else
vcleq_u8(uint8x16_t __p0,uint8x16_t __p1)4081 __ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4082   uint8x16_t __ret;
4083   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4084   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4085   __ret = (uint8x16_t)(__rev0 <= __rev1);
4086   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4087   return __ret;
4088 }
4089 #endif
4090 
4091 #ifdef __LITTLE_ENDIAN__
vcleq_u32(uint32x4_t __p0,uint32x4_t __p1)4092 __ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4093   uint32x4_t __ret;
4094   __ret = (uint32x4_t)(__p0 <= __p1);
4095   return __ret;
4096 }
4097 #else
vcleq_u32(uint32x4_t __p0,uint32x4_t __p1)4098 __ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4099   uint32x4_t __ret;
4100   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4101   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4102   __ret = (uint32x4_t)(__rev0 <= __rev1);
4103   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4104   return __ret;
4105 }
4106 #endif
4107 
4108 #ifdef __LITTLE_ENDIAN__
vcleq_u16(uint16x8_t __p0,uint16x8_t __p1)4109 __ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4110   uint16x8_t __ret;
4111   __ret = (uint16x8_t)(__p0 <= __p1);
4112   return __ret;
4113 }
4114 #else
vcleq_u16(uint16x8_t __p0,uint16x8_t __p1)4115 __ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4116   uint16x8_t __ret;
4117   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4118   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4119   __ret = (uint16x8_t)(__rev0 <= __rev1);
4120   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4121   return __ret;
4122 }
4123 #endif
4124 
4125 #ifdef __LITTLE_ENDIAN__
vcleq_s8(int8x16_t __p0,int8x16_t __p1)4126 __ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
4127   uint8x16_t __ret;
4128   __ret = (uint8x16_t)(__p0 <= __p1);
4129   return __ret;
4130 }
4131 #else
vcleq_s8(int8x16_t __p0,int8x16_t __p1)4132 __ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
4133   uint8x16_t __ret;
4134   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4135   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4136   __ret = (uint8x16_t)(__rev0 <= __rev1);
4137   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4138   return __ret;
4139 }
4140 #endif
4141 
4142 #ifdef __LITTLE_ENDIAN__
vcleq_f32(float32x4_t __p0,float32x4_t __p1)4143 __ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
4144   uint32x4_t __ret;
4145   __ret = (uint32x4_t)(__p0 <= __p1);
4146   return __ret;
4147 }
4148 #else
vcleq_f32(float32x4_t __p0,float32x4_t __p1)4149 __ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
4150   uint32x4_t __ret;
4151   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4152   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4153   __ret = (uint32x4_t)(__rev0 <= __rev1);
4154   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4155   return __ret;
4156 }
4157 #endif
4158 
4159 #ifdef __LITTLE_ENDIAN__
vcleq_s32(int32x4_t __p0,int32x4_t __p1)4160 __ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
4161   uint32x4_t __ret;
4162   __ret = (uint32x4_t)(__p0 <= __p1);
4163   return __ret;
4164 }
4165 #else
vcleq_s32(int32x4_t __p0,int32x4_t __p1)4166 __ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
4167   uint32x4_t __ret;
4168   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4169   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4170   __ret = (uint32x4_t)(__rev0 <= __rev1);
4171   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4172   return __ret;
4173 }
4174 #endif
4175 
4176 #ifdef __LITTLE_ENDIAN__
vcleq_s16(int16x8_t __p0,int16x8_t __p1)4177 __ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
4178   uint16x8_t __ret;
4179   __ret = (uint16x8_t)(__p0 <= __p1);
4180   return __ret;
4181 }
4182 #else
vcleq_s16(int16x8_t __p0,int16x8_t __p1)4183 __ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
4184   uint16x8_t __ret;
4185   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4186   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4187   __ret = (uint16x8_t)(__rev0 <= __rev1);
4188   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4189   return __ret;
4190 }
4191 #endif
4192 
4193 #ifdef __LITTLE_ENDIAN__
vcle_u8(uint8x8_t __p0,uint8x8_t __p1)4194 __ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
4195   uint8x8_t __ret;
4196   __ret = (uint8x8_t)(__p0 <= __p1);
4197   return __ret;
4198 }
4199 #else
vcle_u8(uint8x8_t __p0,uint8x8_t __p1)4200 __ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
4201   uint8x8_t __ret;
4202   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4203   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4204   __ret = (uint8x8_t)(__rev0 <= __rev1);
4205   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4206   return __ret;
4207 }
4208 #endif
4209 
4210 #ifdef __LITTLE_ENDIAN__
vcle_u32(uint32x2_t __p0,uint32x2_t __p1)4211 __ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
4212   uint32x2_t __ret;
4213   __ret = (uint32x2_t)(__p0 <= __p1);
4214   return __ret;
4215 }
4216 #else
vcle_u32(uint32x2_t __p0,uint32x2_t __p1)4217 __ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
4218   uint32x2_t __ret;
4219   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4220   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4221   __ret = (uint32x2_t)(__rev0 <= __rev1);
4222   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4223   return __ret;
4224 }
4225 #endif
4226 
4227 #ifdef __LITTLE_ENDIAN__
vcle_u16(uint16x4_t __p0,uint16x4_t __p1)4228 __ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
4229   uint16x4_t __ret;
4230   __ret = (uint16x4_t)(__p0 <= __p1);
4231   return __ret;
4232 }
4233 #else
vcle_u16(uint16x4_t __p0,uint16x4_t __p1)4234 __ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
4235   uint16x4_t __ret;
4236   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4237   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4238   __ret = (uint16x4_t)(__rev0 <= __rev1);
4239   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4240   return __ret;
4241 }
4242 #endif
4243 
4244 #ifdef __LITTLE_ENDIAN__
vcle_s8(int8x8_t __p0,int8x8_t __p1)4245 __ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
4246   uint8x8_t __ret;
4247   __ret = (uint8x8_t)(__p0 <= __p1);
4248   return __ret;
4249 }
4250 #else
vcle_s8(int8x8_t __p0,int8x8_t __p1)4251 __ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
4252   uint8x8_t __ret;
4253   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4254   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4255   __ret = (uint8x8_t)(__rev0 <= __rev1);
4256   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4257   return __ret;
4258 }
4259 #endif
4260 
4261 #ifdef __LITTLE_ENDIAN__
vcle_f32(float32x2_t __p0,float32x2_t __p1)4262 __ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
4263   uint32x2_t __ret;
4264   __ret = (uint32x2_t)(__p0 <= __p1);
4265   return __ret;
4266 }
4267 #else
vcle_f32(float32x2_t __p0,float32x2_t __p1)4268 __ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
4269   uint32x2_t __ret;
4270   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4271   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4272   __ret = (uint32x2_t)(__rev0 <= __rev1);
4273   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4274   return __ret;
4275 }
4276 #endif
4277 
4278 #ifdef __LITTLE_ENDIAN__
vcle_s32(int32x2_t __p0,int32x2_t __p1)4279 __ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
4280   uint32x2_t __ret;
4281   __ret = (uint32x2_t)(__p0 <= __p1);
4282   return __ret;
4283 }
4284 #else
vcle_s32(int32x2_t __p0,int32x2_t __p1)4285 __ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
4286   uint32x2_t __ret;
4287   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4288   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4289   __ret = (uint32x2_t)(__rev0 <= __rev1);
4290   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4291   return __ret;
4292 }
4293 #endif
4294 
4295 #ifdef __LITTLE_ENDIAN__
vcle_s16(int16x4_t __p0,int16x4_t __p1)4296 __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
4297   uint16x4_t __ret;
4298   __ret = (uint16x4_t)(__p0 <= __p1);
4299   return __ret;
4300 }
4301 #else
vcle_s16(int16x4_t __p0,int16x4_t __p1)4302 __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
4303   uint16x4_t __ret;
4304   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4305   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4306   __ret = (uint16x4_t)(__rev0 <= __rev1);
4307   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4308   return __ret;
4309 }
4310 #endif
4311 
4312 #ifdef __LITTLE_ENDIAN__
vclsq_u8(uint8x16_t __p0)4313 __ai int8x16_t vclsq_u8(uint8x16_t __p0) {
4314   int8x16_t __ret;
4315   __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32);
4316   return __ret;
4317 }
4318 #else
vclsq_u8(uint8x16_t __p0)4319 __ai int8x16_t vclsq_u8(uint8x16_t __p0) {
4320   int8x16_t __ret;
4321   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4322   __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32);
4323   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4324   return __ret;
4325 }
4326 #endif
4327 
4328 #ifdef __LITTLE_ENDIAN__
vclsq_u32(uint32x4_t __p0)4329 __ai int32x4_t vclsq_u32(uint32x4_t __p0) {
4330   int32x4_t __ret;
4331   __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34);
4332   return __ret;
4333 }
4334 #else
vclsq_u32(uint32x4_t __p0)4335 __ai int32x4_t vclsq_u32(uint32x4_t __p0) {
4336   int32x4_t __ret;
4337   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4338   __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34);
4339   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4340   return __ret;
4341 }
4342 #endif
4343 
4344 #ifdef __LITTLE_ENDIAN__
vclsq_u16(uint16x8_t __p0)4345 __ai int16x8_t vclsq_u16(uint16x8_t __p0) {
4346   int16x8_t __ret;
4347   __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33);
4348   return __ret;
4349 }
4350 #else
vclsq_u16(uint16x8_t __p0)4351 __ai int16x8_t vclsq_u16(uint16x8_t __p0) {
4352   int16x8_t __ret;
4353   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4354   __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33);
4355   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4356   return __ret;
4357 }
4358 #endif
4359 
4360 #ifdef __LITTLE_ENDIAN__
vclsq_s8(int8x16_t __p0)4361 __ai int8x16_t vclsq_s8(int8x16_t __p0) {
4362   int8x16_t __ret;
4363   __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32);
4364   return __ret;
4365 }
4366 #else
vclsq_s8(int8x16_t __p0)4367 __ai int8x16_t vclsq_s8(int8x16_t __p0) {
4368   int8x16_t __ret;
4369   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4370   __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32);
4371   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4372   return __ret;
4373 }
4374 #endif
4375 
4376 #ifdef __LITTLE_ENDIAN__
vclsq_s32(int32x4_t __p0)4377 __ai int32x4_t vclsq_s32(int32x4_t __p0) {
4378   int32x4_t __ret;
4379   __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34);
4380   return __ret;
4381 }
4382 #else
vclsq_s32(int32x4_t __p0)4383 __ai int32x4_t vclsq_s32(int32x4_t __p0) {
4384   int32x4_t __ret;
4385   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4386   __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34);
4387   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4388   return __ret;
4389 }
4390 #endif
4391 
4392 #ifdef __LITTLE_ENDIAN__
vclsq_s16(int16x8_t __p0)4393 __ai int16x8_t vclsq_s16(int16x8_t __p0) {
4394   int16x8_t __ret;
4395   __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33);
4396   return __ret;
4397 }
4398 #else
vclsq_s16(int16x8_t __p0)4399 __ai int16x8_t vclsq_s16(int16x8_t __p0) {
4400   int16x8_t __ret;
4401   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4402   __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33);
4403   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4404   return __ret;
4405 }
4406 #endif
4407 
4408 #ifdef __LITTLE_ENDIAN__
vcls_u8(uint8x8_t __p0)4409 __ai int8x8_t vcls_u8(uint8x8_t __p0) {
4410   int8x8_t __ret;
4411   __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0);
4412   return __ret;
4413 }
4414 #else
vcls_u8(uint8x8_t __p0)4415 __ai int8x8_t vcls_u8(uint8x8_t __p0) {
4416   int8x8_t __ret;
4417   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4418   __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0);
4419   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4420   return __ret;
4421 }
4422 #endif
4423 
4424 #ifdef __LITTLE_ENDIAN__
vcls_u32(uint32x2_t __p0)4425 __ai int32x2_t vcls_u32(uint32x2_t __p0) {
4426   int32x2_t __ret;
4427   __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2);
4428   return __ret;
4429 }
4430 #else
vcls_u32(uint32x2_t __p0)4431 __ai int32x2_t vcls_u32(uint32x2_t __p0) {
4432   int32x2_t __ret;
4433   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4434   __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2);
4435   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4436   return __ret;
4437 }
4438 #endif
4439 
4440 #ifdef __LITTLE_ENDIAN__
vcls_u16(uint16x4_t __p0)4441 __ai int16x4_t vcls_u16(uint16x4_t __p0) {
4442   int16x4_t __ret;
4443   __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1);
4444   return __ret;
4445 }
4446 #else
vcls_u16(uint16x4_t __p0)4447 __ai int16x4_t vcls_u16(uint16x4_t __p0) {
4448   int16x4_t __ret;
4449   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4450   __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1);
4451   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4452   return __ret;
4453 }
4454 #endif
4455 
4456 #ifdef __LITTLE_ENDIAN__
vcls_s8(int8x8_t __p0)4457 __ai int8x8_t vcls_s8(int8x8_t __p0) {
4458   int8x8_t __ret;
4459   __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0);
4460   return __ret;
4461 }
4462 #else
vcls_s8(int8x8_t __p0)4463 __ai int8x8_t vcls_s8(int8x8_t __p0) {
4464   int8x8_t __ret;
4465   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4466   __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0);
4467   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4468   return __ret;
4469 }
4470 #endif
4471 
4472 #ifdef __LITTLE_ENDIAN__
vcls_s32(int32x2_t __p0)4473 __ai int32x2_t vcls_s32(int32x2_t __p0) {
4474   int32x2_t __ret;
4475   __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2);
4476   return __ret;
4477 }
4478 #else
vcls_s32(int32x2_t __p0)4479 __ai int32x2_t vcls_s32(int32x2_t __p0) {
4480   int32x2_t __ret;
4481   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4482   __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2);
4483   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4484   return __ret;
4485 }
4486 #endif
4487 
4488 #ifdef __LITTLE_ENDIAN__
vcls_s16(int16x4_t __p0)4489 __ai int16x4_t vcls_s16(int16x4_t __p0) {
4490   int16x4_t __ret;
4491   __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1);
4492   return __ret;
4493 }
4494 #else
vcls_s16(int16x4_t __p0)4495 __ai int16x4_t vcls_s16(int16x4_t __p0) {
4496   int16x4_t __ret;
4497   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4498   __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1);
4499   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4500   return __ret;
4501 }
4502 #endif
4503 
4504 #ifdef __LITTLE_ENDIAN__
vcltq_u8(uint8x16_t __p0,uint8x16_t __p1)4505 __ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4506   uint8x16_t __ret;
4507   __ret = (uint8x16_t)(__p0 < __p1);
4508   return __ret;
4509 }
4510 #else
vcltq_u8(uint8x16_t __p0,uint8x16_t __p1)4511 __ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4512   uint8x16_t __ret;
4513   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4514   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4515   __ret = (uint8x16_t)(__rev0 < __rev1);
4516   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4517   return __ret;
4518 }
4519 #endif
4520 
4521 #ifdef __LITTLE_ENDIAN__
vcltq_u32(uint32x4_t __p0,uint32x4_t __p1)4522 __ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4523   uint32x4_t __ret;
4524   __ret = (uint32x4_t)(__p0 < __p1);
4525   return __ret;
4526 }
4527 #else
vcltq_u32(uint32x4_t __p0,uint32x4_t __p1)4528 __ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4529   uint32x4_t __ret;
4530   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4531   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4532   __ret = (uint32x4_t)(__rev0 < __rev1);
4533   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4534   return __ret;
4535 }
4536 #endif
4537 
4538 #ifdef __LITTLE_ENDIAN__
vcltq_u16(uint16x8_t __p0,uint16x8_t __p1)4539 __ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4540   uint16x8_t __ret;
4541   __ret = (uint16x8_t)(__p0 < __p1);
4542   return __ret;
4543 }
4544 #else
vcltq_u16(uint16x8_t __p0,uint16x8_t __p1)4545 __ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4546   uint16x8_t __ret;
4547   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4548   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4549   __ret = (uint16x8_t)(__rev0 < __rev1);
4550   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4551   return __ret;
4552 }
4553 #endif
4554 
4555 #ifdef __LITTLE_ENDIAN__
vcltq_s8(int8x16_t __p0,int8x16_t __p1)4556 __ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
4557   uint8x16_t __ret;
4558   __ret = (uint8x16_t)(__p0 < __p1);
4559   return __ret;
4560 }
4561 #else
vcltq_s8(int8x16_t __p0,int8x16_t __p1)4562 __ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
4563   uint8x16_t __ret;
4564   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4565   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4566   __ret = (uint8x16_t)(__rev0 < __rev1);
4567   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4568   return __ret;
4569 }
4570 #endif
4571 
4572 #ifdef __LITTLE_ENDIAN__
vcltq_f32(float32x4_t __p0,float32x4_t __p1)4573 __ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
4574   uint32x4_t __ret;
4575   __ret = (uint32x4_t)(__p0 < __p1);
4576   return __ret;
4577 }
4578 #else
vcltq_f32(float32x4_t __p0,float32x4_t __p1)4579 __ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
4580   uint32x4_t __ret;
4581   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4582   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4583   __ret = (uint32x4_t)(__rev0 < __rev1);
4584   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4585   return __ret;
4586 }
4587 #endif
4588 
4589 #ifdef __LITTLE_ENDIAN__
vcltq_s32(int32x4_t __p0,int32x4_t __p1)4590 __ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
4591   uint32x4_t __ret;
4592   __ret = (uint32x4_t)(__p0 < __p1);
4593   return __ret;
4594 }
4595 #else
vcltq_s32(int32x4_t __p0,int32x4_t __p1)4596 __ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
4597   uint32x4_t __ret;
4598   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4599   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4600   __ret = (uint32x4_t)(__rev0 < __rev1);
4601   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4602   return __ret;
4603 }
4604 #endif
4605 
4606 #ifdef __LITTLE_ENDIAN__
vcltq_s16(int16x8_t __p0,int16x8_t __p1)4607 __ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
4608   uint16x8_t __ret;
4609   __ret = (uint16x8_t)(__p0 < __p1);
4610   return __ret;
4611 }
4612 #else
vcltq_s16(int16x8_t __p0,int16x8_t __p1)4613 __ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
4614   uint16x8_t __ret;
4615   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4616   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4617   __ret = (uint16x8_t)(__rev0 < __rev1);
4618   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4619   return __ret;
4620 }
4621 #endif
4622 
4623 #ifdef __LITTLE_ENDIAN__
vclt_u8(uint8x8_t __p0,uint8x8_t __p1)4624 __ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
4625   uint8x8_t __ret;
4626   __ret = (uint8x8_t)(__p0 < __p1);
4627   return __ret;
4628 }
4629 #else
vclt_u8(uint8x8_t __p0,uint8x8_t __p1)4630 __ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
4631   uint8x8_t __ret;
4632   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4633   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4634   __ret = (uint8x8_t)(__rev0 < __rev1);
4635   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4636   return __ret;
4637 }
4638 #endif
4639 
4640 #ifdef __LITTLE_ENDIAN__
vclt_u32(uint32x2_t __p0,uint32x2_t __p1)4641 __ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
4642   uint32x2_t __ret;
4643   __ret = (uint32x2_t)(__p0 < __p1);
4644   return __ret;
4645 }
4646 #else
vclt_u32(uint32x2_t __p0,uint32x2_t __p1)4647 __ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
4648   uint32x2_t __ret;
4649   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4650   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4651   __ret = (uint32x2_t)(__rev0 < __rev1);
4652   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4653   return __ret;
4654 }
4655 #endif
4656 
4657 #ifdef __LITTLE_ENDIAN__
vclt_u16(uint16x4_t __p0,uint16x4_t __p1)4658 __ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
4659   uint16x4_t __ret;
4660   __ret = (uint16x4_t)(__p0 < __p1);
4661   return __ret;
4662 }
4663 #else
vclt_u16(uint16x4_t __p0,uint16x4_t __p1)4664 __ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
4665   uint16x4_t __ret;
4666   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4667   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4668   __ret = (uint16x4_t)(__rev0 < __rev1);
4669   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4670   return __ret;
4671 }
4672 #endif
4673 
4674 #ifdef __LITTLE_ENDIAN__
vclt_s8(int8x8_t __p0,int8x8_t __p1)4675 __ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
4676   uint8x8_t __ret;
4677   __ret = (uint8x8_t)(__p0 < __p1);
4678   return __ret;
4679 }
4680 #else
vclt_s8(int8x8_t __p0,int8x8_t __p1)4681 __ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
4682   uint8x8_t __ret;
4683   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4684   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4685   __ret = (uint8x8_t)(__rev0 < __rev1);
4686   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4687   return __ret;
4688 }
4689 #endif
4690 
4691 #ifdef __LITTLE_ENDIAN__
vclt_f32(float32x2_t __p0,float32x2_t __p1)4692 __ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
4693   uint32x2_t __ret;
4694   __ret = (uint32x2_t)(__p0 < __p1);
4695   return __ret;
4696 }
4697 #else
vclt_f32(float32x2_t __p0,float32x2_t __p1)4698 __ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
4699   uint32x2_t __ret;
4700   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4701   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4702   __ret = (uint32x2_t)(__rev0 < __rev1);
4703   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4704   return __ret;
4705 }
4706 #endif
4707 
4708 #ifdef __LITTLE_ENDIAN__
vclt_s32(int32x2_t __p0,int32x2_t __p1)4709 __ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
4710   uint32x2_t __ret;
4711   __ret = (uint32x2_t)(__p0 < __p1);
4712   return __ret;
4713 }
4714 #else
vclt_s32(int32x2_t __p0,int32x2_t __p1)4715 __ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
4716   uint32x2_t __ret;
4717   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4718   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4719   __ret = (uint32x2_t)(__rev0 < __rev1);
4720   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4721   return __ret;
4722 }
4723 #endif
4724 
4725 #ifdef __LITTLE_ENDIAN__
vclt_s16(int16x4_t __p0,int16x4_t __p1)4726 __ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
4727   uint16x4_t __ret;
4728   __ret = (uint16x4_t)(__p0 < __p1);
4729   return __ret;
4730 }
4731 #else
vclt_s16(int16x4_t __p0,int16x4_t __p1)4732 __ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
4733   uint16x4_t __ret;
4734   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4735   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4736   __ret = (uint16x4_t)(__rev0 < __rev1);
4737   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4738   return __ret;
4739 }
4740 #endif
4741 
4742 #ifdef __LITTLE_ENDIAN__
vclzq_u8(uint8x16_t __p0)4743 __ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
4744   uint8x16_t __ret;
4745   __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 48);
4746   return __ret;
4747 }
4748 #else
vclzq_u8(uint8x16_t __p0)4749 __ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
4750   uint8x16_t __ret;
4751   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4752   __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 48);
4753   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4754   return __ret;
4755 }
4756 #endif
4757 
4758 #ifdef __LITTLE_ENDIAN__
vclzq_u32(uint32x4_t __p0)4759 __ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
4760   uint32x4_t __ret;
4761   __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 50);
4762   return __ret;
4763 }
4764 #else
vclzq_u32(uint32x4_t __p0)4765 __ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
4766   uint32x4_t __ret;
4767   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4768   __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 50);
4769   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4770   return __ret;
4771 }
4772 #endif
4773 
4774 #ifdef __LITTLE_ENDIAN__
vclzq_u16(uint16x8_t __p0)4775 __ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
4776   uint16x8_t __ret;
4777   __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 49);
4778   return __ret;
4779 }
4780 #else
vclzq_u16(uint16x8_t __p0)4781 __ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
4782   uint16x8_t __ret;
4783   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4784   __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 49);
4785   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4786   return __ret;
4787 }
4788 #endif
4789 
4790 #ifdef __LITTLE_ENDIAN__
vclzq_s8(int8x16_t __p0)4791 __ai int8x16_t vclzq_s8(int8x16_t __p0) {
4792   int8x16_t __ret;
4793   __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 32);
4794   return __ret;
4795 }
4796 #else
vclzq_s8(int8x16_t __p0)4797 __ai int8x16_t vclzq_s8(int8x16_t __p0) {
4798   int8x16_t __ret;
4799   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4800   __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 32);
4801   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4802   return __ret;
4803 }
4804 #endif
4805 
4806 #ifdef __LITTLE_ENDIAN__
vclzq_s32(int32x4_t __p0)4807 __ai int32x4_t vclzq_s32(int32x4_t __p0) {
4808   int32x4_t __ret;
4809   __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 34);
4810   return __ret;
4811 }
4812 #else
vclzq_s32(int32x4_t __p0)4813 __ai int32x4_t vclzq_s32(int32x4_t __p0) {
4814   int32x4_t __ret;
4815   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4816   __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 34);
4817   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4818   return __ret;
4819 }
4820 #endif
4821 
4822 #ifdef __LITTLE_ENDIAN__
vclzq_s16(int16x8_t __p0)4823 __ai int16x8_t vclzq_s16(int16x8_t __p0) {
4824   int16x8_t __ret;
4825   __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 33);
4826   return __ret;
4827 }
4828 #else
vclzq_s16(int16x8_t __p0)4829 __ai int16x8_t vclzq_s16(int16x8_t __p0) {
4830   int16x8_t __ret;
4831   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4832   __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 33);
4833   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4834   return __ret;
4835 }
4836 #endif
4837 
4838 #ifdef __LITTLE_ENDIAN__
vclz_u8(uint8x8_t __p0)4839 __ai uint8x8_t vclz_u8(uint8x8_t __p0) {
4840   uint8x8_t __ret;
4841   __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 16);
4842   return __ret;
4843 }
4844 #else
vclz_u8(uint8x8_t __p0)4845 __ai uint8x8_t vclz_u8(uint8x8_t __p0) {
4846   uint8x8_t __ret;
4847   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4848   __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 16);
4849   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4850   return __ret;
4851 }
4852 #endif
4853 
4854 #ifdef __LITTLE_ENDIAN__
vclz_u32(uint32x2_t __p0)4855 __ai uint32x2_t vclz_u32(uint32x2_t __p0) {
4856   uint32x2_t __ret;
4857   __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 18);
4858   return __ret;
4859 }
4860 #else
vclz_u32(uint32x2_t __p0)4861 __ai uint32x2_t vclz_u32(uint32x2_t __p0) {
4862   uint32x2_t __ret;
4863   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4864   __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 18);
4865   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4866   return __ret;
4867 }
4868 #endif
4869 
4870 #ifdef __LITTLE_ENDIAN__
vclz_u16(uint16x4_t __p0)4871 __ai uint16x4_t vclz_u16(uint16x4_t __p0) {
4872   uint16x4_t __ret;
4873   __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 17);
4874   return __ret;
4875 }
4876 #else
vclz_u16(uint16x4_t __p0)4877 __ai uint16x4_t vclz_u16(uint16x4_t __p0) {
4878   uint16x4_t __ret;
4879   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4880   __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 17);
4881   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4882   return __ret;
4883 }
4884 #endif
4885 
4886 #ifdef __LITTLE_ENDIAN__
vclz_s8(int8x8_t __p0)4887 __ai int8x8_t vclz_s8(int8x8_t __p0) {
4888   int8x8_t __ret;
4889   __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 0);
4890   return __ret;
4891 }
4892 #else
vclz_s8(int8x8_t __p0)4893 __ai int8x8_t vclz_s8(int8x8_t __p0) {
4894   int8x8_t __ret;
4895   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4896   __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 0);
4897   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4898   return __ret;
4899 }
4900 #endif
4901 
4902 #ifdef __LITTLE_ENDIAN__
vclz_s32(int32x2_t __p0)4903 __ai int32x2_t vclz_s32(int32x2_t __p0) {
4904   int32x2_t __ret;
4905   __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 2);
4906   return __ret;
4907 }
4908 #else
vclz_s32(int32x2_t __p0)4909 __ai int32x2_t vclz_s32(int32x2_t __p0) {
4910   int32x2_t __ret;
4911   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4912   __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 2);
4913   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4914   return __ret;
4915 }
4916 #endif
4917 
4918 #ifdef __LITTLE_ENDIAN__
vclz_s16(int16x4_t __p0)4919 __ai int16x4_t vclz_s16(int16x4_t __p0) {
4920   int16x4_t __ret;
4921   __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 1);
4922   return __ret;
4923 }
4924 #else
vclz_s16(int16x4_t __p0)4925 __ai int16x4_t vclz_s16(int16x4_t __p0) {
4926   int16x4_t __ret;
4927   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4928   __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 1);
4929   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4930   return __ret;
4931 }
4932 #endif
4933 
4934 #ifdef __LITTLE_ENDIAN__
vcnt_p8(poly8x8_t __p0)4935 __ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
4936   poly8x8_t __ret;
4937   __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 4);
4938   return __ret;
4939 }
4940 #else
vcnt_p8(poly8x8_t __p0)4941 __ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
4942   poly8x8_t __ret;
4943   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4944   __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 4);
4945   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4946   return __ret;
4947 }
4948 #endif
4949 
4950 #ifdef __LITTLE_ENDIAN__
vcntq_p8(poly8x16_t __p0)4951 __ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
4952   poly8x16_t __ret;
4953   __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 36);
4954   return __ret;
4955 }
4956 #else
vcntq_p8(poly8x16_t __p0)4957 __ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
4958   poly8x16_t __ret;
4959   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4960   __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 36);
4961   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4962   return __ret;
4963 }
4964 #endif
4965 
4966 #ifdef __LITTLE_ENDIAN__
vcntq_u8(uint8x16_t __p0)4967 __ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
4968   uint8x16_t __ret;
4969   __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 48);
4970   return __ret;
4971 }
4972 #else
vcntq_u8(uint8x16_t __p0)4973 __ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
4974   uint8x16_t __ret;
4975   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4976   __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 48);
4977   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4978   return __ret;
4979 }
4980 #endif
4981 
4982 #ifdef __LITTLE_ENDIAN__
vcntq_s8(int8x16_t __p0)4983 __ai int8x16_t vcntq_s8(int8x16_t __p0) {
4984   int8x16_t __ret;
4985   __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 32);
4986   return __ret;
4987 }
4988 #else
vcntq_s8(int8x16_t __p0)4989 __ai int8x16_t vcntq_s8(int8x16_t __p0) {
4990   int8x16_t __ret;
4991   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4992   __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 32);
4993   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4994   return __ret;
4995 }
4996 #endif
4997 
4998 #ifdef __LITTLE_ENDIAN__
vcnt_u8(uint8x8_t __p0)4999 __ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
5000   uint8x8_t __ret;
5001   __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 16);
5002   return __ret;
5003 }
5004 #else
vcnt_u8(uint8x8_t __p0)5005 __ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
5006   uint8x8_t __ret;
5007   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5008   __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 16);
5009   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5010   return __ret;
5011 }
5012 #endif
5013 
5014 #ifdef __LITTLE_ENDIAN__
vcnt_s8(int8x8_t __p0)5015 __ai int8x8_t vcnt_s8(int8x8_t __p0) {
5016   int8x8_t __ret;
5017   __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 0);
5018   return __ret;
5019 }
5020 #else
vcnt_s8(int8x8_t __p0)5021 __ai int8x8_t vcnt_s8(int8x8_t __p0) {
5022   int8x8_t __ret;
5023   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5024   __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 0);
5025   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5026   return __ret;
5027 }
5028 #endif
5029 
5030 #ifdef __LITTLE_ENDIAN__
vcombine_p8(poly8x8_t __p0,poly8x8_t __p1)5031 __ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
5032   poly8x16_t __ret;
5033   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5034   return __ret;
5035 }
5036 #else
vcombine_p8(poly8x8_t __p0,poly8x8_t __p1)5037 __ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
5038   poly8x16_t __ret;
5039   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5040   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5041   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5042   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5043   return __ret;
5044 }
5045 #endif
5046 
5047 #ifdef __LITTLE_ENDIAN__
vcombine_p16(poly16x4_t __p0,poly16x4_t __p1)5048 __ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
5049   poly16x8_t __ret;
5050   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5051   return __ret;
5052 }
5053 #else
vcombine_p16(poly16x4_t __p0,poly16x4_t __p1)5054 __ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
5055   poly16x8_t __ret;
5056   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5057   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5058   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
5059   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5060   return __ret;
5061 }
5062 #endif
5063 
5064 #ifdef __LITTLE_ENDIAN__
vcombine_u8(uint8x8_t __p0,uint8x8_t __p1)5065 __ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
5066   uint8x16_t __ret;
5067   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5068   return __ret;
5069 }
5070 #else
vcombine_u8(uint8x8_t __p0,uint8x8_t __p1)5071 __ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
5072   uint8x16_t __ret;
5073   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5074   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5075   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5076   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5077   return __ret;
5078 }
__noswap_vcombine_u8(uint8x8_t __p0,uint8x8_t __p1)5079 __ai uint8x16_t __noswap_vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
5080   uint8x16_t __ret;
5081   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5082   return __ret;
5083 }
5084 #endif
5085 
5086 #ifdef __LITTLE_ENDIAN__
vcombine_u32(uint32x2_t __p0,uint32x2_t __p1)5087 __ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
5088   uint32x4_t __ret;
5089   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5090   return __ret;
5091 }
5092 #else
vcombine_u32(uint32x2_t __p0,uint32x2_t __p1)5093 __ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
5094   uint32x4_t __ret;
5095   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5096   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5097   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
5098   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5099   return __ret;
5100 }
__noswap_vcombine_u32(uint32x2_t __p0,uint32x2_t __p1)5101 __ai uint32x4_t __noswap_vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
5102   uint32x4_t __ret;
5103   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5104   return __ret;
5105 }
5106 #endif
5107 
5108 #ifdef __LITTLE_ENDIAN__
vcombine_u64(uint64x1_t __p0,uint64x1_t __p1)5109 __ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
5110   uint64x2_t __ret;
5111   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
5112   return __ret;
5113 }
5114 #else
vcombine_u64(uint64x1_t __p0,uint64x1_t __p1)5115 __ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
5116   uint64x2_t __ret;
5117   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
5118   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5119   return __ret;
5120 }
5121 #endif
5122 
5123 #ifdef __LITTLE_ENDIAN__
vcombine_u16(uint16x4_t __p0,uint16x4_t __p1)5124 __ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
5125   uint16x8_t __ret;
5126   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5127   return __ret;
5128 }
5129 #else
vcombine_u16(uint16x4_t __p0,uint16x4_t __p1)5130 __ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
5131   uint16x8_t __ret;
5132   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5133   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5134   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
5135   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5136   return __ret;
5137 }
__noswap_vcombine_u16(uint16x4_t __p0,uint16x4_t __p1)5138 __ai uint16x8_t __noswap_vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
5139   uint16x8_t __ret;
5140   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5141   return __ret;
5142 }
5143 #endif
5144 
5145 #ifdef __LITTLE_ENDIAN__
vcombine_s8(int8x8_t __p0,int8x8_t __p1)5146 __ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
5147   int8x16_t __ret;
5148   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5149   return __ret;
5150 }
5151 #else
vcombine_s8(int8x8_t __p0,int8x8_t __p1)5152 __ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
5153   int8x16_t __ret;
5154   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5155   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5156   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5157   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5158   return __ret;
5159 }
__noswap_vcombine_s8(int8x8_t __p0,int8x8_t __p1)5160 __ai int8x16_t __noswap_vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
5161   int8x16_t __ret;
5162   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5163   return __ret;
5164 }
5165 #endif
5166 
5167 #ifdef __LITTLE_ENDIAN__
vcombine_f32(float32x2_t __p0,float32x2_t __p1)5168 __ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
5169   float32x4_t __ret;
5170   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5171   return __ret;
5172 }
5173 #else
vcombine_f32(float32x2_t __p0,float32x2_t __p1)5174 __ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
5175   float32x4_t __ret;
5176   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5177   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5178   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
5179   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5180   return __ret;
5181 }
__noswap_vcombine_f32(float32x2_t __p0,float32x2_t __p1)5182 __ai float32x4_t __noswap_vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
5183   float32x4_t __ret;
5184   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5185   return __ret;
5186 }
5187 #endif
5188 
5189 #ifdef __LITTLE_ENDIAN__
vcombine_f16(float16x4_t __p0,float16x4_t __p1)5190 __ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
5191   float16x8_t __ret;
5192   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5193   return __ret;
5194 }
5195 #else
vcombine_f16(float16x4_t __p0,float16x4_t __p1)5196 __ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
5197   float16x8_t __ret;
5198   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5199   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5200   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
5201   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5202   return __ret;
5203 }
__noswap_vcombine_f16(float16x4_t __p0,float16x4_t __p1)5204 __ai float16x8_t __noswap_vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
5205   float16x8_t __ret;
5206   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5207   return __ret;
5208 }
5209 #endif
5210 
5211 #ifdef __LITTLE_ENDIAN__
vcombine_s32(int32x2_t __p0,int32x2_t __p1)5212 __ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
5213   int32x4_t __ret;
5214   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5215   return __ret;
5216 }
5217 #else
vcombine_s32(int32x2_t __p0,int32x2_t __p1)5218 __ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
5219   int32x4_t __ret;
5220   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5221   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5222   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
5223   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5224   return __ret;
5225 }
__noswap_vcombine_s32(int32x2_t __p0,int32x2_t __p1)5226 __ai int32x4_t __noswap_vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
5227   int32x4_t __ret;
5228   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5229   return __ret;
5230 }
5231 #endif
5232 
5233 #ifdef __LITTLE_ENDIAN__
vcombine_s64(int64x1_t __p0,int64x1_t __p1)5234 __ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
5235   int64x2_t __ret;
5236   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
5237   return __ret;
5238 }
5239 #else
vcombine_s64(int64x1_t __p0,int64x1_t __p1)5240 __ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
5241   int64x2_t __ret;
5242   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
5243   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5244   return __ret;
5245 }
5246 #endif
5247 
5248 #ifdef __LITTLE_ENDIAN__
vcombine_s16(int16x4_t __p0,int16x4_t __p1)5249 __ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
5250   int16x8_t __ret;
5251   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5252   return __ret;
5253 }
5254 #else
vcombine_s16(int16x4_t __p0,int16x4_t __p1)5255 __ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
5256   int16x8_t __ret;
5257   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5258   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5259   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
5260   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5261   return __ret;
5262 }
__noswap_vcombine_s16(int16x4_t __p0,int16x4_t __p1)5263 __ai int16x8_t __noswap_vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
5264   int16x8_t __ret;
5265   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5266   return __ret;
5267 }
5268 #endif
5269 
5270 #define vcreate_p8(__p0) __extension__ ({ \
5271   poly8x8_t __ret; \
5272   uint64_t __promote = __p0; \
5273   __ret = (poly8x8_t)(__promote); \
5274   __ret; \
5275 })
5276 #define vcreate_p16(__p0) __extension__ ({ \
5277   poly16x4_t __ret; \
5278   uint64_t __promote = __p0; \
5279   __ret = (poly16x4_t)(__promote); \
5280   __ret; \
5281 })
5282 #define vcreate_u8(__p0) __extension__ ({ \
5283   uint8x8_t __ret; \
5284   uint64_t __promote = __p0; \
5285   __ret = (uint8x8_t)(__promote); \
5286   __ret; \
5287 })
5288 #define vcreate_u32(__p0) __extension__ ({ \
5289   uint32x2_t __ret; \
5290   uint64_t __promote = __p0; \
5291   __ret = (uint32x2_t)(__promote); \
5292   __ret; \
5293 })
5294 #define vcreate_u64(__p0) __extension__ ({ \
5295   uint64x1_t __ret; \
5296   uint64_t __promote = __p0; \
5297   __ret = (uint64x1_t)(__promote); \
5298   __ret; \
5299 })
5300 #define vcreate_u16(__p0) __extension__ ({ \
5301   uint16x4_t __ret; \
5302   uint64_t __promote = __p0; \
5303   __ret = (uint16x4_t)(__promote); \
5304   __ret; \
5305 })
5306 #define vcreate_s8(__p0) __extension__ ({ \
5307   int8x8_t __ret; \
5308   uint64_t __promote = __p0; \
5309   __ret = (int8x8_t)(__promote); \
5310   __ret; \
5311 })
5312 #define vcreate_f32(__p0) __extension__ ({ \
5313   float32x2_t __ret; \
5314   uint64_t __promote = __p0; \
5315   __ret = (float32x2_t)(__promote); \
5316   __ret; \
5317 })
5318 #define vcreate_f16(__p0) __extension__ ({ \
5319   float16x4_t __ret; \
5320   uint64_t __promote = __p0; \
5321   __ret = (float16x4_t)(__promote); \
5322   __ret; \
5323 })
5324 #define vcreate_s32(__p0) __extension__ ({ \
5325   int32x2_t __ret; \
5326   uint64_t __promote = __p0; \
5327   __ret = (int32x2_t)(__promote); \
5328   __ret; \
5329 })
5330 #define vcreate_s64(__p0) __extension__ ({ \
5331   int64x1_t __ret; \
5332   uint64_t __promote = __p0; \
5333   __ret = (int64x1_t)(__promote); \
5334   __ret; \
5335 })
5336 #define vcreate_s16(__p0) __extension__ ({ \
5337   int16x4_t __ret; \
5338   uint64_t __promote = __p0; \
5339   __ret = (int16x4_t)(__promote); \
5340   __ret; \
5341 })
5342 #ifdef __LITTLE_ENDIAN__
vcvtq_f32_u32(uint32x4_t __p0)5343 __ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
5344   float32x4_t __ret;
5345   __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 50);
5346   return __ret;
5347 }
5348 #else
vcvtq_f32_u32(uint32x4_t __p0)5349 __ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
5350   float32x4_t __ret;
5351   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5352   __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 50);
5353   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5354   return __ret;
5355 }
5356 #endif
5357 
5358 #ifdef __LITTLE_ENDIAN__
vcvtq_f32_s32(int32x4_t __p0)5359 __ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
5360   float32x4_t __ret;
5361   __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 34);
5362   return __ret;
5363 }
5364 #else
vcvtq_f32_s32(int32x4_t __p0)5365 __ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
5366   float32x4_t __ret;
5367   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5368   __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 34);
5369   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5370   return __ret;
5371 }
5372 #endif
5373 
5374 #ifdef __LITTLE_ENDIAN__
vcvt_f32_u32(uint32x2_t __p0)5375 __ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
5376   float32x2_t __ret;
5377   __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 18);
5378   return __ret;
5379 }
5380 #else
vcvt_f32_u32(uint32x2_t __p0)5381 __ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
5382   float32x2_t __ret;
5383   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5384   __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 18);
5385   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5386   return __ret;
5387 }
5388 #endif
5389 
5390 #ifdef __LITTLE_ENDIAN__
vcvt_f32_s32(int32x2_t __p0)5391 __ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
5392   float32x2_t __ret;
5393   __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 2);
5394   return __ret;
5395 }
5396 #else
vcvt_f32_s32(int32x2_t __p0)5397 __ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
5398   float32x2_t __ret;
5399   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5400   __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 2);
5401   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5402   return __ret;
5403 }
5404 #endif
5405 
5406 #ifdef __LITTLE_ENDIAN__
5407 #define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
5408   float32x4_t __ret; \
5409   uint32x4_t __s0 = __p0; \
5410   __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 50); \
5411   __ret; \
5412 })
5413 #else
5414 #define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
5415   float32x4_t __ret; \
5416   uint32x4_t __s0 = __p0; \
5417   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5418   __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 50); \
5419   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5420   __ret; \
5421 })
5422 #endif
5423 
5424 #ifdef __LITTLE_ENDIAN__
5425 #define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
5426   float32x4_t __ret; \
5427   int32x4_t __s0 = __p0; \
5428   __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 34); \
5429   __ret; \
5430 })
5431 #else
5432 #define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
5433   float32x4_t __ret; \
5434   int32x4_t __s0 = __p0; \
5435   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5436   __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 34); \
5437   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5438   __ret; \
5439 })
5440 #endif
5441 
5442 #ifdef __LITTLE_ENDIAN__
5443 #define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
5444   float32x2_t __ret; \
5445   uint32x2_t __s0 = __p0; \
5446   __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 18); \
5447   __ret; \
5448 })
5449 #else
5450 #define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
5451   float32x2_t __ret; \
5452   uint32x2_t __s0 = __p0; \
5453   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5454   __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 18); \
5455   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5456   __ret; \
5457 })
5458 #endif
5459 
5460 #ifdef __LITTLE_ENDIAN__
5461 #define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
5462   float32x2_t __ret; \
5463   int32x2_t __s0 = __p0; \
5464   __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 2); \
5465   __ret; \
5466 })
5467 #else
5468 #define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
5469   float32x2_t __ret; \
5470   int32x2_t __s0 = __p0; \
5471   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5472   __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 2); \
5473   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5474   __ret; \
5475 })
5476 #endif
5477 
5478 #ifdef __LITTLE_ENDIAN__
5479 #define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
5480   int32x4_t __ret; \
5481   float32x4_t __s0 = __p0; \
5482   __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__s0, __p1, 34); \
5483   __ret; \
5484 })
5485 #else
5486 #define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
5487   int32x4_t __ret; \
5488   float32x4_t __s0 = __p0; \
5489   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5490   __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__rev0, __p1, 34); \
5491   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5492   __ret; \
5493 })
5494 #endif
5495 
5496 #ifdef __LITTLE_ENDIAN__
5497 #define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
5498   int32x2_t __ret; \
5499   float32x2_t __s0 = __p0; \
5500   __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__s0, __p1, 2); \
5501   __ret; \
5502 })
5503 #else
5504 #define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
5505   int32x2_t __ret; \
5506   float32x2_t __s0 = __p0; \
5507   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5508   __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__rev0, __p1, 2); \
5509   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5510   __ret; \
5511 })
5512 #endif
5513 
5514 #ifdef __LITTLE_ENDIAN__
5515 #define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
5516   uint32x4_t __ret; \
5517   float32x4_t __s0 = __p0; \
5518   __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__s0, __p1, 50); \
5519   __ret; \
5520 })
5521 #else
5522 #define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
5523   uint32x4_t __ret; \
5524   float32x4_t __s0 = __p0; \
5525   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5526   __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__rev0, __p1, 50); \
5527   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5528   __ret; \
5529 })
5530 #endif
5531 
5532 #ifdef __LITTLE_ENDIAN__
5533 #define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
5534   uint32x2_t __ret; \
5535   float32x2_t __s0 = __p0; \
5536   __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__s0, __p1, 18); \
5537   __ret; \
5538 })
5539 #else
5540 #define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
5541   uint32x2_t __ret; \
5542   float32x2_t __s0 = __p0; \
5543   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5544   __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__rev0, __p1, 18); \
5545   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5546   __ret; \
5547 })
5548 #endif
5549 
5550 #ifdef __LITTLE_ENDIAN__
vcvtq_s32_f32(float32x4_t __p0)5551 __ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
5552   int32x4_t __ret;
5553   __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__p0, 34);
5554   return __ret;
5555 }
5556 #else
vcvtq_s32_f32(float32x4_t __p0)5557 __ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
5558   int32x4_t __ret;
5559   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5560   __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__rev0, 34);
5561   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5562   return __ret;
5563 }
5564 #endif
5565 
5566 #ifdef __LITTLE_ENDIAN__
vcvt_s32_f32(float32x2_t __p0)5567 __ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
5568   int32x2_t __ret;
5569   __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__p0, 2);
5570   return __ret;
5571 }
5572 #else
vcvt_s32_f32(float32x2_t __p0)5573 __ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
5574   int32x2_t __ret;
5575   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5576   __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__rev0, 2);
5577   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5578   return __ret;
5579 }
5580 #endif
5581 
5582 #ifdef __LITTLE_ENDIAN__
vcvtq_u32_f32(float32x4_t __p0)5583 __ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
5584   uint32x4_t __ret;
5585   __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__p0, 50);
5586   return __ret;
5587 }
5588 #else
vcvtq_u32_f32(float32x4_t __p0)5589 __ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
5590   uint32x4_t __ret;
5591   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5592   __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__rev0, 50);
5593   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5594   return __ret;
5595 }
5596 #endif
5597 
5598 #ifdef __LITTLE_ENDIAN__
vcvt_u32_f32(float32x2_t __p0)5599 __ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
5600   uint32x2_t __ret;
5601   __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__p0, 18);
5602   return __ret;
5603 }
5604 #else
vcvt_u32_f32(float32x2_t __p0)5605 __ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
5606   uint32x2_t __ret;
5607   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5608   __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__rev0, 18);
5609   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5610   return __ret;
5611 }
5612 #endif
5613 
5614 #ifdef __LITTLE_ENDIAN__
5615 #define vdup_lane_p8(__p0_0, __p1_0) __extension__ ({ \
5616   poly8x8_t __ret_0; \
5617   poly8x8_t __s0_0 = __p0_0; \
5618   __ret_0 = splat_lane_p8(__s0_0, __p1_0); \
5619   __ret_0; \
5620 })
5621 #else
5622 #define vdup_lane_p8(__p0_1, __p1_1) __extension__ ({ \
5623   poly8x8_t __ret_1; \
5624   poly8x8_t __s0_1 = __p0_1; \
5625   poly8x8_t __rev0_1;  __rev0_1 = __builtin_shufflevector(__s0_1, __s0_1, 7, 6, 5, 4, 3, 2, 1, 0); \
5626   __ret_1 = __noswap_splat_lane_p8(__rev0_1, __p1_1); \
5627   __ret_1 = __builtin_shufflevector(__ret_1, __ret_1, 7, 6, 5, 4, 3, 2, 1, 0); \
5628   __ret_1; \
5629 })
5630 #endif
5631 
5632 #ifdef __LITTLE_ENDIAN__
5633 #define vdup_lane_p16(__p0_2, __p1_2) __extension__ ({ \
5634   poly16x4_t __ret_2; \
5635   poly16x4_t __s0_2 = __p0_2; \
5636   __ret_2 = splat_lane_p16(__s0_2, __p1_2); \
5637   __ret_2; \
5638 })
5639 #else
5640 #define vdup_lane_p16(__p0_3, __p1_3) __extension__ ({ \
5641   poly16x4_t __ret_3; \
5642   poly16x4_t __s0_3 = __p0_3; \
5643   poly16x4_t __rev0_3;  __rev0_3 = __builtin_shufflevector(__s0_3, __s0_3, 3, 2, 1, 0); \
5644   __ret_3 = __noswap_splat_lane_p16(__rev0_3, __p1_3); \
5645   __ret_3 = __builtin_shufflevector(__ret_3, __ret_3, 3, 2, 1, 0); \
5646   __ret_3; \
5647 })
5648 #endif
5649 
5650 #ifdef __LITTLE_ENDIAN__
5651 #define vdupq_lane_p8(__p0_4, __p1_4) __extension__ ({ \
5652   poly8x16_t __ret_4; \
5653   poly8x8_t __s0_4 = __p0_4; \
5654   __ret_4 = splatq_lane_p8(__s0_4, __p1_4); \
5655   __ret_4; \
5656 })
5657 #else
5658 #define vdupq_lane_p8(__p0_5, __p1_5) __extension__ ({ \
5659   poly8x16_t __ret_5; \
5660   poly8x8_t __s0_5 = __p0_5; \
5661   poly8x8_t __rev0_5;  __rev0_5 = __builtin_shufflevector(__s0_5, __s0_5, 7, 6, 5, 4, 3, 2, 1, 0); \
5662   __ret_5 = __noswap_splatq_lane_p8(__rev0_5, __p1_5); \
5663   __ret_5 = __builtin_shufflevector(__ret_5, __ret_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5664   __ret_5; \
5665 })
5666 #endif
5667 
5668 #ifdef __LITTLE_ENDIAN__
5669 #define vdupq_lane_p16(__p0_6, __p1_6) __extension__ ({ \
5670   poly16x8_t __ret_6; \
5671   poly16x4_t __s0_6 = __p0_6; \
5672   __ret_6 = splatq_lane_p16(__s0_6, __p1_6); \
5673   __ret_6; \
5674 })
5675 #else
5676 #define vdupq_lane_p16(__p0_7, __p1_7) __extension__ ({ \
5677   poly16x8_t __ret_7; \
5678   poly16x4_t __s0_7 = __p0_7; \
5679   poly16x4_t __rev0_7;  __rev0_7 = __builtin_shufflevector(__s0_7, __s0_7, 3, 2, 1, 0); \
5680   __ret_7 = __noswap_splatq_lane_p16(__rev0_7, __p1_7); \
5681   __ret_7 = __builtin_shufflevector(__ret_7, __ret_7, 7, 6, 5, 4, 3, 2, 1, 0); \
5682   __ret_7; \
5683 })
5684 #endif
5685 
5686 #ifdef __LITTLE_ENDIAN__
5687 #define vdupq_lane_u8(__p0_8, __p1_8) __extension__ ({ \
5688   uint8x16_t __ret_8; \
5689   uint8x8_t __s0_8 = __p0_8; \
5690   __ret_8 = splatq_lane_u8(__s0_8, __p1_8); \
5691   __ret_8; \
5692 })
5693 #else
5694 #define vdupq_lane_u8(__p0_9, __p1_9) __extension__ ({ \
5695   uint8x16_t __ret_9; \
5696   uint8x8_t __s0_9 = __p0_9; \
5697   uint8x8_t __rev0_9;  __rev0_9 = __builtin_shufflevector(__s0_9, __s0_9, 7, 6, 5, 4, 3, 2, 1, 0); \
5698   __ret_9 = __noswap_splatq_lane_u8(__rev0_9, __p1_9); \
5699   __ret_9 = __builtin_shufflevector(__ret_9, __ret_9, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5700   __ret_9; \
5701 })
5702 #endif
5703 
5704 #ifdef __LITTLE_ENDIAN__
5705 #define vdupq_lane_u32(__p0_10, __p1_10) __extension__ ({ \
5706   uint32x4_t __ret_10; \
5707   uint32x2_t __s0_10 = __p0_10; \
5708   __ret_10 = splatq_lane_u32(__s0_10, __p1_10); \
5709   __ret_10; \
5710 })
5711 #else
5712 #define vdupq_lane_u32(__p0_11, __p1_11) __extension__ ({ \
5713   uint32x4_t __ret_11; \
5714   uint32x2_t __s0_11 = __p0_11; \
5715   uint32x2_t __rev0_11;  __rev0_11 = __builtin_shufflevector(__s0_11, __s0_11, 1, 0); \
5716   __ret_11 = __noswap_splatq_lane_u32(__rev0_11, __p1_11); \
5717   __ret_11 = __builtin_shufflevector(__ret_11, __ret_11, 3, 2, 1, 0); \
5718   __ret_11; \
5719 })
5720 #endif
5721 
5722 #ifdef __LITTLE_ENDIAN__
5723 #define vdupq_lane_u64(__p0_12, __p1_12) __extension__ ({ \
5724   uint64x2_t __ret_12; \
5725   uint64x1_t __s0_12 = __p0_12; \
5726   __ret_12 = splatq_lane_u64(__s0_12, __p1_12); \
5727   __ret_12; \
5728 })
5729 #else
5730 #define vdupq_lane_u64(__p0_13, __p1_13) __extension__ ({ \
5731   uint64x2_t __ret_13; \
5732   uint64x1_t __s0_13 = __p0_13; \
5733   __ret_13 = __noswap_splatq_lane_u64(__s0_13, __p1_13); \
5734   __ret_13 = __builtin_shufflevector(__ret_13, __ret_13, 1, 0); \
5735   __ret_13; \
5736 })
5737 #endif
5738 
5739 #ifdef __LITTLE_ENDIAN__
5740 #define vdupq_lane_u16(__p0_14, __p1_14) __extension__ ({ \
5741   uint16x8_t __ret_14; \
5742   uint16x4_t __s0_14 = __p0_14; \
5743   __ret_14 = splatq_lane_u16(__s0_14, __p1_14); \
5744   __ret_14; \
5745 })
5746 #else
5747 #define vdupq_lane_u16(__p0_15, __p1_15) __extension__ ({ \
5748   uint16x8_t __ret_15; \
5749   uint16x4_t __s0_15 = __p0_15; \
5750   uint16x4_t __rev0_15;  __rev0_15 = __builtin_shufflevector(__s0_15, __s0_15, 3, 2, 1, 0); \
5751   __ret_15 = __noswap_splatq_lane_u16(__rev0_15, __p1_15); \
5752   __ret_15 = __builtin_shufflevector(__ret_15, __ret_15, 7, 6, 5, 4, 3, 2, 1, 0); \
5753   __ret_15; \
5754 })
5755 #endif
5756 
5757 #ifdef __LITTLE_ENDIAN__
5758 #define vdupq_lane_s8(__p0_16, __p1_16) __extension__ ({ \
5759   int8x16_t __ret_16; \
5760   int8x8_t __s0_16 = __p0_16; \
5761   __ret_16 = splatq_lane_s8(__s0_16, __p1_16); \
5762   __ret_16; \
5763 })
5764 #else
5765 #define vdupq_lane_s8(__p0_17, __p1_17) __extension__ ({ \
5766   int8x16_t __ret_17; \
5767   int8x8_t __s0_17 = __p0_17; \
5768   int8x8_t __rev0_17;  __rev0_17 = __builtin_shufflevector(__s0_17, __s0_17, 7, 6, 5, 4, 3, 2, 1, 0); \
5769   __ret_17 = __noswap_splatq_lane_s8(__rev0_17, __p1_17); \
5770   __ret_17 = __builtin_shufflevector(__ret_17, __ret_17, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5771   __ret_17; \
5772 })
5773 #endif
5774 
5775 #ifdef __LITTLE_ENDIAN__
5776 #define vdupq_lane_f32(__p0_18, __p1_18) __extension__ ({ \
5777   float32x4_t __ret_18; \
5778   float32x2_t __s0_18 = __p0_18; \
5779   __ret_18 = splatq_lane_f32(__s0_18, __p1_18); \
5780   __ret_18; \
5781 })
5782 #else
5783 #define vdupq_lane_f32(__p0_19, __p1_19) __extension__ ({ \
5784   float32x4_t __ret_19; \
5785   float32x2_t __s0_19 = __p0_19; \
5786   float32x2_t __rev0_19;  __rev0_19 = __builtin_shufflevector(__s0_19, __s0_19, 1, 0); \
5787   __ret_19 = __noswap_splatq_lane_f32(__rev0_19, __p1_19); \
5788   __ret_19 = __builtin_shufflevector(__ret_19, __ret_19, 3, 2, 1, 0); \
5789   __ret_19; \
5790 })
5791 #endif
5792 
5793 #ifdef __LITTLE_ENDIAN__
5794 #define vdupq_lane_f16(__p0_20, __p1_20) __extension__ ({ \
5795   float16x8_t __ret_20; \
5796   float16x4_t __s0_20 = __p0_20; \
5797   __ret_20 = splatq_lane_f16(__s0_20, __p1_20); \
5798   __ret_20; \
5799 })
5800 #else
5801 #define vdupq_lane_f16(__p0_21, __p1_21) __extension__ ({ \
5802   float16x8_t __ret_21; \
5803   float16x4_t __s0_21 = __p0_21; \
5804   float16x4_t __rev0_21;  __rev0_21 = __builtin_shufflevector(__s0_21, __s0_21, 3, 2, 1, 0); \
5805   __ret_21 = __noswap_splatq_lane_f16(__rev0_21, __p1_21); \
5806   __ret_21 = __builtin_shufflevector(__ret_21, __ret_21, 7, 6, 5, 4, 3, 2, 1, 0); \
5807   __ret_21; \
5808 })
5809 #endif
5810 
5811 #ifdef __LITTLE_ENDIAN__
5812 #define vdupq_lane_s32(__p0_22, __p1_22) __extension__ ({ \
5813   int32x4_t __ret_22; \
5814   int32x2_t __s0_22 = __p0_22; \
5815   __ret_22 = splatq_lane_s32(__s0_22, __p1_22); \
5816   __ret_22; \
5817 })
5818 #else
5819 #define vdupq_lane_s32(__p0_23, __p1_23) __extension__ ({ \
5820   int32x4_t __ret_23; \
5821   int32x2_t __s0_23 = __p0_23; \
5822   int32x2_t __rev0_23;  __rev0_23 = __builtin_shufflevector(__s0_23, __s0_23, 1, 0); \
5823   __ret_23 = __noswap_splatq_lane_s32(__rev0_23, __p1_23); \
5824   __ret_23 = __builtin_shufflevector(__ret_23, __ret_23, 3, 2, 1, 0); \
5825   __ret_23; \
5826 })
5827 #endif
5828 
5829 #ifdef __LITTLE_ENDIAN__
5830 #define vdupq_lane_s64(__p0_24, __p1_24) __extension__ ({ \
5831   int64x2_t __ret_24; \
5832   int64x1_t __s0_24 = __p0_24; \
5833   __ret_24 = splatq_lane_s64(__s0_24, __p1_24); \
5834   __ret_24; \
5835 })
5836 #else
5837 #define vdupq_lane_s64(__p0_25, __p1_25) __extension__ ({ \
5838   int64x2_t __ret_25; \
5839   int64x1_t __s0_25 = __p0_25; \
5840   __ret_25 = __noswap_splatq_lane_s64(__s0_25, __p1_25); \
5841   __ret_25 = __builtin_shufflevector(__ret_25, __ret_25, 1, 0); \
5842   __ret_25; \
5843 })
5844 #endif
5845 
5846 #ifdef __LITTLE_ENDIAN__
5847 #define vdupq_lane_s16(__p0_26, __p1_26) __extension__ ({ \
5848   int16x8_t __ret_26; \
5849   int16x4_t __s0_26 = __p0_26; \
5850   __ret_26 = splatq_lane_s16(__s0_26, __p1_26); \
5851   __ret_26; \
5852 })
5853 #else
5854 #define vdupq_lane_s16(__p0_27, __p1_27) __extension__ ({ \
5855   int16x8_t __ret_27; \
5856   int16x4_t __s0_27 = __p0_27; \
5857   int16x4_t __rev0_27;  __rev0_27 = __builtin_shufflevector(__s0_27, __s0_27, 3, 2, 1, 0); \
5858   __ret_27 = __noswap_splatq_lane_s16(__rev0_27, __p1_27); \
5859   __ret_27 = __builtin_shufflevector(__ret_27, __ret_27, 7, 6, 5, 4, 3, 2, 1, 0); \
5860   __ret_27; \
5861 })
5862 #endif
5863 
5864 #ifdef __LITTLE_ENDIAN__
5865 #define vdup_lane_u8(__p0_28, __p1_28) __extension__ ({ \
5866   uint8x8_t __ret_28; \
5867   uint8x8_t __s0_28 = __p0_28; \
5868   __ret_28 = splat_lane_u8(__s0_28, __p1_28); \
5869   __ret_28; \
5870 })
5871 #else
5872 #define vdup_lane_u8(__p0_29, __p1_29) __extension__ ({ \
5873   uint8x8_t __ret_29; \
5874   uint8x8_t __s0_29 = __p0_29; \
5875   uint8x8_t __rev0_29;  __rev0_29 = __builtin_shufflevector(__s0_29, __s0_29, 7, 6, 5, 4, 3, 2, 1, 0); \
5876   __ret_29 = __noswap_splat_lane_u8(__rev0_29, __p1_29); \
5877   __ret_29 = __builtin_shufflevector(__ret_29, __ret_29, 7, 6, 5, 4, 3, 2, 1, 0); \
5878   __ret_29; \
5879 })
5880 #endif
5881 
5882 #ifdef __LITTLE_ENDIAN__
5883 #define vdup_lane_u32(__p0_30, __p1_30) __extension__ ({ \
5884   uint32x2_t __ret_30; \
5885   uint32x2_t __s0_30 = __p0_30; \
5886   __ret_30 = splat_lane_u32(__s0_30, __p1_30); \
5887   __ret_30; \
5888 })
5889 #else
5890 #define vdup_lane_u32(__p0_31, __p1_31) __extension__ ({ \
5891   uint32x2_t __ret_31; \
5892   uint32x2_t __s0_31 = __p0_31; \
5893   uint32x2_t __rev0_31;  __rev0_31 = __builtin_shufflevector(__s0_31, __s0_31, 1, 0); \
5894   __ret_31 = __noswap_splat_lane_u32(__rev0_31, __p1_31); \
5895   __ret_31 = __builtin_shufflevector(__ret_31, __ret_31, 1, 0); \
5896   __ret_31; \
5897 })
5898 #endif
5899 
5900 #define vdup_lane_u64(__p0_32, __p1_32) __extension__ ({ \
5901   uint64x1_t __ret_32; \
5902   uint64x1_t __s0_32 = __p0_32; \
5903   __ret_32 = splat_lane_u64(__s0_32, __p1_32); \
5904   __ret_32; \
5905 })
5906 #ifdef __LITTLE_ENDIAN__
5907 #define vdup_lane_u16(__p0_33, __p1_33) __extension__ ({ \
5908   uint16x4_t __ret_33; \
5909   uint16x4_t __s0_33 = __p0_33; \
5910   __ret_33 = splat_lane_u16(__s0_33, __p1_33); \
5911   __ret_33; \
5912 })
5913 #else
5914 #define vdup_lane_u16(__p0_34, __p1_34) __extension__ ({ \
5915   uint16x4_t __ret_34; \
5916   uint16x4_t __s0_34 = __p0_34; \
5917   uint16x4_t __rev0_34;  __rev0_34 = __builtin_shufflevector(__s0_34, __s0_34, 3, 2, 1, 0); \
5918   __ret_34 = __noswap_splat_lane_u16(__rev0_34, __p1_34); \
5919   __ret_34 = __builtin_shufflevector(__ret_34, __ret_34, 3, 2, 1, 0); \
5920   __ret_34; \
5921 })
5922 #endif
5923 
5924 #ifdef __LITTLE_ENDIAN__
5925 #define vdup_lane_s8(__p0_35, __p1_35) __extension__ ({ \
5926   int8x8_t __ret_35; \
5927   int8x8_t __s0_35 = __p0_35; \
5928   __ret_35 = splat_lane_s8(__s0_35, __p1_35); \
5929   __ret_35; \
5930 })
5931 #else
5932 #define vdup_lane_s8(__p0_36, __p1_36) __extension__ ({ \
5933   int8x8_t __ret_36; \
5934   int8x8_t __s0_36 = __p0_36; \
5935   int8x8_t __rev0_36;  __rev0_36 = __builtin_shufflevector(__s0_36, __s0_36, 7, 6, 5, 4, 3, 2, 1, 0); \
5936   __ret_36 = __noswap_splat_lane_s8(__rev0_36, __p1_36); \
5937   __ret_36 = __builtin_shufflevector(__ret_36, __ret_36, 7, 6, 5, 4, 3, 2, 1, 0); \
5938   __ret_36; \
5939 })
5940 #endif
5941 
5942 #ifdef __LITTLE_ENDIAN__
5943 #define vdup_lane_f32(__p0_37, __p1_37) __extension__ ({ \
5944   float32x2_t __ret_37; \
5945   float32x2_t __s0_37 = __p0_37; \
5946   __ret_37 = splat_lane_f32(__s0_37, __p1_37); \
5947   __ret_37; \
5948 })
5949 #else
5950 #define vdup_lane_f32(__p0_38, __p1_38) __extension__ ({ \
5951   float32x2_t __ret_38; \
5952   float32x2_t __s0_38 = __p0_38; \
5953   float32x2_t __rev0_38;  __rev0_38 = __builtin_shufflevector(__s0_38, __s0_38, 1, 0); \
5954   __ret_38 = __noswap_splat_lane_f32(__rev0_38, __p1_38); \
5955   __ret_38 = __builtin_shufflevector(__ret_38, __ret_38, 1, 0); \
5956   __ret_38; \
5957 })
5958 #endif
5959 
5960 #ifdef __LITTLE_ENDIAN__
5961 #define vdup_lane_f16(__p0_39, __p1_39) __extension__ ({ \
5962   float16x4_t __ret_39; \
5963   float16x4_t __s0_39 = __p0_39; \
5964   __ret_39 = splat_lane_f16(__s0_39, __p1_39); \
5965   __ret_39; \
5966 })
5967 #else
5968 #define vdup_lane_f16(__p0_40, __p1_40) __extension__ ({ \
5969   float16x4_t __ret_40; \
5970   float16x4_t __s0_40 = __p0_40; \
5971   float16x4_t __rev0_40;  __rev0_40 = __builtin_shufflevector(__s0_40, __s0_40, 3, 2, 1, 0); \
5972   __ret_40 = __noswap_splat_lane_f16(__rev0_40, __p1_40); \
5973   __ret_40 = __builtin_shufflevector(__ret_40, __ret_40, 3, 2, 1, 0); \
5974   __ret_40; \
5975 })
5976 #endif
5977 
5978 #ifdef __LITTLE_ENDIAN__
5979 #define vdup_lane_s32(__p0_41, __p1_41) __extension__ ({ \
5980   int32x2_t __ret_41; \
5981   int32x2_t __s0_41 = __p0_41; \
5982   __ret_41 = splat_lane_s32(__s0_41, __p1_41); \
5983   __ret_41; \
5984 })
5985 #else
5986 #define vdup_lane_s32(__p0_42, __p1_42) __extension__ ({ \
5987   int32x2_t __ret_42; \
5988   int32x2_t __s0_42 = __p0_42; \
5989   int32x2_t __rev0_42;  __rev0_42 = __builtin_shufflevector(__s0_42, __s0_42, 1, 0); \
5990   __ret_42 = __noswap_splat_lane_s32(__rev0_42, __p1_42); \
5991   __ret_42 = __builtin_shufflevector(__ret_42, __ret_42, 1, 0); \
5992   __ret_42; \
5993 })
5994 #endif
5995 
5996 #define vdup_lane_s64(__p0_43, __p1_43) __extension__ ({ \
5997   int64x1_t __ret_43; \
5998   int64x1_t __s0_43 = __p0_43; \
5999   __ret_43 = splat_lane_s64(__s0_43, __p1_43); \
6000   __ret_43; \
6001 })
6002 #ifdef __LITTLE_ENDIAN__
6003 #define vdup_lane_s16(__p0_44, __p1_44) __extension__ ({ \
6004   int16x4_t __ret_44; \
6005   int16x4_t __s0_44 = __p0_44; \
6006   __ret_44 = splat_lane_s16(__s0_44, __p1_44); \
6007   __ret_44; \
6008 })
6009 #else
6010 #define vdup_lane_s16(__p0_45, __p1_45) __extension__ ({ \
6011   int16x4_t __ret_45; \
6012   int16x4_t __s0_45 = __p0_45; \
6013   int16x4_t __rev0_45;  __rev0_45 = __builtin_shufflevector(__s0_45, __s0_45, 3, 2, 1, 0); \
6014   __ret_45 = __noswap_splat_lane_s16(__rev0_45, __p1_45); \
6015   __ret_45 = __builtin_shufflevector(__ret_45, __ret_45, 3, 2, 1, 0); \
6016   __ret_45; \
6017 })
6018 #endif
6019 
6020 #ifdef __LITTLE_ENDIAN__
vdup_n_p8(poly8_t __p0)6021 __ai poly8x8_t vdup_n_p8(poly8_t __p0) {
6022   poly8x8_t __ret;
6023   __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6024   return __ret;
6025 }
6026 #else
vdup_n_p8(poly8_t __p0)6027 __ai poly8x8_t vdup_n_p8(poly8_t __p0) {
6028   poly8x8_t __ret;
6029   __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6030   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6031   return __ret;
6032 }
6033 #endif
6034 
6035 #ifdef __LITTLE_ENDIAN__
vdup_n_p16(poly16_t __p0)6036 __ai poly16x4_t vdup_n_p16(poly16_t __p0) {
6037   poly16x4_t __ret;
6038   __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
6039   return __ret;
6040 }
6041 #else
vdup_n_p16(poly16_t __p0)6042 __ai poly16x4_t vdup_n_p16(poly16_t __p0) {
6043   poly16x4_t __ret;
6044   __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
6045   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6046   return __ret;
6047 }
6048 #endif
6049 
6050 #ifdef __LITTLE_ENDIAN__
vdupq_n_p8(poly8_t __p0)6051 __ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
6052   poly8x16_t __ret;
6053   __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6054   return __ret;
6055 }
6056 #else
vdupq_n_p8(poly8_t __p0)6057 __ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
6058   poly8x16_t __ret;
6059   __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6060   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6061   return __ret;
6062 }
6063 #endif
6064 
6065 #ifdef __LITTLE_ENDIAN__
vdupq_n_p16(poly16_t __p0)6066 __ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
6067   poly16x8_t __ret;
6068   __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6069   return __ret;
6070 }
6071 #else
vdupq_n_p16(poly16_t __p0)6072 __ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
6073   poly16x8_t __ret;
6074   __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6075   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6076   return __ret;
6077 }
6078 #endif
6079 
6080 #ifdef __LITTLE_ENDIAN__
vdupq_n_u8(uint8_t __p0)6081 __ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
6082   uint8x16_t __ret;
6083   __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6084   return __ret;
6085 }
6086 #else
vdupq_n_u8(uint8_t __p0)6087 __ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
6088   uint8x16_t __ret;
6089   __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6090   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6091   return __ret;
6092 }
6093 #endif
6094 
6095 #ifdef __LITTLE_ENDIAN__
vdupq_n_u32(uint32_t __p0)6096 __ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
6097   uint32x4_t __ret;
6098   __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
6099   return __ret;
6100 }
6101 #else
vdupq_n_u32(uint32_t __p0)6102 __ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
6103   uint32x4_t __ret;
6104   __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
6105   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6106   return __ret;
6107 }
6108 #endif
6109 
6110 #ifdef __LITTLE_ENDIAN__
vdupq_n_u64(uint64_t __p0)6111 __ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
6112   uint64x2_t __ret;
6113   __ret = (uint64x2_t) {__p0, __p0};
6114   return __ret;
6115 }
6116 #else
vdupq_n_u64(uint64_t __p0)6117 __ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
6118   uint64x2_t __ret;
6119   __ret = (uint64x2_t) {__p0, __p0};
6120   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6121   return __ret;
6122 }
6123 #endif
6124 
6125 #ifdef __LITTLE_ENDIAN__
vdupq_n_u16(uint16_t __p0)6126 __ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
6127   uint16x8_t __ret;
6128   __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6129   return __ret;
6130 }
6131 #else
vdupq_n_u16(uint16_t __p0)6132 __ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
6133   uint16x8_t __ret;
6134   __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6135   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6136   return __ret;
6137 }
6138 #endif
6139 
6140 #ifdef __LITTLE_ENDIAN__
vdupq_n_s8(int8_t __p0)6141 __ai int8x16_t vdupq_n_s8(int8_t __p0) {
6142   int8x16_t __ret;
6143   __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6144   return __ret;
6145 }
6146 #else
vdupq_n_s8(int8_t __p0)6147 __ai int8x16_t vdupq_n_s8(int8_t __p0) {
6148   int8x16_t __ret;
6149   __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6150   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6151   return __ret;
6152 }
6153 #endif
6154 
6155 #ifdef __LITTLE_ENDIAN__
vdupq_n_f32(float32_t __p0)6156 __ai float32x4_t vdupq_n_f32(float32_t __p0) {
6157   float32x4_t __ret;
6158   __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
6159   return __ret;
6160 }
6161 #else
vdupq_n_f32(float32_t __p0)6162 __ai float32x4_t vdupq_n_f32(float32_t __p0) {
6163   float32x4_t __ret;
6164   __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
6165   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6166   return __ret;
6167 }
6168 #endif
6169 
6170 #ifdef __LITTLE_ENDIAN__
6171 #define vdupq_n_f16(__p0) __extension__ ({ \
6172   float16x8_t __ret; \
6173   float16_t __s0 = __p0; \
6174   __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
6175   __ret; \
6176 })
6177 #else
6178 #define vdupq_n_f16(__p0) __extension__ ({ \
6179   float16x8_t __ret; \
6180   float16_t __s0 = __p0; \
6181   __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
6182   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6183   __ret; \
6184 })
6185 #endif
6186 
6187 #ifdef __LITTLE_ENDIAN__
vdupq_n_s32(int32_t __p0)6188 __ai int32x4_t vdupq_n_s32(int32_t __p0) {
6189   int32x4_t __ret;
6190   __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
6191   return __ret;
6192 }
6193 #else
vdupq_n_s32(int32_t __p0)6194 __ai int32x4_t vdupq_n_s32(int32_t __p0) {
6195   int32x4_t __ret;
6196   __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
6197   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6198   return __ret;
6199 }
6200 #endif
6201 
6202 #ifdef __LITTLE_ENDIAN__
vdupq_n_s64(int64_t __p0)6203 __ai int64x2_t vdupq_n_s64(int64_t __p0) {
6204   int64x2_t __ret;
6205   __ret = (int64x2_t) {__p0, __p0};
6206   return __ret;
6207 }
6208 #else
vdupq_n_s64(int64_t __p0)6209 __ai int64x2_t vdupq_n_s64(int64_t __p0) {
6210   int64x2_t __ret;
6211   __ret = (int64x2_t) {__p0, __p0};
6212   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6213   return __ret;
6214 }
6215 #endif
6216 
6217 #ifdef __LITTLE_ENDIAN__
vdupq_n_s16(int16_t __p0)6218 __ai int16x8_t vdupq_n_s16(int16_t __p0) {
6219   int16x8_t __ret;
6220   __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6221   return __ret;
6222 }
6223 #else
vdupq_n_s16(int16_t __p0)6224 __ai int16x8_t vdupq_n_s16(int16_t __p0) {
6225   int16x8_t __ret;
6226   __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6227   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6228   return __ret;
6229 }
6230 #endif
6231 
6232 #ifdef __LITTLE_ENDIAN__
vdup_n_u8(uint8_t __p0)6233 __ai uint8x8_t vdup_n_u8(uint8_t __p0) {
6234   uint8x8_t __ret;
6235   __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6236   return __ret;
6237 }
6238 #else
vdup_n_u8(uint8_t __p0)6239 __ai uint8x8_t vdup_n_u8(uint8_t __p0) {
6240   uint8x8_t __ret;
6241   __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6242   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6243   return __ret;
6244 }
6245 #endif
6246 
6247 #ifdef __LITTLE_ENDIAN__
vdup_n_u32(uint32_t __p0)6248 __ai uint32x2_t vdup_n_u32(uint32_t __p0) {
6249   uint32x2_t __ret;
6250   __ret = (uint32x2_t) {__p0, __p0};
6251   return __ret;
6252 }
6253 #else
vdup_n_u32(uint32_t __p0)6254 __ai uint32x2_t vdup_n_u32(uint32_t __p0) {
6255   uint32x2_t __ret;
6256   __ret = (uint32x2_t) {__p0, __p0};
6257   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6258   return __ret;
6259 }
6260 #endif
6261 
vdup_n_u64(uint64_t __p0)6262 __ai uint64x1_t vdup_n_u64(uint64_t __p0) {
6263   uint64x1_t __ret;
6264   __ret = (uint64x1_t) {__p0};
6265   return __ret;
6266 }
6267 #ifdef __LITTLE_ENDIAN__
vdup_n_u16(uint16_t __p0)6268 __ai uint16x4_t vdup_n_u16(uint16_t __p0) {
6269   uint16x4_t __ret;
6270   __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
6271   return __ret;
6272 }
6273 #else
vdup_n_u16(uint16_t __p0)6274 __ai uint16x4_t vdup_n_u16(uint16_t __p0) {
6275   uint16x4_t __ret;
6276   __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
6277   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6278   return __ret;
6279 }
6280 #endif
6281 
6282 #ifdef __LITTLE_ENDIAN__
vdup_n_s8(int8_t __p0)6283 __ai int8x8_t vdup_n_s8(int8_t __p0) {
6284   int8x8_t __ret;
6285   __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6286   return __ret;
6287 }
6288 #else
vdup_n_s8(int8_t __p0)6289 __ai int8x8_t vdup_n_s8(int8_t __p0) {
6290   int8x8_t __ret;
6291   __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6292   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6293   return __ret;
6294 }
6295 #endif
6296 
6297 #ifdef __LITTLE_ENDIAN__
vdup_n_f32(float32_t __p0)6298 __ai float32x2_t vdup_n_f32(float32_t __p0) {
6299   float32x2_t __ret;
6300   __ret = (float32x2_t) {__p0, __p0};
6301   return __ret;
6302 }
6303 #else
vdup_n_f32(float32_t __p0)6304 __ai float32x2_t vdup_n_f32(float32_t __p0) {
6305   float32x2_t __ret;
6306   __ret = (float32x2_t) {__p0, __p0};
6307   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6308   return __ret;
6309 }
6310 #endif
6311 
6312 #ifdef __LITTLE_ENDIAN__
6313 #define vdup_n_f16(__p0) __extension__ ({ \
6314   float16x4_t __ret; \
6315   float16_t __s0 = __p0; \
6316   __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
6317   __ret; \
6318 })
6319 #else
6320 #define vdup_n_f16(__p0) __extension__ ({ \
6321   float16x4_t __ret; \
6322   float16_t __s0 = __p0; \
6323   __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
6324   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6325   __ret; \
6326 })
6327 #endif
6328 
6329 #ifdef __LITTLE_ENDIAN__
vdup_n_s32(int32_t __p0)6330 __ai int32x2_t vdup_n_s32(int32_t __p0) {
6331   int32x2_t __ret;
6332   __ret = (int32x2_t) {__p0, __p0};
6333   return __ret;
6334 }
6335 #else
vdup_n_s32(int32_t __p0)6336 __ai int32x2_t vdup_n_s32(int32_t __p0) {
6337   int32x2_t __ret;
6338   __ret = (int32x2_t) {__p0, __p0};
6339   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6340   return __ret;
6341 }
6342 #endif
6343 
vdup_n_s64(int64_t __p0)6344 __ai int64x1_t vdup_n_s64(int64_t __p0) {
6345   int64x1_t __ret;
6346   __ret = (int64x1_t) {__p0};
6347   return __ret;
6348 }
6349 #ifdef __LITTLE_ENDIAN__
vdup_n_s16(int16_t __p0)6350 __ai int16x4_t vdup_n_s16(int16_t __p0) {
6351   int16x4_t __ret;
6352   __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
6353   return __ret;
6354 }
6355 #else
vdup_n_s16(int16_t __p0)6356 __ai int16x4_t vdup_n_s16(int16_t __p0) {
6357   int16x4_t __ret;
6358   __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
6359   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6360   return __ret;
6361 }
6362 #endif
6363 
6364 #ifdef __LITTLE_ENDIAN__
veorq_u8(uint8x16_t __p0,uint8x16_t __p1)6365 __ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
6366   uint8x16_t __ret;
6367   __ret = __p0 ^ __p1;
6368   return __ret;
6369 }
6370 #else
veorq_u8(uint8x16_t __p0,uint8x16_t __p1)6371 __ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
6372   uint8x16_t __ret;
6373   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6374   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6375   __ret = __rev0 ^ __rev1;
6376   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6377   return __ret;
6378 }
6379 #endif
6380 
6381 #ifdef __LITTLE_ENDIAN__
veorq_u32(uint32x4_t __p0,uint32x4_t __p1)6382 __ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
6383   uint32x4_t __ret;
6384   __ret = __p0 ^ __p1;
6385   return __ret;
6386 }
6387 #else
veorq_u32(uint32x4_t __p0,uint32x4_t __p1)6388 __ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
6389   uint32x4_t __ret;
6390   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6391   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6392   __ret = __rev0 ^ __rev1;
6393   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6394   return __ret;
6395 }
6396 #endif
6397 
6398 #ifdef __LITTLE_ENDIAN__
veorq_u64(uint64x2_t __p0,uint64x2_t __p1)6399 __ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
6400   uint64x2_t __ret;
6401   __ret = __p0 ^ __p1;
6402   return __ret;
6403 }
6404 #else
veorq_u64(uint64x2_t __p0,uint64x2_t __p1)6405 __ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
6406   uint64x2_t __ret;
6407   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6408   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6409   __ret = __rev0 ^ __rev1;
6410   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6411   return __ret;
6412 }
6413 #endif
6414 
6415 #ifdef __LITTLE_ENDIAN__
veorq_u16(uint16x8_t __p0,uint16x8_t __p1)6416 __ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
6417   uint16x8_t __ret;
6418   __ret = __p0 ^ __p1;
6419   return __ret;
6420 }
6421 #else
veorq_u16(uint16x8_t __p0,uint16x8_t __p1)6422 __ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
6423   uint16x8_t __ret;
6424   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6425   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
6426   __ret = __rev0 ^ __rev1;
6427   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6428   return __ret;
6429 }
6430 #endif
6431 
6432 #ifdef __LITTLE_ENDIAN__
veorq_s8(int8x16_t __p0,int8x16_t __p1)6433 __ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
6434   int8x16_t __ret;
6435   __ret = __p0 ^ __p1;
6436   return __ret;
6437 }
6438 #else
veorq_s8(int8x16_t __p0,int8x16_t __p1)6439 __ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
6440   int8x16_t __ret;
6441   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6442   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6443   __ret = __rev0 ^ __rev1;
6444   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6445   return __ret;
6446 }
6447 #endif
6448 
6449 #ifdef __LITTLE_ENDIAN__
veorq_s32(int32x4_t __p0,int32x4_t __p1)6450 __ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
6451   int32x4_t __ret;
6452   __ret = __p0 ^ __p1;
6453   return __ret;
6454 }
6455 #else
veorq_s32(int32x4_t __p0,int32x4_t __p1)6456 __ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
6457   int32x4_t __ret;
6458   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6459   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6460   __ret = __rev0 ^ __rev1;
6461   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6462   return __ret;
6463 }
6464 #endif
6465 
6466 #ifdef __LITTLE_ENDIAN__
veorq_s64(int64x2_t __p0,int64x2_t __p1)6467 __ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
6468   int64x2_t __ret;
6469   __ret = __p0 ^ __p1;
6470   return __ret;
6471 }
6472 #else
veorq_s64(int64x2_t __p0,int64x2_t __p1)6473 __ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
6474   int64x2_t __ret;
6475   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6476   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6477   __ret = __rev0 ^ __rev1;
6478   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6479   return __ret;
6480 }
6481 #endif
6482 
6483 #ifdef __LITTLE_ENDIAN__
veorq_s16(int16x8_t __p0,int16x8_t __p1)6484 __ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
6485   int16x8_t __ret;
6486   __ret = __p0 ^ __p1;
6487   return __ret;
6488 }
6489 #else
veorq_s16(int16x8_t __p0,int16x8_t __p1)6490 __ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
6491   int16x8_t __ret;
6492   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6493   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
6494   __ret = __rev0 ^ __rev1;
6495   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6496   return __ret;
6497 }
6498 #endif
6499 
6500 #ifdef __LITTLE_ENDIAN__
veor_u8(uint8x8_t __p0,uint8x8_t __p1)6501 __ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
6502   uint8x8_t __ret;
6503   __ret = __p0 ^ __p1;
6504   return __ret;
6505 }
6506 #else
veor_u8(uint8x8_t __p0,uint8x8_t __p1)6507 __ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
6508   uint8x8_t __ret;
6509   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6510   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
6511   __ret = __rev0 ^ __rev1;
6512   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6513   return __ret;
6514 }
6515 #endif
6516 
6517 #ifdef __LITTLE_ENDIAN__
veor_u32(uint32x2_t __p0,uint32x2_t __p1)6518 __ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
6519   uint32x2_t __ret;
6520   __ret = __p0 ^ __p1;
6521   return __ret;
6522 }
6523 #else
veor_u32(uint32x2_t __p0,uint32x2_t __p1)6524 __ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
6525   uint32x2_t __ret;
6526   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6527   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6528   __ret = __rev0 ^ __rev1;
6529   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6530   return __ret;
6531 }
6532 #endif
6533 
veor_u64(uint64x1_t __p0,uint64x1_t __p1)6534 __ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
6535   uint64x1_t __ret;
6536   __ret = __p0 ^ __p1;
6537   return __ret;
6538 }
6539 #ifdef __LITTLE_ENDIAN__
veor_u16(uint16x4_t __p0,uint16x4_t __p1)6540 __ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
6541   uint16x4_t __ret;
6542   __ret = __p0 ^ __p1;
6543   return __ret;
6544 }
6545 #else
veor_u16(uint16x4_t __p0,uint16x4_t __p1)6546 __ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
6547   uint16x4_t __ret;
6548   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6549   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6550   __ret = __rev0 ^ __rev1;
6551   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6552   return __ret;
6553 }
6554 #endif
6555 
6556 #ifdef __LITTLE_ENDIAN__
veor_s8(int8x8_t __p0,int8x8_t __p1)6557 __ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
6558   int8x8_t __ret;
6559   __ret = __p0 ^ __p1;
6560   return __ret;
6561 }
6562 #else
veor_s8(int8x8_t __p0,int8x8_t __p1)6563 __ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
6564   int8x8_t __ret;
6565   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6566   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
6567   __ret = __rev0 ^ __rev1;
6568   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6569   return __ret;
6570 }
6571 #endif
6572 
6573 #ifdef __LITTLE_ENDIAN__
veor_s32(int32x2_t __p0,int32x2_t __p1)6574 __ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
6575   int32x2_t __ret;
6576   __ret = __p0 ^ __p1;
6577   return __ret;
6578 }
6579 #else
veor_s32(int32x2_t __p0,int32x2_t __p1)6580 __ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
6581   int32x2_t __ret;
6582   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6583   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6584   __ret = __rev0 ^ __rev1;
6585   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6586   return __ret;
6587 }
6588 #endif
6589 
veor_s64(int64x1_t __p0,int64x1_t __p1)6590 __ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
6591   int64x1_t __ret;
6592   __ret = __p0 ^ __p1;
6593   return __ret;
6594 }
6595 #ifdef __LITTLE_ENDIAN__
veor_s16(int16x4_t __p0,int16x4_t __p1)6596 __ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
6597   int16x4_t __ret;
6598   __ret = __p0 ^ __p1;
6599   return __ret;
6600 }
6601 #else
veor_s16(int16x4_t __p0,int16x4_t __p1)6602 __ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
6603   int16x4_t __ret;
6604   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6605   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6606   __ret = __rev0 ^ __rev1;
6607   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6608   return __ret;
6609 }
6610 #endif
6611 
6612 #ifdef __LITTLE_ENDIAN__
6613 #define vext_p8(__p0, __p1, __p2) __extension__ ({ \
6614   poly8x8_t __ret; \
6615   poly8x8_t __s0 = __p0; \
6616   poly8x8_t __s1 = __p1; \
6617   __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
6618   __ret; \
6619 })
6620 #else
6621 #define vext_p8(__p0, __p1, __p2) __extension__ ({ \
6622   poly8x8_t __ret; \
6623   poly8x8_t __s0 = __p0; \
6624   poly8x8_t __s1 = __p1; \
6625   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6626   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6627   __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
6628   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6629   __ret; \
6630 })
6631 #endif
6632 
6633 #ifdef __LITTLE_ENDIAN__
6634 #define vext_p16(__p0, __p1, __p2) __extension__ ({ \
6635   poly16x4_t __ret; \
6636   poly16x4_t __s0 = __p0; \
6637   poly16x4_t __s1 = __p1; \
6638   __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
6639   __ret; \
6640 })
6641 #else
6642 #define vext_p16(__p0, __p1, __p2) __extension__ ({ \
6643   poly16x4_t __ret; \
6644   poly16x4_t __s0 = __p0; \
6645   poly16x4_t __s1 = __p1; \
6646   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6647   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6648   __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
6649   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6650   __ret; \
6651 })
6652 #endif
6653 
6654 #ifdef __LITTLE_ENDIAN__
6655 #define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
6656   poly8x16_t __ret; \
6657   poly8x16_t __s0 = __p0; \
6658   poly8x16_t __s1 = __p1; \
6659   __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
6660   __ret; \
6661 })
6662 #else
6663 #define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
6664   poly8x16_t __ret; \
6665   poly8x16_t __s0 = __p0; \
6666   poly8x16_t __s1 = __p1; \
6667   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6668   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6669   __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
6670   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6671   __ret; \
6672 })
6673 #endif
6674 
6675 #ifdef __LITTLE_ENDIAN__
6676 #define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
6677   poly16x8_t __ret; \
6678   poly16x8_t __s0 = __p0; \
6679   poly16x8_t __s1 = __p1; \
6680   __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
6681   __ret; \
6682 })
6683 #else
6684 #define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
6685   poly16x8_t __ret; \
6686   poly16x8_t __s0 = __p0; \
6687   poly16x8_t __s1 = __p1; \
6688   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6689   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6690   __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
6691   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6692   __ret; \
6693 })
6694 #endif
6695 
6696 #ifdef __LITTLE_ENDIAN__
6697 #define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
6698   uint8x16_t __ret; \
6699   uint8x16_t __s0 = __p0; \
6700   uint8x16_t __s1 = __p1; \
6701   __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
6702   __ret; \
6703 })
6704 #else
6705 #define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
6706   uint8x16_t __ret; \
6707   uint8x16_t __s0 = __p0; \
6708   uint8x16_t __s1 = __p1; \
6709   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6710   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6711   __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
6712   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6713   __ret; \
6714 })
6715 #endif
6716 
6717 #ifdef __LITTLE_ENDIAN__
6718 #define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
6719   uint32x4_t __ret; \
6720   uint32x4_t __s0 = __p0; \
6721   uint32x4_t __s1 = __p1; \
6722   __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
6723   __ret; \
6724 })
6725 #else
6726 #define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
6727   uint32x4_t __ret; \
6728   uint32x4_t __s0 = __p0; \
6729   uint32x4_t __s1 = __p1; \
6730   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6731   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6732   __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
6733   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6734   __ret; \
6735 })
6736 #endif
6737 
6738 #ifdef __LITTLE_ENDIAN__
6739 #define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
6740   uint64x2_t __ret; \
6741   uint64x2_t __s0 = __p0; \
6742   uint64x2_t __s1 = __p1; \
6743   __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
6744   __ret; \
6745 })
6746 #else
6747 #define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
6748   uint64x2_t __ret; \
6749   uint64x2_t __s0 = __p0; \
6750   uint64x2_t __s1 = __p1; \
6751   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6752   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6753   __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
6754   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6755   __ret; \
6756 })
6757 #endif
6758 
6759 #ifdef __LITTLE_ENDIAN__
6760 #define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
6761   uint16x8_t __ret; \
6762   uint16x8_t __s0 = __p0; \
6763   uint16x8_t __s1 = __p1; \
6764   __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
6765   __ret; \
6766 })
6767 #else
6768 #define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
6769   uint16x8_t __ret; \
6770   uint16x8_t __s0 = __p0; \
6771   uint16x8_t __s1 = __p1; \
6772   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6773   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6774   __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
6775   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6776   __ret; \
6777 })
6778 #endif
6779 
6780 #ifdef __LITTLE_ENDIAN__
6781 #define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
6782   int8x16_t __ret; \
6783   int8x16_t __s0 = __p0; \
6784   int8x16_t __s1 = __p1; \
6785   __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
6786   __ret; \
6787 })
6788 #else
6789 #define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
6790   int8x16_t __ret; \
6791   int8x16_t __s0 = __p0; \
6792   int8x16_t __s1 = __p1; \
6793   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6794   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6795   __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
6796   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6797   __ret; \
6798 })
6799 #endif
6800 
6801 #ifdef __LITTLE_ENDIAN__
6802 #define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
6803   float32x4_t __ret; \
6804   float32x4_t __s0 = __p0; \
6805   float32x4_t __s1 = __p1; \
6806   __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 41); \
6807   __ret; \
6808 })
6809 #else
6810 #define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
6811   float32x4_t __ret; \
6812   float32x4_t __s0 = __p0; \
6813   float32x4_t __s1 = __p1; \
6814   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6815   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6816   __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 41); \
6817   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6818   __ret; \
6819 })
6820 #endif
6821 
6822 #ifdef __LITTLE_ENDIAN__
6823 #define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
6824   int32x4_t __ret; \
6825   int32x4_t __s0 = __p0; \
6826   int32x4_t __s1 = __p1; \
6827   __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
6828   __ret; \
6829 })
6830 #else
6831 #define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
6832   int32x4_t __ret; \
6833   int32x4_t __s0 = __p0; \
6834   int32x4_t __s1 = __p1; \
6835   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6836   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6837   __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
6838   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6839   __ret; \
6840 })
6841 #endif
6842 
6843 #ifdef __LITTLE_ENDIAN__
6844 #define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
6845   int64x2_t __ret; \
6846   int64x2_t __s0 = __p0; \
6847   int64x2_t __s1 = __p1; \
6848   __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
6849   __ret; \
6850 })
6851 #else
6852 #define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
6853   int64x2_t __ret; \
6854   int64x2_t __s0 = __p0; \
6855   int64x2_t __s1 = __p1; \
6856   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6857   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6858   __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
6859   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6860   __ret; \
6861 })
6862 #endif
6863 
6864 #ifdef __LITTLE_ENDIAN__
6865 #define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
6866   int16x8_t __ret; \
6867   int16x8_t __s0 = __p0; \
6868   int16x8_t __s1 = __p1; \
6869   __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
6870   __ret; \
6871 })
6872 #else
6873 #define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
6874   int16x8_t __ret; \
6875   int16x8_t __s0 = __p0; \
6876   int16x8_t __s1 = __p1; \
6877   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6878   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6879   __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
6880   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6881   __ret; \
6882 })
6883 #endif
6884 
6885 #ifdef __LITTLE_ENDIAN__
6886 #define vext_u8(__p0, __p1, __p2) __extension__ ({ \
6887   uint8x8_t __ret; \
6888   uint8x8_t __s0 = __p0; \
6889   uint8x8_t __s1 = __p1; \
6890   __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
6891   __ret; \
6892 })
6893 #else
6894 #define vext_u8(__p0, __p1, __p2) __extension__ ({ \
6895   uint8x8_t __ret; \
6896   uint8x8_t __s0 = __p0; \
6897   uint8x8_t __s1 = __p1; \
6898   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6899   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6900   __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
6901   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6902   __ret; \
6903 })
6904 #endif
6905 
6906 #ifdef __LITTLE_ENDIAN__
6907 #define vext_u32(__p0, __p1, __p2) __extension__ ({ \
6908   uint32x2_t __ret; \
6909   uint32x2_t __s0 = __p0; \
6910   uint32x2_t __s1 = __p1; \
6911   __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
6912   __ret; \
6913 })
6914 #else
6915 #define vext_u32(__p0, __p1, __p2) __extension__ ({ \
6916   uint32x2_t __ret; \
6917   uint32x2_t __s0 = __p0; \
6918   uint32x2_t __s1 = __p1; \
6919   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6920   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6921   __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
6922   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6923   __ret; \
6924 })
6925 #endif
6926 
6927 #define vext_u64(__p0, __p1, __p2) __extension__ ({ \
6928   uint64x1_t __ret; \
6929   uint64x1_t __s0 = __p0; \
6930   uint64x1_t __s1 = __p1; \
6931   __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
6932   __ret; \
6933 })
6934 #ifdef __LITTLE_ENDIAN__
6935 #define vext_u16(__p0, __p1, __p2) __extension__ ({ \
6936   uint16x4_t __ret; \
6937   uint16x4_t __s0 = __p0; \
6938   uint16x4_t __s1 = __p1; \
6939   __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
6940   __ret; \
6941 })
6942 #else
6943 #define vext_u16(__p0, __p1, __p2) __extension__ ({ \
6944   uint16x4_t __ret; \
6945   uint16x4_t __s0 = __p0; \
6946   uint16x4_t __s1 = __p1; \
6947   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6948   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6949   __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
6950   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6951   __ret; \
6952 })
6953 #endif
6954 
6955 #ifdef __LITTLE_ENDIAN__
6956 #define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6957   int8x8_t __ret; \
6958   int8x8_t __s0 = __p0; \
6959   int8x8_t __s1 = __p1; \
6960   __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
6961   __ret; \
6962 })
6963 #else
6964 #define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6965   int8x8_t __ret; \
6966   int8x8_t __s0 = __p0; \
6967   int8x8_t __s1 = __p1; \
6968   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6969   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6970   __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
6971   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6972   __ret; \
6973 })
6974 #endif
6975 
6976 #ifdef __LITTLE_ENDIAN__
6977 #define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6978   float32x2_t __ret; \
6979   float32x2_t __s0 = __p0; \
6980   float32x2_t __s1 = __p1; \
6981   __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 9); \
6982   __ret; \
6983 })
6984 #else
6985 #define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6986   float32x2_t __ret; \
6987   float32x2_t __s0 = __p0; \
6988   float32x2_t __s1 = __p1; \
6989   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6990   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6991   __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 9); \
6992   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6993   __ret; \
6994 })
6995 #endif
6996 
6997 #ifdef __LITTLE_ENDIAN__
6998 #define vext_s32(__p0, __p1, __p2) __extension__ ({ \
6999   int32x2_t __ret; \
7000   int32x2_t __s0 = __p0; \
7001   int32x2_t __s1 = __p1; \
7002   __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
7003   __ret; \
7004 })
7005 #else
7006 #define vext_s32(__p0, __p1, __p2) __extension__ ({ \
7007   int32x2_t __ret; \
7008   int32x2_t __s0 = __p0; \
7009   int32x2_t __s1 = __p1; \
7010   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7011   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
7012   __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
7013   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7014   __ret; \
7015 })
7016 #endif
7017 
7018 #define vext_s64(__p0, __p1, __p2) __extension__ ({ \
7019   int64x1_t __ret; \
7020   int64x1_t __s0 = __p0; \
7021   int64x1_t __s1 = __p1; \
7022   __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
7023   __ret; \
7024 })
7025 #ifdef __LITTLE_ENDIAN__
7026 #define vext_s16(__p0, __p1, __p2) __extension__ ({ \
7027   int16x4_t __ret; \
7028   int16x4_t __s0 = __p0; \
7029   int16x4_t __s1 = __p1; \
7030   __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
7031   __ret; \
7032 })
7033 #else
7034 #define vext_s16(__p0, __p1, __p2) __extension__ ({ \
7035   int16x4_t __ret; \
7036   int16x4_t __s0 = __p0; \
7037   int16x4_t __s1 = __p1; \
7038   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7039   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
7040   __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
7041   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7042   __ret; \
7043 })
7044 #endif
7045 
7046 #ifdef __LITTLE_ENDIAN__
vget_high_p8(poly8x16_t __p0)7047 __ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
7048   poly8x8_t __ret;
7049   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7050   return __ret;
7051 }
7052 #else
vget_high_p8(poly8x16_t __p0)7053 __ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
7054   poly8x8_t __ret;
7055   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7056   __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
7057   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7058   return __ret;
7059 }
__noswap_vget_high_p8(poly8x16_t __p0)7060 __ai poly8x8_t __noswap_vget_high_p8(poly8x16_t __p0) {
7061   poly8x8_t __ret;
7062   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7063   return __ret;
7064 }
7065 #endif
7066 
7067 #ifdef __LITTLE_ENDIAN__
vget_high_p16(poly16x8_t __p0)7068 __ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
7069   poly16x4_t __ret;
7070   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7071   return __ret;
7072 }
7073 #else
vget_high_p16(poly16x8_t __p0)7074 __ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
7075   poly16x4_t __ret;
7076   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7077   __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
7078   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7079   return __ret;
7080 }
7081 #endif
7082 
7083 #ifdef __LITTLE_ENDIAN__
vget_high_u8(uint8x16_t __p0)7084 __ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
7085   uint8x8_t __ret;
7086   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7087   return __ret;
7088 }
7089 #else
vget_high_u8(uint8x16_t __p0)7090 __ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
7091   uint8x8_t __ret;
7092   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7093   __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
7094   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7095   return __ret;
7096 }
__noswap_vget_high_u8(uint8x16_t __p0)7097 __ai uint8x8_t __noswap_vget_high_u8(uint8x16_t __p0) {
7098   uint8x8_t __ret;
7099   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7100   return __ret;
7101 }
7102 #endif
7103 
7104 #ifdef __LITTLE_ENDIAN__
vget_high_u32(uint32x4_t __p0)7105 __ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
7106   uint32x2_t __ret;
7107   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7108   return __ret;
7109 }
7110 #else
vget_high_u32(uint32x4_t __p0)7111 __ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
7112   uint32x2_t __ret;
7113   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7114   __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
7115   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7116   return __ret;
7117 }
__noswap_vget_high_u32(uint32x4_t __p0)7118 __ai uint32x2_t __noswap_vget_high_u32(uint32x4_t __p0) {
7119   uint32x2_t __ret;
7120   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7121   return __ret;
7122 }
7123 #endif
7124 
7125 #ifdef __LITTLE_ENDIAN__
vget_high_u64(uint64x2_t __p0)7126 __ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
7127   uint64x1_t __ret;
7128   __ret = __builtin_shufflevector(__p0, __p0, 1);
7129   return __ret;
7130 }
7131 #else
vget_high_u64(uint64x2_t __p0)7132 __ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
7133   uint64x1_t __ret;
7134   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7135   __ret = __builtin_shufflevector(__rev0, __rev0, 1);
7136   return __ret;
7137 }
7138 #endif
7139 
7140 #ifdef __LITTLE_ENDIAN__
vget_high_u16(uint16x8_t __p0)7141 __ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
7142   uint16x4_t __ret;
7143   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7144   return __ret;
7145 }
7146 #else
vget_high_u16(uint16x8_t __p0)7147 __ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
7148   uint16x4_t __ret;
7149   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7150   __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
7151   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7152   return __ret;
7153 }
__noswap_vget_high_u16(uint16x8_t __p0)7154 __ai uint16x4_t __noswap_vget_high_u16(uint16x8_t __p0) {
7155   uint16x4_t __ret;
7156   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7157   return __ret;
7158 }
7159 #endif
7160 
7161 #ifdef __LITTLE_ENDIAN__
vget_high_s8(int8x16_t __p0)7162 __ai int8x8_t vget_high_s8(int8x16_t __p0) {
7163   int8x8_t __ret;
7164   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7165   return __ret;
7166 }
7167 #else
vget_high_s8(int8x16_t __p0)7168 __ai int8x8_t vget_high_s8(int8x16_t __p0) {
7169   int8x8_t __ret;
7170   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7171   __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
7172   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7173   return __ret;
7174 }
__noswap_vget_high_s8(int8x16_t __p0)7175 __ai int8x8_t __noswap_vget_high_s8(int8x16_t __p0) {
7176   int8x8_t __ret;
7177   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7178   return __ret;
7179 }
7180 #endif
7181 
7182 #ifdef __LITTLE_ENDIAN__
vget_high_f32(float32x4_t __p0)7183 __ai float32x2_t vget_high_f32(float32x4_t __p0) {
7184   float32x2_t __ret;
7185   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7186   return __ret;
7187 }
7188 #else
vget_high_f32(float32x4_t __p0)7189 __ai float32x2_t vget_high_f32(float32x4_t __p0) {
7190   float32x2_t __ret;
7191   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7192   __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
7193   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7194   return __ret;
7195 }
__noswap_vget_high_f32(float32x4_t __p0)7196 __ai float32x2_t __noswap_vget_high_f32(float32x4_t __p0) {
7197   float32x2_t __ret;
7198   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7199   return __ret;
7200 }
7201 #endif
7202 
7203 #ifdef __LITTLE_ENDIAN__
vget_high_f16(float16x8_t __p0)7204 __ai float16x4_t vget_high_f16(float16x8_t __p0) {
7205   float16x4_t __ret;
7206   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7207   return __ret;
7208 }
7209 #else
vget_high_f16(float16x8_t __p0)7210 __ai float16x4_t vget_high_f16(float16x8_t __p0) {
7211   float16x4_t __ret;
7212   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7213   __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
7214   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7215   return __ret;
7216 }
__noswap_vget_high_f16(float16x8_t __p0)7217 __ai float16x4_t __noswap_vget_high_f16(float16x8_t __p0) {
7218   float16x4_t __ret;
7219   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7220   return __ret;
7221 }
7222 #endif
7223 
7224 #ifdef __LITTLE_ENDIAN__
vget_high_s32(int32x4_t __p0)7225 __ai int32x2_t vget_high_s32(int32x4_t __p0) {
7226   int32x2_t __ret;
7227   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7228   return __ret;
7229 }
7230 #else
vget_high_s32(int32x4_t __p0)7231 __ai int32x2_t vget_high_s32(int32x4_t __p0) {
7232   int32x2_t __ret;
7233   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7234   __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
7235   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7236   return __ret;
7237 }
__noswap_vget_high_s32(int32x4_t __p0)7238 __ai int32x2_t __noswap_vget_high_s32(int32x4_t __p0) {
7239   int32x2_t __ret;
7240   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7241   return __ret;
7242 }
7243 #endif
7244 
7245 #ifdef __LITTLE_ENDIAN__
vget_high_s64(int64x2_t __p0)7246 __ai int64x1_t vget_high_s64(int64x2_t __p0) {
7247   int64x1_t __ret;
7248   __ret = __builtin_shufflevector(__p0, __p0, 1);
7249   return __ret;
7250 }
7251 #else
vget_high_s64(int64x2_t __p0)7252 __ai int64x1_t vget_high_s64(int64x2_t __p0) {
7253   int64x1_t __ret;
7254   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7255   __ret = __builtin_shufflevector(__rev0, __rev0, 1);
7256   return __ret;
7257 }
7258 #endif
7259 
7260 #ifdef __LITTLE_ENDIAN__
vget_high_s16(int16x8_t __p0)7261 __ai int16x4_t vget_high_s16(int16x8_t __p0) {
7262   int16x4_t __ret;
7263   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7264   return __ret;
7265 }
7266 #else
vget_high_s16(int16x8_t __p0)7267 __ai int16x4_t vget_high_s16(int16x8_t __p0) {
7268   int16x4_t __ret;
7269   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7270   __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
7271   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7272   return __ret;
7273 }
__noswap_vget_high_s16(int16x8_t __p0)7274 __ai int16x4_t __noswap_vget_high_s16(int16x8_t __p0) {
7275   int16x4_t __ret;
7276   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7277   return __ret;
7278 }
7279 #endif
7280 
7281 #ifdef __LITTLE_ENDIAN__
7282 #define vget_lane_p8(__p0, __p1) __extension__ ({ \
7283   poly8_t __ret; \
7284   poly8x8_t __s0 = __p0; \
7285   __ret = (poly8_t) __builtin_neon_vget_lane_i8((poly8x8_t)__s0, __p1); \
7286   __ret; \
7287 })
7288 #else
7289 #define vget_lane_p8(__p0, __p1) __extension__ ({ \
7290   poly8_t __ret; \
7291   poly8x8_t __s0 = __p0; \
7292   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7293   __ret = (poly8_t) __builtin_neon_vget_lane_i8((poly8x8_t)__rev0, __p1); \
7294   __ret; \
7295 })
7296 #define __noswap_vget_lane_p8(__p0, __p1) __extension__ ({ \
7297   poly8_t __ret; \
7298   poly8x8_t __s0 = __p0; \
7299   __ret = (poly8_t) __builtin_neon_vget_lane_i8((poly8x8_t)__s0, __p1); \
7300   __ret; \
7301 })
7302 #endif
7303 
7304 #ifdef __LITTLE_ENDIAN__
7305 #define vget_lane_p16(__p0, __p1) __extension__ ({ \
7306   poly16_t __ret; \
7307   poly16x4_t __s0 = __p0; \
7308   __ret = (poly16_t) __builtin_neon_vget_lane_i16((poly16x4_t)__s0, __p1); \
7309   __ret; \
7310 })
7311 #else
7312 #define vget_lane_p16(__p0, __p1) __extension__ ({ \
7313   poly16_t __ret; \
7314   poly16x4_t __s0 = __p0; \
7315   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7316   __ret = (poly16_t) __builtin_neon_vget_lane_i16((poly16x4_t)__rev0, __p1); \
7317   __ret; \
7318 })
7319 #define __noswap_vget_lane_p16(__p0, __p1) __extension__ ({ \
7320   poly16_t __ret; \
7321   poly16x4_t __s0 = __p0; \
7322   __ret = (poly16_t) __builtin_neon_vget_lane_i16((poly16x4_t)__s0, __p1); \
7323   __ret; \
7324 })
7325 #endif
7326 
7327 #ifdef __LITTLE_ENDIAN__
7328 #define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
7329   poly8_t __ret; \
7330   poly8x16_t __s0 = __p0; \
7331   __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((poly8x16_t)__s0, __p1); \
7332   __ret; \
7333 })
7334 #else
7335 #define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
7336   poly8_t __ret; \
7337   poly8x16_t __s0 = __p0; \
7338   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7339   __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((poly8x16_t)__rev0, __p1); \
7340   __ret; \
7341 })
7342 #define __noswap_vgetq_lane_p8(__p0, __p1) __extension__ ({ \
7343   poly8_t __ret; \
7344   poly8x16_t __s0 = __p0; \
7345   __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((poly8x16_t)__s0, __p1); \
7346   __ret; \
7347 })
7348 #endif
7349 
7350 #ifdef __LITTLE_ENDIAN__
7351 #define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
7352   poly16_t __ret; \
7353   poly16x8_t __s0 = __p0; \
7354   __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((poly16x8_t)__s0, __p1); \
7355   __ret; \
7356 })
7357 #else
7358 #define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
7359   poly16_t __ret; \
7360   poly16x8_t __s0 = __p0; \
7361   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7362   __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((poly16x8_t)__rev0, __p1); \
7363   __ret; \
7364 })
7365 #define __noswap_vgetq_lane_p16(__p0, __p1) __extension__ ({ \
7366   poly16_t __ret; \
7367   poly16x8_t __s0 = __p0; \
7368   __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((poly16x8_t)__s0, __p1); \
7369   __ret; \
7370 })
7371 #endif
7372 
7373 #ifdef __LITTLE_ENDIAN__
7374 #define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
7375   uint8_t __ret; \
7376   uint8x16_t __s0 = __p0; \
7377   __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
7378   __ret; \
7379 })
7380 #else
7381 #define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
7382   uint8_t __ret; \
7383   uint8x16_t __s0 = __p0; \
7384   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7385   __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
7386   __ret; \
7387 })
7388 #define __noswap_vgetq_lane_u8(__p0, __p1) __extension__ ({ \
7389   uint8_t __ret; \
7390   uint8x16_t __s0 = __p0; \
7391   __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
7392   __ret; \
7393 })
7394 #endif
7395 
7396 #ifdef __LITTLE_ENDIAN__
7397 #define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
7398   uint32_t __ret; \
7399   uint32x4_t __s0 = __p0; \
7400   __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__s0, __p1); \
7401   __ret; \
7402 })
7403 #else
7404 #define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
7405   uint32_t __ret; \
7406   uint32x4_t __s0 = __p0; \
7407   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7408   __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__rev0, __p1); \
7409   __ret; \
7410 })
7411 #define __noswap_vgetq_lane_u32(__p0, __p1) __extension__ ({ \
7412   uint32_t __ret; \
7413   uint32x4_t __s0 = __p0; \
7414   __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__s0, __p1); \
7415   __ret; \
7416 })
7417 #endif
7418 
7419 #ifdef __LITTLE_ENDIAN__
7420 #define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
7421   uint64_t __ret; \
7422   uint64x2_t __s0 = __p0; \
7423   __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__s0, __p1); \
7424   __ret; \
7425 })
7426 #else
7427 #define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
7428   uint64_t __ret; \
7429   uint64x2_t __s0 = __p0; \
7430   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7431   __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__rev0, __p1); \
7432   __ret; \
7433 })
7434 #define __noswap_vgetq_lane_u64(__p0, __p1) __extension__ ({ \
7435   uint64_t __ret; \
7436   uint64x2_t __s0 = __p0; \
7437   __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__s0, __p1); \
7438   __ret; \
7439 })
7440 #endif
7441 
7442 #ifdef __LITTLE_ENDIAN__
7443 #define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
7444   uint16_t __ret; \
7445   uint16x8_t __s0 = __p0; \
7446   __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__s0, __p1); \
7447   __ret; \
7448 })
7449 #else
7450 #define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
7451   uint16_t __ret; \
7452   uint16x8_t __s0 = __p0; \
7453   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7454   __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__rev0, __p1); \
7455   __ret; \
7456 })
7457 #define __noswap_vgetq_lane_u16(__p0, __p1) __extension__ ({ \
7458   uint16_t __ret; \
7459   uint16x8_t __s0 = __p0; \
7460   __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__s0, __p1); \
7461   __ret; \
7462 })
7463 #endif
7464 
7465 #ifdef __LITTLE_ENDIAN__
7466 #define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
7467   int8_t __ret; \
7468   int8x16_t __s0 = __p0; \
7469   __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
7470   __ret; \
7471 })
7472 #else
7473 #define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
7474   int8_t __ret; \
7475   int8x16_t __s0 = __p0; \
7476   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7477   __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
7478   __ret; \
7479 })
7480 #define __noswap_vgetq_lane_s8(__p0, __p1) __extension__ ({ \
7481   int8_t __ret; \
7482   int8x16_t __s0 = __p0; \
7483   __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
7484   __ret; \
7485 })
7486 #endif
7487 
7488 #ifdef __LITTLE_ENDIAN__
7489 #define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
7490   float32_t __ret; \
7491   float32x4_t __s0 = __p0; \
7492   __ret = (float32_t) __builtin_neon_vgetq_lane_f32((float32x4_t)__s0, __p1); \
7493   __ret; \
7494 })
7495 #else
7496 #define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
7497   float32_t __ret; \
7498   float32x4_t __s0 = __p0; \
7499   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7500   __ret = (float32_t) __builtin_neon_vgetq_lane_f32((float32x4_t)__rev0, __p1); \
7501   __ret; \
7502 })
7503 #define __noswap_vgetq_lane_f32(__p0, __p1) __extension__ ({ \
7504   float32_t __ret; \
7505   float32x4_t __s0 = __p0; \
7506   __ret = (float32_t) __builtin_neon_vgetq_lane_f32((float32x4_t)__s0, __p1); \
7507   __ret; \
7508 })
7509 #endif
7510 
7511 #ifdef __LITTLE_ENDIAN__
7512 #define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
7513   int32_t __ret; \
7514   int32x4_t __s0 = __p0; \
7515   __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__s0, __p1); \
7516   __ret; \
7517 })
7518 #else
7519 #define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
7520   int32_t __ret; \
7521   int32x4_t __s0 = __p0; \
7522   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7523   __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__rev0, __p1); \
7524   __ret; \
7525 })
7526 #define __noswap_vgetq_lane_s32(__p0, __p1) __extension__ ({ \
7527   int32_t __ret; \
7528   int32x4_t __s0 = __p0; \
7529   __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__s0, __p1); \
7530   __ret; \
7531 })
7532 #endif
7533 
7534 #ifdef __LITTLE_ENDIAN__
7535 #define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
7536   int64_t __ret; \
7537   int64x2_t __s0 = __p0; \
7538   __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__s0, __p1); \
7539   __ret; \
7540 })
7541 #else
7542 #define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
7543   int64_t __ret; \
7544   int64x2_t __s0 = __p0; \
7545   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7546   __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__rev0, __p1); \
7547   __ret; \
7548 })
7549 #define __noswap_vgetq_lane_s64(__p0, __p1) __extension__ ({ \
7550   int64_t __ret; \
7551   int64x2_t __s0 = __p0; \
7552   __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__s0, __p1); \
7553   __ret; \
7554 })
7555 #endif
7556 
7557 #ifdef __LITTLE_ENDIAN__
7558 #define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
7559   int16_t __ret; \
7560   int16x8_t __s0 = __p0; \
7561   __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__s0, __p1); \
7562   __ret; \
7563 })
7564 #else
7565 #define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
7566   int16_t __ret; \
7567   int16x8_t __s0 = __p0; \
7568   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7569   __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__rev0, __p1); \
7570   __ret; \
7571 })
7572 #define __noswap_vgetq_lane_s16(__p0, __p1) __extension__ ({ \
7573   int16_t __ret; \
7574   int16x8_t __s0 = __p0; \
7575   __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__s0, __p1); \
7576   __ret; \
7577 })
7578 #endif
7579 
7580 #ifdef __LITTLE_ENDIAN__
7581 #define vget_lane_u8(__p0, __p1) __extension__ ({ \
7582   uint8_t __ret; \
7583   uint8x8_t __s0 = __p0; \
7584   __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
7585   __ret; \
7586 })
7587 #else
7588 #define vget_lane_u8(__p0, __p1) __extension__ ({ \
7589   uint8_t __ret; \
7590   uint8x8_t __s0 = __p0; \
7591   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7592   __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
7593   __ret; \
7594 })
7595 #define __noswap_vget_lane_u8(__p0, __p1) __extension__ ({ \
7596   uint8_t __ret; \
7597   uint8x8_t __s0 = __p0; \
7598   __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
7599   __ret; \
7600 })
7601 #endif
7602 
7603 #ifdef __LITTLE_ENDIAN__
7604 #define vget_lane_u32(__p0, __p1) __extension__ ({ \
7605   uint32_t __ret; \
7606   uint32x2_t __s0 = __p0; \
7607   __ret = (uint32_t) __builtin_neon_vget_lane_i32((int32x2_t)__s0, __p1); \
7608   __ret; \
7609 })
7610 #else
7611 #define vget_lane_u32(__p0, __p1) __extension__ ({ \
7612   uint32_t __ret; \
7613   uint32x2_t __s0 = __p0; \
7614   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7615   __ret = (uint32_t) __builtin_neon_vget_lane_i32((int32x2_t)__rev0, __p1); \
7616   __ret; \
7617 })
7618 #define __noswap_vget_lane_u32(__p0, __p1) __extension__ ({ \
7619   uint32_t __ret; \
7620   uint32x2_t __s0 = __p0; \
7621   __ret = (uint32_t) __builtin_neon_vget_lane_i32((int32x2_t)__s0, __p1); \
7622   __ret; \
7623 })
7624 #endif
7625 
7626 #define vget_lane_u64(__p0, __p1) __extension__ ({ \
7627   uint64_t __ret; \
7628   uint64x1_t __s0 = __p0; \
7629   __ret = (uint64_t) __builtin_neon_vget_lane_i64((int64x1_t)__s0, __p1); \
7630   __ret; \
7631 })
7632 #ifdef __LITTLE_ENDIAN__
7633 #define vget_lane_u16(__p0, __p1) __extension__ ({ \
7634   uint16_t __ret; \
7635   uint16x4_t __s0 = __p0; \
7636   __ret = (uint16_t) __builtin_neon_vget_lane_i16((int16x4_t)__s0, __p1); \
7637   __ret; \
7638 })
7639 #else
7640 #define vget_lane_u16(__p0, __p1) __extension__ ({ \
7641   uint16_t __ret; \
7642   uint16x4_t __s0 = __p0; \
7643   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7644   __ret = (uint16_t) __builtin_neon_vget_lane_i16((int16x4_t)__rev0, __p1); \
7645   __ret; \
7646 })
7647 #define __noswap_vget_lane_u16(__p0, __p1) __extension__ ({ \
7648   uint16_t __ret; \
7649   uint16x4_t __s0 = __p0; \
7650   __ret = (uint16_t) __builtin_neon_vget_lane_i16((int16x4_t)__s0, __p1); \
7651   __ret; \
7652 })
7653 #endif
7654 
7655 #ifdef __LITTLE_ENDIAN__
7656 #define vget_lane_s8(__p0, __p1) __extension__ ({ \
7657   int8_t __ret; \
7658   int8x8_t __s0 = __p0; \
7659   __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
7660   __ret; \
7661 })
7662 #else
7663 #define vget_lane_s8(__p0, __p1) __extension__ ({ \
7664   int8_t __ret; \
7665   int8x8_t __s0 = __p0; \
7666   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7667   __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
7668   __ret; \
7669 })
7670 #define __noswap_vget_lane_s8(__p0, __p1) __extension__ ({ \
7671   int8_t __ret; \
7672   int8x8_t __s0 = __p0; \
7673   __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
7674   __ret; \
7675 })
7676 #endif
7677 
7678 #ifdef __LITTLE_ENDIAN__
7679 #define vget_lane_f32(__p0, __p1) __extension__ ({ \
7680   float32_t __ret; \
7681   float32x2_t __s0 = __p0; \
7682   __ret = (float32_t) __builtin_neon_vget_lane_f32((float32x2_t)__s0, __p1); \
7683   __ret; \
7684 })
7685 #else
7686 #define vget_lane_f32(__p0, __p1) __extension__ ({ \
7687   float32_t __ret; \
7688   float32x2_t __s0 = __p0; \
7689   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7690   __ret = (float32_t) __builtin_neon_vget_lane_f32((float32x2_t)__rev0, __p1); \
7691   __ret; \
7692 })
7693 #define __noswap_vget_lane_f32(__p0, __p1) __extension__ ({ \
7694   float32_t __ret; \
7695   float32x2_t __s0 = __p0; \
7696   __ret = (float32_t) __builtin_neon_vget_lane_f32((float32x2_t)__s0, __p1); \
7697   __ret; \
7698 })
7699 #endif
7700 
7701 #ifdef __LITTLE_ENDIAN__
7702 #define vget_lane_s32(__p0, __p1) __extension__ ({ \
7703   int32_t __ret; \
7704   int32x2_t __s0 = __p0; \
7705   __ret = (int32_t) __builtin_neon_vget_lane_i32((int32x2_t)__s0, __p1); \
7706   __ret; \
7707 })
7708 #else
7709 #define vget_lane_s32(__p0, __p1) __extension__ ({ \
7710   int32_t __ret; \
7711   int32x2_t __s0 = __p0; \
7712   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7713   __ret = (int32_t) __builtin_neon_vget_lane_i32((int32x2_t)__rev0, __p1); \
7714   __ret; \
7715 })
7716 #define __noswap_vget_lane_s32(__p0, __p1) __extension__ ({ \
7717   int32_t __ret; \
7718   int32x2_t __s0 = __p0; \
7719   __ret = (int32_t) __builtin_neon_vget_lane_i32((int32x2_t)__s0, __p1); \
7720   __ret; \
7721 })
7722 #endif
7723 
7724 #define vget_lane_s64(__p0, __p1) __extension__ ({ \
7725   int64_t __ret; \
7726   int64x1_t __s0 = __p0; \
7727   __ret = (int64_t) __builtin_neon_vget_lane_i64((int64x1_t)__s0, __p1); \
7728   __ret; \
7729 })
7730 #ifdef __LITTLE_ENDIAN__
7731 #define vget_lane_s16(__p0, __p1) __extension__ ({ \
7732   int16_t __ret; \
7733   int16x4_t __s0 = __p0; \
7734   __ret = (int16_t) __builtin_neon_vget_lane_i16((int16x4_t)__s0, __p1); \
7735   __ret; \
7736 })
7737 #else
7738 #define vget_lane_s16(__p0, __p1) __extension__ ({ \
7739   int16_t __ret; \
7740   int16x4_t __s0 = __p0; \
7741   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7742   __ret = (int16_t) __builtin_neon_vget_lane_i16((int16x4_t)__rev0, __p1); \
7743   __ret; \
7744 })
7745 #define __noswap_vget_lane_s16(__p0, __p1) __extension__ ({ \
7746   int16_t __ret; \
7747   int16x4_t __s0 = __p0; \
7748   __ret = (int16_t) __builtin_neon_vget_lane_i16((int16x4_t)__s0, __p1); \
7749   __ret; \
7750 })
7751 #endif
7752 
7753 #ifdef __LITTLE_ENDIAN__
vget_low_p8(poly8x16_t __p0)7754 __ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
7755   poly8x8_t __ret;
7756   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
7757   return __ret;
7758 }
7759 #else
vget_low_p8(poly8x16_t __p0)7760 __ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
7761   poly8x8_t __ret;
7762   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7763   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
7764   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7765   return __ret;
7766 }
7767 #endif
7768 
7769 #ifdef __LITTLE_ENDIAN__
vget_low_p16(poly16x8_t __p0)7770 __ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
7771   poly16x4_t __ret;
7772   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7773   return __ret;
7774 }
7775 #else
vget_low_p16(poly16x8_t __p0)7776 __ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
7777   poly16x4_t __ret;
7778   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7779   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7780   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7781   return __ret;
7782 }
7783 #endif
7784 
7785 #ifdef __LITTLE_ENDIAN__
vget_low_u8(uint8x16_t __p0)7786 __ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
7787   uint8x8_t __ret;
7788   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
7789   return __ret;
7790 }
7791 #else
vget_low_u8(uint8x16_t __p0)7792 __ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
7793   uint8x8_t __ret;
7794   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7795   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
7796   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7797   return __ret;
7798 }
7799 #endif
7800 
7801 #ifdef __LITTLE_ENDIAN__
vget_low_u32(uint32x4_t __p0)7802 __ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
7803   uint32x2_t __ret;
7804   __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7805   return __ret;
7806 }
7807 #else
vget_low_u32(uint32x4_t __p0)7808 __ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
7809   uint32x2_t __ret;
7810   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7811   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7812   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7813   return __ret;
7814 }
7815 #endif
7816 
7817 #ifdef __LITTLE_ENDIAN__
vget_low_u64(uint64x2_t __p0)7818 __ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7819   uint64x1_t __ret;
7820   __ret = __builtin_shufflevector(__p0, __p0, 0);
7821   return __ret;
7822 }
7823 #else
vget_low_u64(uint64x2_t __p0)7824 __ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7825   uint64x1_t __ret;
7826   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7827   __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7828   return __ret;
7829 }
7830 #endif
7831 
7832 #ifdef __LITTLE_ENDIAN__
vget_low_u16(uint16x8_t __p0)7833 __ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7834   uint16x4_t __ret;
7835   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7836   return __ret;
7837 }
7838 #else
vget_low_u16(uint16x8_t __p0)7839 __ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7840   uint16x4_t __ret;
7841   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7842   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7843   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7844   return __ret;
7845 }
7846 #endif
7847 
7848 #ifdef __LITTLE_ENDIAN__
vget_low_s8(int8x16_t __p0)7849 __ai int8x8_t vget_low_s8(int8x16_t __p0) {
7850   int8x8_t __ret;
7851   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
7852   return __ret;
7853 }
7854 #else
vget_low_s8(int8x16_t __p0)7855 __ai int8x8_t vget_low_s8(int8x16_t __p0) {
7856   int8x8_t __ret;
7857   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7858   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
7859   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7860   return __ret;
7861 }
7862 #endif
7863 
7864 #ifdef __LITTLE_ENDIAN__
vget_low_f32(float32x4_t __p0)7865 __ai float32x2_t vget_low_f32(float32x4_t __p0) {
7866   float32x2_t __ret;
7867   __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7868   return __ret;
7869 }
7870 #else
vget_low_f32(float32x4_t __p0)7871 __ai float32x2_t vget_low_f32(float32x4_t __p0) {
7872   float32x2_t __ret;
7873   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7874   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7875   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7876   return __ret;
7877 }
7878 #endif
7879 
7880 #ifdef __LITTLE_ENDIAN__
vget_low_f16(float16x8_t __p0)7881 __ai float16x4_t vget_low_f16(float16x8_t __p0) {
7882   float16x4_t __ret;
7883   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7884   return __ret;
7885 }
7886 #else
vget_low_f16(float16x8_t __p0)7887 __ai float16x4_t vget_low_f16(float16x8_t __p0) {
7888   float16x4_t __ret;
7889   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7890   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7891   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7892   return __ret;
7893 }
7894 #endif
7895 
7896 #ifdef __LITTLE_ENDIAN__
vget_low_s32(int32x4_t __p0)7897 __ai int32x2_t vget_low_s32(int32x4_t __p0) {
7898   int32x2_t __ret;
7899   __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7900   return __ret;
7901 }
7902 #else
vget_low_s32(int32x4_t __p0)7903 __ai int32x2_t vget_low_s32(int32x4_t __p0) {
7904   int32x2_t __ret;
7905   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7906   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7907   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7908   return __ret;
7909 }
7910 #endif
7911 
7912 #ifdef __LITTLE_ENDIAN__
vget_low_s64(int64x2_t __p0)7913 __ai int64x1_t vget_low_s64(int64x2_t __p0) {
7914   int64x1_t __ret;
7915   __ret = __builtin_shufflevector(__p0, __p0, 0);
7916   return __ret;
7917 }
7918 #else
vget_low_s64(int64x2_t __p0)7919 __ai int64x1_t vget_low_s64(int64x2_t __p0) {
7920   int64x1_t __ret;
7921   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7922   __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7923   return __ret;
7924 }
7925 #endif
7926 
7927 #ifdef __LITTLE_ENDIAN__
vget_low_s16(int16x8_t __p0)7928 __ai int16x4_t vget_low_s16(int16x8_t __p0) {
7929   int16x4_t __ret;
7930   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7931   return __ret;
7932 }
7933 #else
vget_low_s16(int16x8_t __p0)7934 __ai int16x4_t vget_low_s16(int16x8_t __p0) {
7935   int16x4_t __ret;
7936   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7937   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7938   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7939   return __ret;
7940 }
7941 #endif
7942 
7943 #ifdef __LITTLE_ENDIAN__
vhaddq_u8(uint8x16_t __p0,uint8x16_t __p1)7944 __ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7945   uint8x16_t __ret;
7946   __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
7947   return __ret;
7948 }
7949 #else
vhaddq_u8(uint8x16_t __p0,uint8x16_t __p1)7950 __ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7951   uint8x16_t __ret;
7952   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7953   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7954   __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
7955   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7956   return __ret;
7957 }
7958 #endif
7959 
7960 #ifdef __LITTLE_ENDIAN__
vhaddq_u32(uint32x4_t __p0,uint32x4_t __p1)7961 __ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7962   uint32x4_t __ret;
7963   __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
7964   return __ret;
7965 }
7966 #else
vhaddq_u32(uint32x4_t __p0,uint32x4_t __p1)7967 __ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7968   uint32x4_t __ret;
7969   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7970   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7971   __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
7972   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7973   return __ret;
7974 }
7975 #endif
7976 
7977 #ifdef __LITTLE_ENDIAN__
vhaddq_u16(uint16x8_t __p0,uint16x8_t __p1)7978 __ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7979   uint16x8_t __ret;
7980   __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
7981   return __ret;
7982 }
7983 #else
vhaddq_u16(uint16x8_t __p0,uint16x8_t __p1)7984 __ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7985   uint16x8_t __ret;
7986   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7987   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7988   __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
7989   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7990   return __ret;
7991 }
7992 #endif
7993 
7994 #ifdef __LITTLE_ENDIAN__
vhaddq_s8(int8x16_t __p0,int8x16_t __p1)7995 __ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
7996   int8x16_t __ret;
7997   __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
7998   return __ret;
7999 }
8000 #else
vhaddq_s8(int8x16_t __p0,int8x16_t __p1)8001 __ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
8002   int8x16_t __ret;
8003   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8004   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8005   __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
8006   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8007   return __ret;
8008 }
8009 #endif
8010 
8011 #ifdef __LITTLE_ENDIAN__
vhaddq_s32(int32x4_t __p0,int32x4_t __p1)8012 __ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
8013   int32x4_t __ret;
8014   __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
8015   return __ret;
8016 }
8017 #else
vhaddq_s32(int32x4_t __p0,int32x4_t __p1)8018 __ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
8019   int32x4_t __ret;
8020   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8021   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8022   __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
8023   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8024   return __ret;
8025 }
8026 #endif
8027 
8028 #ifdef __LITTLE_ENDIAN__
vhaddq_s16(int16x8_t __p0,int16x8_t __p1)8029 __ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
8030   int16x8_t __ret;
8031   __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
8032   return __ret;
8033 }
8034 #else
vhaddq_s16(int16x8_t __p0,int16x8_t __p1)8035 __ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
8036   int16x8_t __ret;
8037   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8038   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8039   __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
8040   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8041   return __ret;
8042 }
8043 #endif
8044 
8045 #ifdef __LITTLE_ENDIAN__
vhadd_u8(uint8x8_t __p0,uint8x8_t __p1)8046 __ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
8047   uint8x8_t __ret;
8048   __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
8049   return __ret;
8050 }
8051 #else
vhadd_u8(uint8x8_t __p0,uint8x8_t __p1)8052 __ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
8053   uint8x8_t __ret;
8054   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8055   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8056   __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
8057   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8058   return __ret;
8059 }
8060 #endif
8061 
8062 #ifdef __LITTLE_ENDIAN__
vhadd_u32(uint32x2_t __p0,uint32x2_t __p1)8063 __ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
8064   uint32x2_t __ret;
8065   __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
8066   return __ret;
8067 }
8068 #else
vhadd_u32(uint32x2_t __p0,uint32x2_t __p1)8069 __ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
8070   uint32x2_t __ret;
8071   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
8072   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
8073   __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
8074   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
8075   return __ret;
8076 }
8077 #endif
8078 
8079 #ifdef __LITTLE_ENDIAN__
vhadd_u16(uint16x4_t __p0,uint16x4_t __p1)8080 __ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
8081   uint16x4_t __ret;
8082   __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
8083   return __ret;
8084 }
8085 #else
vhadd_u16(uint16x4_t __p0,uint16x4_t __p1)8086 __ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
8087   uint16x4_t __ret;
8088   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8089   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8090   __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
8091   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8092   return __ret;
8093 }
8094 #endif
8095 
8096 #ifdef __LITTLE_ENDIAN__
vhadd_s8(int8x8_t __p0,int8x8_t __p1)8097 __ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
8098   int8x8_t __ret;
8099   __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
8100   return __ret;
8101 }
8102 #else
vhadd_s8(int8x8_t __p0,int8x8_t __p1)8103 __ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
8104   int8x8_t __ret;
8105   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8106   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8107   __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
8108   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8109   return __ret;
8110 }
8111 #endif
8112 
8113 #ifdef __LITTLE_ENDIAN__
vhadd_s32(int32x2_t __p0,int32x2_t __p1)8114 __ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
8115   int32x2_t __ret;
8116   __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
8117   return __ret;
8118 }
8119 #else
vhadd_s32(int32x2_t __p0,int32x2_t __p1)8120 __ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
8121   int32x2_t __ret;
8122   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
8123   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
8124   __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
8125   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
8126   return __ret;
8127 }
8128 #endif
8129 
8130 #ifdef __LITTLE_ENDIAN__
vhadd_s16(int16x4_t __p0,int16x4_t __p1)8131 __ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
8132   int16x4_t __ret;
8133   __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
8134   return __ret;
8135 }
8136 #else
vhadd_s16(int16x4_t __p0,int16x4_t __p1)8137 __ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
8138   int16x4_t __ret;
8139   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8140   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8141   __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
8142   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8143   return __ret;
8144 }
8145 #endif
8146 
8147 #ifdef __LITTLE_ENDIAN__
vhsubq_u8(uint8x16_t __p0,uint8x16_t __p1)8148 __ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
8149   uint8x16_t __ret;
8150   __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
8151   return __ret;
8152 }
8153 #else
vhsubq_u8(uint8x16_t __p0,uint8x16_t __p1)8154 __ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
8155   uint8x16_t __ret;
8156   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8157   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8158   __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
8159   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8160   return __ret;
8161 }
8162 #endif
8163 
8164 #ifdef __LITTLE_ENDIAN__
vhsubq_u32(uint32x4_t __p0,uint32x4_t __p1)8165 __ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
8166   uint32x4_t __ret;
8167   __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
8168   return __ret;
8169 }
8170 #else
vhsubq_u32(uint32x4_t __p0,uint32x4_t __p1)8171 __ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
8172   uint32x4_t __ret;
8173   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8174   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8175   __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
8176   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8177   return __ret;
8178 }
8179 #endif
8180 
8181 #ifdef __LITTLE_ENDIAN__
vhsubq_u16(uint16x8_t __p0,uint16x8_t __p1)8182 __ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
8183   uint16x8_t __ret;
8184   __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
8185   return __ret;
8186 }
8187 #else
vhsubq_u16(uint16x8_t __p0,uint16x8_t __p1)8188 __ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
8189   uint16x8_t __ret;
8190   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8191   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8192   __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
8193   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8194   return __ret;
8195 }
8196 #endif
8197 
8198 #ifdef __LITTLE_ENDIAN__
vhsubq_s8(int8x16_t __p0,int8x16_t __p1)8199 __ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
8200   int8x16_t __ret;
8201   __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
8202   return __ret;
8203 }
8204 #else
vhsubq_s8(int8x16_t __p0,int8x16_t __p1)8205 __ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
8206   int8x16_t __ret;
8207   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8208   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8209   __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
8210   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8211   return __ret;
8212 }
8213 #endif
8214 
8215 #ifdef __LITTLE_ENDIAN__
vhsubq_s32(int32x4_t __p0,int32x4_t __p1)8216 __ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
8217   int32x4_t __ret;
8218   __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
8219   return __ret;
8220 }
8221 #else
vhsubq_s32(int32x4_t __p0,int32x4_t __p1)8222 __ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
8223   int32x4_t __ret;
8224   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8225   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8226   __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
8227   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8228   return __ret;
8229 }
8230 #endif
8231 
8232 #ifdef __LITTLE_ENDIAN__
vhsubq_s16(int16x8_t __p0,int16x8_t __p1)8233 __ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
8234   int16x8_t __ret;
8235   __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
8236   return __ret;
8237 }
8238 #else
vhsubq_s16(int16x8_t __p0,int16x8_t __p1)8239 __ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
8240   int16x8_t __ret;
8241   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8242   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8243   __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
8244   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8245   return __ret;
8246 }
8247 #endif
8248 
8249 #ifdef __LITTLE_ENDIAN__
vhsub_u8(uint8x8_t __p0,uint8x8_t __p1)8250 __ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
8251   uint8x8_t __ret;
8252   __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
8253   return __ret;
8254 }
8255 #else
vhsub_u8(uint8x8_t __p0,uint8x8_t __p1)8256 __ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
8257   uint8x8_t __ret;
8258   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8259   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8260   __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
8261   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8262   return __ret;
8263 }
8264 #endif
8265 
8266 #ifdef __LITTLE_ENDIAN__
vhsub_u32(uint32x2_t __p0,uint32x2_t __p1)8267 __ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
8268   uint32x2_t __ret;
8269   __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
8270   return __ret;
8271 }
8272 #else
vhsub_u32(uint32x2_t __p0,uint32x2_t __p1)8273 __ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
8274   uint32x2_t __ret;
8275   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
8276   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
8277   __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
8278   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
8279   return __ret;
8280 }
8281 #endif
8282 
8283 #ifdef __LITTLE_ENDIAN__
vhsub_u16(uint16x4_t __p0,uint16x4_t __p1)8284 __ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
8285   uint16x4_t __ret;
8286   __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
8287   return __ret;
8288 }
8289 #else
vhsub_u16(uint16x4_t __p0,uint16x4_t __p1)8290 __ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
8291   uint16x4_t __ret;
8292   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8293   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8294   __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
8295   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8296   return __ret;
8297 }
8298 #endif
8299 
8300 #ifdef __LITTLE_ENDIAN__
vhsub_s8(int8x8_t __p0,int8x8_t __p1)8301 __ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
8302   int8x8_t __ret;
8303   __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
8304   return __ret;
8305 }
8306 #else
vhsub_s8(int8x8_t __p0,int8x8_t __p1)8307 __ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
8308   int8x8_t __ret;
8309   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8310   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8311   __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
8312   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8313   return __ret;
8314 }
8315 #endif
8316 
8317 #ifdef __LITTLE_ENDIAN__
vhsub_s32(int32x2_t __p0,int32x2_t __p1)8318 __ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
8319   int32x2_t __ret;
8320   __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
8321   return __ret;
8322 }
8323 #else
vhsub_s32(int32x2_t __p0,int32x2_t __p1)8324 __ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
8325   int32x2_t __ret;
8326   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
8327   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
8328   __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
8329   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
8330   return __ret;
8331 }
8332 #endif
8333 
8334 #ifdef __LITTLE_ENDIAN__
vhsub_s16(int16x4_t __p0,int16x4_t __p1)8335 __ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
8336   int16x4_t __ret;
8337   __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
8338   return __ret;
8339 }
8340 #else
vhsub_s16(int16x4_t __p0,int16x4_t __p1)8341 __ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
8342   int16x4_t __ret;
8343   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8344   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8345   __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
8346   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8347   return __ret;
8348 }
8349 #endif
8350 
8351 #ifdef __LITTLE_ENDIAN__
8352 #define vld1_p8(__p0) __extension__ ({ \
8353   poly8x8_t __ret; \
8354   __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
8355   __ret; \
8356 })
8357 #else
8358 #define vld1_p8(__p0) __extension__ ({ \
8359   poly8x8_t __ret; \
8360   __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
8361   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8362   __ret; \
8363 })
8364 #endif
8365 
8366 #ifdef __LITTLE_ENDIAN__
8367 #define vld1_p16(__p0) __extension__ ({ \
8368   poly16x4_t __ret; \
8369   __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
8370   __ret; \
8371 })
8372 #else
8373 #define vld1_p16(__p0) __extension__ ({ \
8374   poly16x4_t __ret; \
8375   __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
8376   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8377   __ret; \
8378 })
8379 #endif
8380 
8381 #ifdef __LITTLE_ENDIAN__
8382 #define vld1q_p8(__p0) __extension__ ({ \
8383   poly8x16_t __ret; \
8384   __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
8385   __ret; \
8386 })
8387 #else
8388 #define vld1q_p8(__p0) __extension__ ({ \
8389   poly8x16_t __ret; \
8390   __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
8391   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8392   __ret; \
8393 })
8394 #endif
8395 
8396 #ifdef __LITTLE_ENDIAN__
8397 #define vld1q_p16(__p0) __extension__ ({ \
8398   poly16x8_t __ret; \
8399   __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
8400   __ret; \
8401 })
8402 #else
8403 #define vld1q_p16(__p0) __extension__ ({ \
8404   poly16x8_t __ret; \
8405   __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
8406   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8407   __ret; \
8408 })
8409 #endif
8410 
8411 #ifdef __LITTLE_ENDIAN__
8412 #define vld1q_u8(__p0) __extension__ ({ \
8413   uint8x16_t __ret; \
8414   __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
8415   __ret; \
8416 })
8417 #else
8418 #define vld1q_u8(__p0) __extension__ ({ \
8419   uint8x16_t __ret; \
8420   __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
8421   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8422   __ret; \
8423 })
8424 #endif
8425 
8426 #ifdef __LITTLE_ENDIAN__
8427 #define vld1q_u32(__p0) __extension__ ({ \
8428   uint32x4_t __ret; \
8429   __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
8430   __ret; \
8431 })
8432 #else
8433 #define vld1q_u32(__p0) __extension__ ({ \
8434   uint32x4_t __ret; \
8435   __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
8436   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8437   __ret; \
8438 })
8439 #endif
8440 
8441 #ifdef __LITTLE_ENDIAN__
8442 #define vld1q_u64(__p0) __extension__ ({ \
8443   uint64x2_t __ret; \
8444   __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
8445   __ret; \
8446 })
8447 #else
8448 #define vld1q_u64(__p0) __extension__ ({ \
8449   uint64x2_t __ret; \
8450   __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
8451   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8452   __ret; \
8453 })
8454 #endif
8455 
8456 #ifdef __LITTLE_ENDIAN__
8457 #define vld1q_u16(__p0) __extension__ ({ \
8458   uint16x8_t __ret; \
8459   __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
8460   __ret; \
8461 })
8462 #else
8463 #define vld1q_u16(__p0) __extension__ ({ \
8464   uint16x8_t __ret; \
8465   __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
8466   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8467   __ret; \
8468 })
8469 #endif
8470 
8471 #ifdef __LITTLE_ENDIAN__
8472 #define vld1q_s8(__p0) __extension__ ({ \
8473   int8x16_t __ret; \
8474   __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
8475   __ret; \
8476 })
8477 #else
8478 #define vld1q_s8(__p0) __extension__ ({ \
8479   int8x16_t __ret; \
8480   __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
8481   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8482   __ret; \
8483 })
8484 #endif
8485 
8486 #ifdef __LITTLE_ENDIAN__
8487 #define vld1q_f32(__p0) __extension__ ({ \
8488   float32x4_t __ret; \
8489   __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
8490   __ret; \
8491 })
8492 #else
8493 #define vld1q_f32(__p0) __extension__ ({ \
8494   float32x4_t __ret; \
8495   __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
8496   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8497   __ret; \
8498 })
8499 #endif
8500 
8501 #ifdef __LITTLE_ENDIAN__
8502 #define vld1q_s32(__p0) __extension__ ({ \
8503   int32x4_t __ret; \
8504   __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
8505   __ret; \
8506 })
8507 #else
8508 #define vld1q_s32(__p0) __extension__ ({ \
8509   int32x4_t __ret; \
8510   __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
8511   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8512   __ret; \
8513 })
8514 #endif
8515 
8516 #ifdef __LITTLE_ENDIAN__
8517 #define vld1q_s64(__p0) __extension__ ({ \
8518   int64x2_t __ret; \
8519   __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
8520   __ret; \
8521 })
8522 #else
8523 #define vld1q_s64(__p0) __extension__ ({ \
8524   int64x2_t __ret; \
8525   __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
8526   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8527   __ret; \
8528 })
8529 #endif
8530 
8531 #ifdef __LITTLE_ENDIAN__
8532 #define vld1q_s16(__p0) __extension__ ({ \
8533   int16x8_t __ret; \
8534   __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
8535   __ret; \
8536 })
8537 #else
8538 #define vld1q_s16(__p0) __extension__ ({ \
8539   int16x8_t __ret; \
8540   __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
8541   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8542   __ret; \
8543 })
8544 #endif
8545 
8546 #ifdef __LITTLE_ENDIAN__
8547 #define vld1_u8(__p0) __extension__ ({ \
8548   uint8x8_t __ret; \
8549   __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
8550   __ret; \
8551 })
8552 #else
8553 #define vld1_u8(__p0) __extension__ ({ \
8554   uint8x8_t __ret; \
8555   __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
8556   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8557   __ret; \
8558 })
8559 #endif
8560 
8561 #ifdef __LITTLE_ENDIAN__
8562 #define vld1_u32(__p0) __extension__ ({ \
8563   uint32x2_t __ret; \
8564   __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
8565   __ret; \
8566 })
8567 #else
8568 #define vld1_u32(__p0) __extension__ ({ \
8569   uint32x2_t __ret; \
8570   __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
8571   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8572   __ret; \
8573 })
8574 #endif
8575 
8576 #define vld1_u64(__p0) __extension__ ({ \
8577   uint64x1_t __ret; \
8578   __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
8579   __ret; \
8580 })
8581 #ifdef __LITTLE_ENDIAN__
8582 #define vld1_u16(__p0) __extension__ ({ \
8583   uint16x4_t __ret; \
8584   __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
8585   __ret; \
8586 })
8587 #else
8588 #define vld1_u16(__p0) __extension__ ({ \
8589   uint16x4_t __ret; \
8590   __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
8591   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8592   __ret; \
8593 })
8594 #endif
8595 
8596 #ifdef __LITTLE_ENDIAN__
8597 #define vld1_s8(__p0) __extension__ ({ \
8598   int8x8_t __ret; \
8599   __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
8600   __ret; \
8601 })
8602 #else
8603 #define vld1_s8(__p0) __extension__ ({ \
8604   int8x8_t __ret; \
8605   __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
8606   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8607   __ret; \
8608 })
8609 #endif
8610 
8611 #ifdef __LITTLE_ENDIAN__
8612 #define vld1_f32(__p0) __extension__ ({ \
8613   float32x2_t __ret; \
8614   __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
8615   __ret; \
8616 })
8617 #else
8618 #define vld1_f32(__p0) __extension__ ({ \
8619   float32x2_t __ret; \
8620   __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
8621   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8622   __ret; \
8623 })
8624 #endif
8625 
8626 #ifdef __LITTLE_ENDIAN__
8627 #define vld1_s32(__p0) __extension__ ({ \
8628   int32x2_t __ret; \
8629   __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
8630   __ret; \
8631 })
8632 #else
8633 #define vld1_s32(__p0) __extension__ ({ \
8634   int32x2_t __ret; \
8635   __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
8636   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8637   __ret; \
8638 })
8639 #endif
8640 
8641 #define vld1_s64(__p0) __extension__ ({ \
8642   int64x1_t __ret; \
8643   __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
8644   __ret; \
8645 })
8646 #ifdef __LITTLE_ENDIAN__
8647 #define vld1_s16(__p0) __extension__ ({ \
8648   int16x4_t __ret; \
8649   __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
8650   __ret; \
8651 })
8652 #else
8653 #define vld1_s16(__p0) __extension__ ({ \
8654   int16x4_t __ret; \
8655   __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
8656   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8657   __ret; \
8658 })
8659 #endif
8660 
8661 #ifdef __LITTLE_ENDIAN__
8662 #define vld1_dup_p8(__p0) __extension__ ({ \
8663   poly8x8_t __ret; \
8664   __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
8665   __ret; \
8666 })
8667 #else
8668 #define vld1_dup_p8(__p0) __extension__ ({ \
8669   poly8x8_t __ret; \
8670   __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
8671   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8672   __ret; \
8673 })
8674 #endif
8675 
8676 #ifdef __LITTLE_ENDIAN__
8677 #define vld1_dup_p16(__p0) __extension__ ({ \
8678   poly16x4_t __ret; \
8679   __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
8680   __ret; \
8681 })
8682 #else
8683 #define vld1_dup_p16(__p0) __extension__ ({ \
8684   poly16x4_t __ret; \
8685   __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
8686   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8687   __ret; \
8688 })
8689 #endif
8690 
8691 #ifdef __LITTLE_ENDIAN__
8692 #define vld1q_dup_p8(__p0) __extension__ ({ \
8693   poly8x16_t __ret; \
8694   __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
8695   __ret; \
8696 })
8697 #else
8698 #define vld1q_dup_p8(__p0) __extension__ ({ \
8699   poly8x16_t __ret; \
8700   __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
8701   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8702   __ret; \
8703 })
8704 #endif
8705 
8706 #ifdef __LITTLE_ENDIAN__
8707 #define vld1q_dup_p16(__p0) __extension__ ({ \
8708   poly16x8_t __ret; \
8709   __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
8710   __ret; \
8711 })
8712 #else
8713 #define vld1q_dup_p16(__p0) __extension__ ({ \
8714   poly16x8_t __ret; \
8715   __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
8716   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8717   __ret; \
8718 })
8719 #endif
8720 
8721 #ifdef __LITTLE_ENDIAN__
8722 #define vld1q_dup_u8(__p0) __extension__ ({ \
8723   uint8x16_t __ret; \
8724   __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
8725   __ret; \
8726 })
8727 #else
8728 #define vld1q_dup_u8(__p0) __extension__ ({ \
8729   uint8x16_t __ret; \
8730   __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
8731   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8732   __ret; \
8733 })
8734 #endif
8735 
8736 #ifdef __LITTLE_ENDIAN__
8737 #define vld1q_dup_u32(__p0) __extension__ ({ \
8738   uint32x4_t __ret; \
8739   __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
8740   __ret; \
8741 })
8742 #else
8743 #define vld1q_dup_u32(__p0) __extension__ ({ \
8744   uint32x4_t __ret; \
8745   __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
8746   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8747   __ret; \
8748 })
8749 #endif
8750 
8751 #ifdef __LITTLE_ENDIAN__
8752 #define vld1q_dup_u64(__p0) __extension__ ({ \
8753   uint64x2_t __ret; \
8754   __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
8755   __ret; \
8756 })
8757 #else
8758 #define vld1q_dup_u64(__p0) __extension__ ({ \
8759   uint64x2_t __ret; \
8760   __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
8761   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8762   __ret; \
8763 })
8764 #endif
8765 
8766 #ifdef __LITTLE_ENDIAN__
8767 #define vld1q_dup_u16(__p0) __extension__ ({ \
8768   uint16x8_t __ret; \
8769   __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8770   __ret; \
8771 })
8772 #else
8773 #define vld1q_dup_u16(__p0) __extension__ ({ \
8774   uint16x8_t __ret; \
8775   __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8776   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8777   __ret; \
8778 })
8779 #endif
8780 
8781 #ifdef __LITTLE_ENDIAN__
8782 #define vld1q_dup_s8(__p0) __extension__ ({ \
8783   int8x16_t __ret; \
8784   __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8785   __ret; \
8786 })
8787 #else
8788 #define vld1q_dup_s8(__p0) __extension__ ({ \
8789   int8x16_t __ret; \
8790   __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8791   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8792   __ret; \
8793 })
8794 #endif
8795 
8796 #ifdef __LITTLE_ENDIAN__
8797 #define vld1q_dup_f32(__p0) __extension__ ({ \
8798   float32x4_t __ret; \
8799   __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8800   __ret; \
8801 })
8802 #else
8803 #define vld1q_dup_f32(__p0) __extension__ ({ \
8804   float32x4_t __ret; \
8805   __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8806   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8807   __ret; \
8808 })
8809 #endif
8810 
8811 #ifdef __LITTLE_ENDIAN__
8812 #define vld1q_dup_s32(__p0) __extension__ ({ \
8813   int32x4_t __ret; \
8814   __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8815   __ret; \
8816 })
8817 #else
8818 #define vld1q_dup_s32(__p0) __extension__ ({ \
8819   int32x4_t __ret; \
8820   __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8821   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8822   __ret; \
8823 })
8824 #endif
8825 
8826 #ifdef __LITTLE_ENDIAN__
8827 #define vld1q_dup_s64(__p0) __extension__ ({ \
8828   int64x2_t __ret; \
8829   __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8830   __ret; \
8831 })
8832 #else
8833 #define vld1q_dup_s64(__p0) __extension__ ({ \
8834   int64x2_t __ret; \
8835   __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8836   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8837   __ret; \
8838 })
8839 #endif
8840 
8841 #ifdef __LITTLE_ENDIAN__
8842 #define vld1q_dup_s16(__p0) __extension__ ({ \
8843   int16x8_t __ret; \
8844   __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8845   __ret; \
8846 })
8847 #else
8848 #define vld1q_dup_s16(__p0) __extension__ ({ \
8849   int16x8_t __ret; \
8850   __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8851   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8852   __ret; \
8853 })
8854 #endif
8855 
8856 #ifdef __LITTLE_ENDIAN__
8857 #define vld1_dup_u8(__p0) __extension__ ({ \
8858   uint8x8_t __ret; \
8859   __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8860   __ret; \
8861 })
8862 #else
8863 #define vld1_dup_u8(__p0) __extension__ ({ \
8864   uint8x8_t __ret; \
8865   __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8866   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8867   __ret; \
8868 })
8869 #endif
8870 
8871 #ifdef __LITTLE_ENDIAN__
8872 #define vld1_dup_u32(__p0) __extension__ ({ \
8873   uint32x2_t __ret; \
8874   __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8875   __ret; \
8876 })
8877 #else
8878 #define vld1_dup_u32(__p0) __extension__ ({ \
8879   uint32x2_t __ret; \
8880   __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8881   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8882   __ret; \
8883 })
8884 #endif
8885 
8886 #define vld1_dup_u64(__p0) __extension__ ({ \
8887   uint64x1_t __ret; \
8888   __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
8889   __ret; \
8890 })
8891 #ifdef __LITTLE_ENDIAN__
8892 #define vld1_dup_u16(__p0) __extension__ ({ \
8893   uint16x4_t __ret; \
8894   __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8895   __ret; \
8896 })
8897 #else
8898 #define vld1_dup_u16(__p0) __extension__ ({ \
8899   uint16x4_t __ret; \
8900   __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8901   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8902   __ret; \
8903 })
8904 #endif
8905 
8906 #ifdef __LITTLE_ENDIAN__
8907 #define vld1_dup_s8(__p0) __extension__ ({ \
8908   int8x8_t __ret; \
8909   __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8910   __ret; \
8911 })
8912 #else
8913 #define vld1_dup_s8(__p0) __extension__ ({ \
8914   int8x8_t __ret; \
8915   __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8916   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8917   __ret; \
8918 })
8919 #endif
8920 
8921 #ifdef __LITTLE_ENDIAN__
8922 #define vld1_dup_f32(__p0) __extension__ ({ \
8923   float32x2_t __ret; \
8924   __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8925   __ret; \
8926 })
8927 #else
8928 #define vld1_dup_f32(__p0) __extension__ ({ \
8929   float32x2_t __ret; \
8930   __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8931   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8932   __ret; \
8933 })
8934 #endif
8935 
8936 #ifdef __LITTLE_ENDIAN__
8937 #define vld1_dup_s32(__p0) __extension__ ({ \
8938   int32x2_t __ret; \
8939   __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8940   __ret; \
8941 })
8942 #else
8943 #define vld1_dup_s32(__p0) __extension__ ({ \
8944   int32x2_t __ret; \
8945   __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8946   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8947   __ret; \
8948 })
8949 #endif
8950 
8951 #define vld1_dup_s64(__p0) __extension__ ({ \
8952   int64x1_t __ret; \
8953   __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
8954   __ret; \
8955 })
8956 #ifdef __LITTLE_ENDIAN__
8957 #define vld1_dup_s16(__p0) __extension__ ({ \
8958   int16x4_t __ret; \
8959   __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8960   __ret; \
8961 })
8962 #else
8963 #define vld1_dup_s16(__p0) __extension__ ({ \
8964   int16x4_t __ret; \
8965   __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8966   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8967   __ret; \
8968 })
8969 #endif
8970 
8971 #ifdef __LITTLE_ENDIAN__
8972 #define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8973   poly8x8_t __ret; \
8974   poly8x8_t __s1 = __p1; \
8975   __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
8976   __ret; \
8977 })
8978 #else
8979 #define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8980   poly8x8_t __ret; \
8981   poly8x8_t __s1 = __p1; \
8982   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8983   __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
8984   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8985   __ret; \
8986 })
8987 #endif
8988 
8989 #ifdef __LITTLE_ENDIAN__
8990 #define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8991   poly16x4_t __ret; \
8992   poly16x4_t __s1 = __p1; \
8993   __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
8994   __ret; \
8995 })
8996 #else
8997 #define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8998   poly16x4_t __ret; \
8999   poly16x4_t __s1 = __p1; \
9000   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9001   __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
9002   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9003   __ret; \
9004 })
9005 #endif
9006 
9007 #ifdef __LITTLE_ENDIAN__
9008 #define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9009   poly8x16_t __ret; \
9010   poly8x16_t __s1 = __p1; \
9011   __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
9012   __ret; \
9013 })
9014 #else
9015 #define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9016   poly8x16_t __ret; \
9017   poly8x16_t __s1 = __p1; \
9018   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9019   __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
9020   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9021   __ret; \
9022 })
9023 #endif
9024 
9025 #ifdef __LITTLE_ENDIAN__
9026 #define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9027   poly16x8_t __ret; \
9028   poly16x8_t __s1 = __p1; \
9029   __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
9030   __ret; \
9031 })
9032 #else
9033 #define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9034   poly16x8_t __ret; \
9035   poly16x8_t __s1 = __p1; \
9036   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9037   __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
9038   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9039   __ret; \
9040 })
9041 #endif
9042 
9043 #ifdef __LITTLE_ENDIAN__
9044 #define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9045   uint8x16_t __ret; \
9046   uint8x16_t __s1 = __p1; \
9047   __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
9048   __ret; \
9049 })
9050 #else
9051 #define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9052   uint8x16_t __ret; \
9053   uint8x16_t __s1 = __p1; \
9054   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9055   __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
9056   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9057   __ret; \
9058 })
9059 #endif
9060 
9061 #ifdef __LITTLE_ENDIAN__
9062 #define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9063   uint32x4_t __ret; \
9064   uint32x4_t __s1 = __p1; \
9065   __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
9066   __ret; \
9067 })
9068 #else
9069 #define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9070   uint32x4_t __ret; \
9071   uint32x4_t __s1 = __p1; \
9072   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9073   __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
9074   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9075   __ret; \
9076 })
9077 #endif
9078 
9079 #ifdef __LITTLE_ENDIAN__
9080 #define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
9081   uint64x2_t __ret; \
9082   uint64x2_t __s1 = __p1; \
9083   __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
9084   __ret; \
9085 })
9086 #else
9087 #define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
9088   uint64x2_t __ret; \
9089   uint64x2_t __s1 = __p1; \
9090   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9091   __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
9092   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9093   __ret; \
9094 })
9095 #endif
9096 
9097 #ifdef __LITTLE_ENDIAN__
9098 #define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9099   uint16x8_t __ret; \
9100   uint16x8_t __s1 = __p1; \
9101   __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
9102   __ret; \
9103 })
9104 #else
9105 #define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9106   uint16x8_t __ret; \
9107   uint16x8_t __s1 = __p1; \
9108   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9109   __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
9110   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9111   __ret; \
9112 })
9113 #endif
9114 
9115 #ifdef __LITTLE_ENDIAN__
9116 #define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9117   int8x16_t __ret; \
9118   int8x16_t __s1 = __p1; \
9119   __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
9120   __ret; \
9121 })
9122 #else
9123 #define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9124   int8x16_t __ret; \
9125   int8x16_t __s1 = __p1; \
9126   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9127   __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
9128   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9129   __ret; \
9130 })
9131 #endif
9132 
9133 #ifdef __LITTLE_ENDIAN__
9134 #define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9135   float32x4_t __ret; \
9136   float32x4_t __s1 = __p1; \
9137   __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
9138   __ret; \
9139 })
9140 #else
9141 #define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9142   float32x4_t __ret; \
9143   float32x4_t __s1 = __p1; \
9144   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9145   __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
9146   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9147   __ret; \
9148 })
9149 #endif
9150 
9151 #ifdef __LITTLE_ENDIAN__
9152 #define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9153   int32x4_t __ret; \
9154   int32x4_t __s1 = __p1; \
9155   __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
9156   __ret; \
9157 })
9158 #else
9159 #define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9160   int32x4_t __ret; \
9161   int32x4_t __s1 = __p1; \
9162   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9163   __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
9164   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9165   __ret; \
9166 })
9167 #endif
9168 
9169 #ifdef __LITTLE_ENDIAN__
9170 #define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
9171   int64x2_t __ret; \
9172   int64x2_t __s1 = __p1; \
9173   __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
9174   __ret; \
9175 })
9176 #else
9177 #define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
9178   int64x2_t __ret; \
9179   int64x2_t __s1 = __p1; \
9180   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9181   __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
9182   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9183   __ret; \
9184 })
9185 #endif
9186 
9187 #ifdef __LITTLE_ENDIAN__
9188 #define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9189   int16x8_t __ret; \
9190   int16x8_t __s1 = __p1; \
9191   __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
9192   __ret; \
9193 })
9194 #else
9195 #define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9196   int16x8_t __ret; \
9197   int16x8_t __s1 = __p1; \
9198   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9199   __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
9200   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9201   __ret; \
9202 })
9203 #endif
9204 
9205 #ifdef __LITTLE_ENDIAN__
9206 #define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9207   uint8x8_t __ret; \
9208   uint8x8_t __s1 = __p1; \
9209   __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
9210   __ret; \
9211 })
9212 #else
9213 #define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9214   uint8x8_t __ret; \
9215   uint8x8_t __s1 = __p1; \
9216   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9217   __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
9218   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9219   __ret; \
9220 })
9221 #endif
9222 
9223 #ifdef __LITTLE_ENDIAN__
9224 #define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9225   uint32x2_t __ret; \
9226   uint32x2_t __s1 = __p1; \
9227   __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
9228   __ret; \
9229 })
9230 #else
9231 #define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9232   uint32x2_t __ret; \
9233   uint32x2_t __s1 = __p1; \
9234   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9235   __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
9236   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9237   __ret; \
9238 })
9239 #endif
9240 
9241 #define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
9242   uint64x1_t __ret; \
9243   uint64x1_t __s1 = __p1; \
9244   __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
9245   __ret; \
9246 })
9247 #ifdef __LITTLE_ENDIAN__
9248 #define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9249   uint16x4_t __ret; \
9250   uint16x4_t __s1 = __p1; \
9251   __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
9252   __ret; \
9253 })
9254 #else
9255 #define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9256   uint16x4_t __ret; \
9257   uint16x4_t __s1 = __p1; \
9258   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9259   __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
9260   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9261   __ret; \
9262 })
9263 #endif
9264 
9265 #ifdef __LITTLE_ENDIAN__
9266 #define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9267   int8x8_t __ret; \
9268   int8x8_t __s1 = __p1; \
9269   __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
9270   __ret; \
9271 })
9272 #else
9273 #define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9274   int8x8_t __ret; \
9275   int8x8_t __s1 = __p1; \
9276   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9277   __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
9278   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9279   __ret; \
9280 })
9281 #endif
9282 
9283 #ifdef __LITTLE_ENDIAN__
9284 #define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9285   float32x2_t __ret; \
9286   float32x2_t __s1 = __p1; \
9287   __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
9288   __ret; \
9289 })
9290 #else
9291 #define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9292   float32x2_t __ret; \
9293   float32x2_t __s1 = __p1; \
9294   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9295   __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
9296   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9297   __ret; \
9298 })
9299 #endif
9300 
9301 #ifdef __LITTLE_ENDIAN__
9302 #define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9303   int32x2_t __ret; \
9304   int32x2_t __s1 = __p1; \
9305   __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
9306   __ret; \
9307 })
9308 #else
9309 #define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9310   int32x2_t __ret; \
9311   int32x2_t __s1 = __p1; \
9312   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9313   __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
9314   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9315   __ret; \
9316 })
9317 #endif
9318 
9319 #define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
9320   int64x1_t __ret; \
9321   int64x1_t __s1 = __p1; \
9322   __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
9323   __ret; \
9324 })
9325 #ifdef __LITTLE_ENDIAN__
9326 #define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9327   int16x4_t __ret; \
9328   int16x4_t __s1 = __p1; \
9329   __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
9330   __ret; \
9331 })
9332 #else
9333 #define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9334   int16x4_t __ret; \
9335   int16x4_t __s1 = __p1; \
9336   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9337   __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
9338   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9339   __ret; \
9340 })
9341 #endif
9342 
9343 #ifdef __LITTLE_ENDIAN__
9344 #define vld1_p8_x2(__p0) __extension__ ({ \
9345   poly8x8x2_t __ret; \
9346   __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
9347   __ret; \
9348 })
9349 #else
9350 #define vld1_p8_x2(__p0) __extension__ ({ \
9351   poly8x8x2_t __ret; \
9352   __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
9353  \
9354   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9355   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9356   __ret; \
9357 })
9358 #endif
9359 
9360 #ifdef __LITTLE_ENDIAN__
9361 #define vld1_p16_x2(__p0) __extension__ ({ \
9362   poly16x4x2_t __ret; \
9363   __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
9364   __ret; \
9365 })
9366 #else
9367 #define vld1_p16_x2(__p0) __extension__ ({ \
9368   poly16x4x2_t __ret; \
9369   __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
9370  \
9371   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9372   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9373   __ret; \
9374 })
9375 #endif
9376 
9377 #ifdef __LITTLE_ENDIAN__
9378 #define vld1q_p8_x2(__p0) __extension__ ({ \
9379   poly8x16x2_t __ret; \
9380   __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
9381   __ret; \
9382 })
9383 #else
9384 #define vld1q_p8_x2(__p0) __extension__ ({ \
9385   poly8x16x2_t __ret; \
9386   __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
9387  \
9388   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9389   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9390   __ret; \
9391 })
9392 #endif
9393 
9394 #ifdef __LITTLE_ENDIAN__
9395 #define vld1q_p16_x2(__p0) __extension__ ({ \
9396   poly16x8x2_t __ret; \
9397   __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
9398   __ret; \
9399 })
9400 #else
9401 #define vld1q_p16_x2(__p0) __extension__ ({ \
9402   poly16x8x2_t __ret; \
9403   __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
9404  \
9405   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9406   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9407   __ret; \
9408 })
9409 #endif
9410 
9411 #ifdef __LITTLE_ENDIAN__
9412 #define vld1q_u8_x2(__p0) __extension__ ({ \
9413   uint8x16x2_t __ret; \
9414   __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
9415   __ret; \
9416 })
9417 #else
9418 #define vld1q_u8_x2(__p0) __extension__ ({ \
9419   uint8x16x2_t __ret; \
9420   __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
9421  \
9422   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9423   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9424   __ret; \
9425 })
9426 #endif
9427 
9428 #ifdef __LITTLE_ENDIAN__
9429 #define vld1q_u32_x2(__p0) __extension__ ({ \
9430   uint32x4x2_t __ret; \
9431   __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
9432   __ret; \
9433 })
9434 #else
9435 #define vld1q_u32_x2(__p0) __extension__ ({ \
9436   uint32x4x2_t __ret; \
9437   __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
9438  \
9439   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9440   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9441   __ret; \
9442 })
9443 #endif
9444 
9445 #ifdef __LITTLE_ENDIAN__
9446 #define vld1q_u64_x2(__p0) __extension__ ({ \
9447   uint64x2x2_t __ret; \
9448   __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
9449   __ret; \
9450 })
9451 #else
9452 #define vld1q_u64_x2(__p0) __extension__ ({ \
9453   uint64x2x2_t __ret; \
9454   __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
9455  \
9456   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9457   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9458   __ret; \
9459 })
9460 #endif
9461 
9462 #ifdef __LITTLE_ENDIAN__
9463 #define vld1q_u16_x2(__p0) __extension__ ({ \
9464   uint16x8x2_t __ret; \
9465   __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
9466   __ret; \
9467 })
9468 #else
9469 #define vld1q_u16_x2(__p0) __extension__ ({ \
9470   uint16x8x2_t __ret; \
9471   __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
9472  \
9473   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9474   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9475   __ret; \
9476 })
9477 #endif
9478 
9479 #ifdef __LITTLE_ENDIAN__
9480 #define vld1q_s8_x2(__p0) __extension__ ({ \
9481   int8x16x2_t __ret; \
9482   __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
9483   __ret; \
9484 })
9485 #else
9486 #define vld1q_s8_x2(__p0) __extension__ ({ \
9487   int8x16x2_t __ret; \
9488   __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
9489  \
9490   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9491   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9492   __ret; \
9493 })
9494 #endif
9495 
9496 #ifdef __LITTLE_ENDIAN__
9497 #define vld1q_f32_x2(__p0) __extension__ ({ \
9498   float32x4x2_t __ret; \
9499   __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
9500   __ret; \
9501 })
9502 #else
9503 #define vld1q_f32_x2(__p0) __extension__ ({ \
9504   float32x4x2_t __ret; \
9505   __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
9506  \
9507   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9508   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9509   __ret; \
9510 })
9511 #endif
9512 
9513 #ifdef __LITTLE_ENDIAN__
9514 #define vld1q_s32_x2(__p0) __extension__ ({ \
9515   int32x4x2_t __ret; \
9516   __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
9517   __ret; \
9518 })
9519 #else
9520 #define vld1q_s32_x2(__p0) __extension__ ({ \
9521   int32x4x2_t __ret; \
9522   __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
9523  \
9524   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9525   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9526   __ret; \
9527 })
9528 #endif
9529 
9530 #ifdef __LITTLE_ENDIAN__
9531 #define vld1q_s64_x2(__p0) __extension__ ({ \
9532   int64x2x2_t __ret; \
9533   __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
9534   __ret; \
9535 })
9536 #else
9537 #define vld1q_s64_x2(__p0) __extension__ ({ \
9538   int64x2x2_t __ret; \
9539   __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
9540  \
9541   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9542   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9543   __ret; \
9544 })
9545 #endif
9546 
9547 #ifdef __LITTLE_ENDIAN__
9548 #define vld1q_s16_x2(__p0) __extension__ ({ \
9549   int16x8x2_t __ret; \
9550   __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
9551   __ret; \
9552 })
9553 #else
9554 #define vld1q_s16_x2(__p0) __extension__ ({ \
9555   int16x8x2_t __ret; \
9556   __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
9557  \
9558   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9559   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9560   __ret; \
9561 })
9562 #endif
9563 
9564 #ifdef __LITTLE_ENDIAN__
9565 #define vld1_u8_x2(__p0) __extension__ ({ \
9566   uint8x8x2_t __ret; \
9567   __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
9568   __ret; \
9569 })
9570 #else
9571 #define vld1_u8_x2(__p0) __extension__ ({ \
9572   uint8x8x2_t __ret; \
9573   __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
9574  \
9575   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9576   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9577   __ret; \
9578 })
9579 #endif
9580 
9581 #ifdef __LITTLE_ENDIAN__
9582 #define vld1_u32_x2(__p0) __extension__ ({ \
9583   uint32x2x2_t __ret; \
9584   __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
9585   __ret; \
9586 })
9587 #else
9588 #define vld1_u32_x2(__p0) __extension__ ({ \
9589   uint32x2x2_t __ret; \
9590   __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
9591  \
9592   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9593   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9594   __ret; \
9595 })
9596 #endif
9597 
9598 #define vld1_u64_x2(__p0) __extension__ ({ \
9599   uint64x1x2_t __ret; \
9600   __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
9601   __ret; \
9602 })
9603 #ifdef __LITTLE_ENDIAN__
9604 #define vld1_u16_x2(__p0) __extension__ ({ \
9605   uint16x4x2_t __ret; \
9606   __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
9607   __ret; \
9608 })
9609 #else
9610 #define vld1_u16_x2(__p0) __extension__ ({ \
9611   uint16x4x2_t __ret; \
9612   __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
9613  \
9614   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9615   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9616   __ret; \
9617 })
9618 #endif
9619 
9620 #ifdef __LITTLE_ENDIAN__
9621 #define vld1_s8_x2(__p0) __extension__ ({ \
9622   int8x8x2_t __ret; \
9623   __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
9624   __ret; \
9625 })
9626 #else
9627 #define vld1_s8_x2(__p0) __extension__ ({ \
9628   int8x8x2_t __ret; \
9629   __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
9630  \
9631   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9632   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9633   __ret; \
9634 })
9635 #endif
9636 
9637 #ifdef __LITTLE_ENDIAN__
9638 #define vld1_f32_x2(__p0) __extension__ ({ \
9639   float32x2x2_t __ret; \
9640   __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
9641   __ret; \
9642 })
9643 #else
9644 #define vld1_f32_x2(__p0) __extension__ ({ \
9645   float32x2x2_t __ret; \
9646   __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
9647  \
9648   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9649   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9650   __ret; \
9651 })
9652 #endif
9653 
9654 #ifdef __LITTLE_ENDIAN__
9655 #define vld1_s32_x2(__p0) __extension__ ({ \
9656   int32x2x2_t __ret; \
9657   __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
9658   __ret; \
9659 })
9660 #else
9661 #define vld1_s32_x2(__p0) __extension__ ({ \
9662   int32x2x2_t __ret; \
9663   __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
9664  \
9665   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9666   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9667   __ret; \
9668 })
9669 #endif
9670 
9671 #define vld1_s64_x2(__p0) __extension__ ({ \
9672   int64x1x2_t __ret; \
9673   __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
9674   __ret; \
9675 })
9676 #ifdef __LITTLE_ENDIAN__
9677 #define vld1_s16_x2(__p0) __extension__ ({ \
9678   int16x4x2_t __ret; \
9679   __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
9680   __ret; \
9681 })
9682 #else
9683 #define vld1_s16_x2(__p0) __extension__ ({ \
9684   int16x4x2_t __ret; \
9685   __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
9686  \
9687   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9688   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9689   __ret; \
9690 })
9691 #endif
9692 
9693 #ifdef __LITTLE_ENDIAN__
9694 #define vld1_p8_x3(__p0) __extension__ ({ \
9695   poly8x8x3_t __ret; \
9696   __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
9697   __ret; \
9698 })
9699 #else
9700 #define vld1_p8_x3(__p0) __extension__ ({ \
9701   poly8x8x3_t __ret; \
9702   __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
9703  \
9704   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9705   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9706   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9707   __ret; \
9708 })
9709 #endif
9710 
9711 #ifdef __LITTLE_ENDIAN__
9712 #define vld1_p16_x3(__p0) __extension__ ({ \
9713   poly16x4x3_t __ret; \
9714   __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
9715   __ret; \
9716 })
9717 #else
9718 #define vld1_p16_x3(__p0) __extension__ ({ \
9719   poly16x4x3_t __ret; \
9720   __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
9721  \
9722   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9723   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9724   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9725   __ret; \
9726 })
9727 #endif
9728 
9729 #ifdef __LITTLE_ENDIAN__
9730 #define vld1q_p8_x3(__p0) __extension__ ({ \
9731   poly8x16x3_t __ret; \
9732   __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
9733   __ret; \
9734 })
9735 #else
9736 #define vld1q_p8_x3(__p0) __extension__ ({ \
9737   poly8x16x3_t __ret; \
9738   __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
9739  \
9740   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9741   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9742   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9743   __ret; \
9744 })
9745 #endif
9746 
9747 #ifdef __LITTLE_ENDIAN__
9748 #define vld1q_p16_x3(__p0) __extension__ ({ \
9749   poly16x8x3_t __ret; \
9750   __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
9751   __ret; \
9752 })
9753 #else
9754 #define vld1q_p16_x3(__p0) __extension__ ({ \
9755   poly16x8x3_t __ret; \
9756   __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
9757  \
9758   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9759   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9760   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9761   __ret; \
9762 })
9763 #endif
9764 
9765 #ifdef __LITTLE_ENDIAN__
9766 #define vld1q_u8_x3(__p0) __extension__ ({ \
9767   uint8x16x3_t __ret; \
9768   __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
9769   __ret; \
9770 })
9771 #else
9772 #define vld1q_u8_x3(__p0) __extension__ ({ \
9773   uint8x16x3_t __ret; \
9774   __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
9775  \
9776   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9777   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9778   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9779   __ret; \
9780 })
9781 #endif
9782 
9783 #ifdef __LITTLE_ENDIAN__
9784 #define vld1q_u32_x3(__p0) __extension__ ({ \
9785   uint32x4x3_t __ret; \
9786   __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
9787   __ret; \
9788 })
9789 #else
9790 #define vld1q_u32_x3(__p0) __extension__ ({ \
9791   uint32x4x3_t __ret; \
9792   __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
9793  \
9794   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9795   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9796   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9797   __ret; \
9798 })
9799 #endif
9800 
9801 #ifdef __LITTLE_ENDIAN__
9802 #define vld1q_u64_x3(__p0) __extension__ ({ \
9803   uint64x2x3_t __ret; \
9804   __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
9805   __ret; \
9806 })
9807 #else
9808 #define vld1q_u64_x3(__p0) __extension__ ({ \
9809   uint64x2x3_t __ret; \
9810   __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
9811  \
9812   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9813   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9814   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9815   __ret; \
9816 })
9817 #endif
9818 
9819 #ifdef __LITTLE_ENDIAN__
9820 #define vld1q_u16_x3(__p0) __extension__ ({ \
9821   uint16x8x3_t __ret; \
9822   __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
9823   __ret; \
9824 })
9825 #else
9826 #define vld1q_u16_x3(__p0) __extension__ ({ \
9827   uint16x8x3_t __ret; \
9828   __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
9829  \
9830   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9831   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9832   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9833   __ret; \
9834 })
9835 #endif
9836 
9837 #ifdef __LITTLE_ENDIAN__
9838 #define vld1q_s8_x3(__p0) __extension__ ({ \
9839   int8x16x3_t __ret; \
9840   __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
9841   __ret; \
9842 })
9843 #else
9844 #define vld1q_s8_x3(__p0) __extension__ ({ \
9845   int8x16x3_t __ret; \
9846   __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
9847  \
9848   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9849   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9850   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9851   __ret; \
9852 })
9853 #endif
9854 
9855 #ifdef __LITTLE_ENDIAN__
9856 #define vld1q_f32_x3(__p0) __extension__ ({ \
9857   float32x4x3_t __ret; \
9858   __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
9859   __ret; \
9860 })
9861 #else
9862 #define vld1q_f32_x3(__p0) __extension__ ({ \
9863   float32x4x3_t __ret; \
9864   __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
9865  \
9866   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9867   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9868   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9869   __ret; \
9870 })
9871 #endif
9872 
9873 #ifdef __LITTLE_ENDIAN__
9874 #define vld1q_s32_x3(__p0) __extension__ ({ \
9875   int32x4x3_t __ret; \
9876   __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
9877   __ret; \
9878 })
9879 #else
9880 #define vld1q_s32_x3(__p0) __extension__ ({ \
9881   int32x4x3_t __ret; \
9882   __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
9883  \
9884   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9885   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9886   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9887   __ret; \
9888 })
9889 #endif
9890 
9891 #ifdef __LITTLE_ENDIAN__
9892 #define vld1q_s64_x3(__p0) __extension__ ({ \
9893   int64x2x3_t __ret; \
9894   __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
9895   __ret; \
9896 })
9897 #else
9898 #define vld1q_s64_x3(__p0) __extension__ ({ \
9899   int64x2x3_t __ret; \
9900   __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
9901  \
9902   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9903   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9904   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9905   __ret; \
9906 })
9907 #endif
9908 
9909 #ifdef __LITTLE_ENDIAN__
9910 #define vld1q_s16_x3(__p0) __extension__ ({ \
9911   int16x8x3_t __ret; \
9912   __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
9913   __ret; \
9914 })
9915 #else
9916 #define vld1q_s16_x3(__p0) __extension__ ({ \
9917   int16x8x3_t __ret; \
9918   __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
9919  \
9920   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9921   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9922   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9923   __ret; \
9924 })
9925 #endif
9926 
9927 #ifdef __LITTLE_ENDIAN__
9928 #define vld1_u8_x3(__p0) __extension__ ({ \
9929   uint8x8x3_t __ret; \
9930   __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
9931   __ret; \
9932 })
9933 #else
9934 #define vld1_u8_x3(__p0) __extension__ ({ \
9935   uint8x8x3_t __ret; \
9936   __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
9937  \
9938   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9939   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9940   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9941   __ret; \
9942 })
9943 #endif
9944 
9945 #ifdef __LITTLE_ENDIAN__
9946 #define vld1_u32_x3(__p0) __extension__ ({ \
9947   uint32x2x3_t __ret; \
9948   __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
9949   __ret; \
9950 })
9951 #else
9952 #define vld1_u32_x3(__p0) __extension__ ({ \
9953   uint32x2x3_t __ret; \
9954   __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
9955  \
9956   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9957   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9958   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9959   __ret; \
9960 })
9961 #endif
9962 
9963 #define vld1_u64_x3(__p0) __extension__ ({ \
9964   uint64x1x3_t __ret; \
9965   __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
9966   __ret; \
9967 })
9968 #ifdef __LITTLE_ENDIAN__
9969 #define vld1_u16_x3(__p0) __extension__ ({ \
9970   uint16x4x3_t __ret; \
9971   __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
9972   __ret; \
9973 })
9974 #else
9975 #define vld1_u16_x3(__p0) __extension__ ({ \
9976   uint16x4x3_t __ret; \
9977   __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
9978  \
9979   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9980   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9981   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9982   __ret; \
9983 })
9984 #endif
9985 
9986 #ifdef __LITTLE_ENDIAN__
9987 #define vld1_s8_x3(__p0) __extension__ ({ \
9988   int8x8x3_t __ret; \
9989   __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
9990   __ret; \
9991 })
9992 #else
9993 #define vld1_s8_x3(__p0) __extension__ ({ \
9994   int8x8x3_t __ret; \
9995   __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
9996  \
9997   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9998   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9999   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10000   __ret; \
10001 })
10002 #endif
10003 
10004 #ifdef __LITTLE_ENDIAN__
10005 #define vld1_f32_x3(__p0) __extension__ ({ \
10006   float32x2x3_t __ret; \
10007   __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
10008   __ret; \
10009 })
10010 #else
10011 #define vld1_f32_x3(__p0) __extension__ ({ \
10012   float32x2x3_t __ret; \
10013   __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
10014  \
10015   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10016   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10017   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10018   __ret; \
10019 })
10020 #endif
10021 
10022 #ifdef __LITTLE_ENDIAN__
10023 #define vld1_s32_x3(__p0) __extension__ ({ \
10024   int32x2x3_t __ret; \
10025   __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
10026   __ret; \
10027 })
10028 #else
10029 #define vld1_s32_x3(__p0) __extension__ ({ \
10030   int32x2x3_t __ret; \
10031   __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
10032  \
10033   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10034   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10035   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10036   __ret; \
10037 })
10038 #endif
10039 
10040 #define vld1_s64_x3(__p0) __extension__ ({ \
10041   int64x1x3_t __ret; \
10042   __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
10043   __ret; \
10044 })
10045 #ifdef __LITTLE_ENDIAN__
10046 #define vld1_s16_x3(__p0) __extension__ ({ \
10047   int16x4x3_t __ret; \
10048   __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
10049   __ret; \
10050 })
10051 #else
10052 #define vld1_s16_x3(__p0) __extension__ ({ \
10053   int16x4x3_t __ret; \
10054   __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
10055  \
10056   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10057   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10058   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10059   __ret; \
10060 })
10061 #endif
10062 
10063 #ifdef __LITTLE_ENDIAN__
10064 #define vld1_p8_x4(__p0) __extension__ ({ \
10065   poly8x8x4_t __ret; \
10066   __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
10067   __ret; \
10068 })
10069 #else
10070 #define vld1_p8_x4(__p0) __extension__ ({ \
10071   poly8x8x4_t __ret; \
10072   __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
10073  \
10074   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10075   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10076   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10077   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10078   __ret; \
10079 })
10080 #endif
10081 
10082 #ifdef __LITTLE_ENDIAN__
10083 #define vld1_p16_x4(__p0) __extension__ ({ \
10084   poly16x4x4_t __ret; \
10085   __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
10086   __ret; \
10087 })
10088 #else
10089 #define vld1_p16_x4(__p0) __extension__ ({ \
10090   poly16x4x4_t __ret; \
10091   __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
10092  \
10093   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10094   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10095   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10096   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10097   __ret; \
10098 })
10099 #endif
10100 
10101 #ifdef __LITTLE_ENDIAN__
10102 #define vld1q_p8_x4(__p0) __extension__ ({ \
10103   poly8x16x4_t __ret; \
10104   __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
10105   __ret; \
10106 })
10107 #else
10108 #define vld1q_p8_x4(__p0) __extension__ ({ \
10109   poly8x16x4_t __ret; \
10110   __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
10111  \
10112   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10113   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10114   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10115   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10116   __ret; \
10117 })
10118 #endif
10119 
10120 #ifdef __LITTLE_ENDIAN__
10121 #define vld1q_p16_x4(__p0) __extension__ ({ \
10122   poly16x8x4_t __ret; \
10123   __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
10124   __ret; \
10125 })
10126 #else
10127 #define vld1q_p16_x4(__p0) __extension__ ({ \
10128   poly16x8x4_t __ret; \
10129   __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
10130  \
10131   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10132   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10133   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10134   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10135   __ret; \
10136 })
10137 #endif
10138 
10139 #ifdef __LITTLE_ENDIAN__
10140 #define vld1q_u8_x4(__p0) __extension__ ({ \
10141   uint8x16x4_t __ret; \
10142   __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
10143   __ret; \
10144 })
10145 #else
10146 #define vld1q_u8_x4(__p0) __extension__ ({ \
10147   uint8x16x4_t __ret; \
10148   __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
10149  \
10150   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10151   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10152   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10153   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10154   __ret; \
10155 })
10156 #endif
10157 
10158 #ifdef __LITTLE_ENDIAN__
10159 #define vld1q_u32_x4(__p0) __extension__ ({ \
10160   uint32x4x4_t __ret; \
10161   __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
10162   __ret; \
10163 })
10164 #else
10165 #define vld1q_u32_x4(__p0) __extension__ ({ \
10166   uint32x4x4_t __ret; \
10167   __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
10168  \
10169   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10170   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10171   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10172   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10173   __ret; \
10174 })
10175 #endif
10176 
10177 #ifdef __LITTLE_ENDIAN__
10178 #define vld1q_u64_x4(__p0) __extension__ ({ \
10179   uint64x2x4_t __ret; \
10180   __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
10181   __ret; \
10182 })
10183 #else
10184 #define vld1q_u64_x4(__p0) __extension__ ({ \
10185   uint64x2x4_t __ret; \
10186   __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
10187  \
10188   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10189   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10190   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10191   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10192   __ret; \
10193 })
10194 #endif
10195 
10196 #ifdef __LITTLE_ENDIAN__
10197 #define vld1q_u16_x4(__p0) __extension__ ({ \
10198   uint16x8x4_t __ret; \
10199   __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
10200   __ret; \
10201 })
10202 #else
10203 #define vld1q_u16_x4(__p0) __extension__ ({ \
10204   uint16x8x4_t __ret; \
10205   __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
10206  \
10207   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10208   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10209   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10210   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10211   __ret; \
10212 })
10213 #endif
10214 
10215 #ifdef __LITTLE_ENDIAN__
10216 #define vld1q_s8_x4(__p0) __extension__ ({ \
10217   int8x16x4_t __ret; \
10218   __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
10219   __ret; \
10220 })
10221 #else
10222 #define vld1q_s8_x4(__p0) __extension__ ({ \
10223   int8x16x4_t __ret; \
10224   __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
10225  \
10226   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10227   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10228   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10229   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10230   __ret; \
10231 })
10232 #endif
10233 
10234 #ifdef __LITTLE_ENDIAN__
10235 #define vld1q_f32_x4(__p0) __extension__ ({ \
10236   float32x4x4_t __ret; \
10237   __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
10238   __ret; \
10239 })
10240 #else
10241 #define vld1q_f32_x4(__p0) __extension__ ({ \
10242   float32x4x4_t __ret; \
10243   __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
10244  \
10245   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10246   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10247   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10248   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10249   __ret; \
10250 })
10251 #endif
10252 
10253 #ifdef __LITTLE_ENDIAN__
10254 #define vld1q_s32_x4(__p0) __extension__ ({ \
10255   int32x4x4_t __ret; \
10256   __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
10257   __ret; \
10258 })
10259 #else
10260 #define vld1q_s32_x4(__p0) __extension__ ({ \
10261   int32x4x4_t __ret; \
10262   __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
10263  \
10264   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10265   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10266   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10267   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10268   __ret; \
10269 })
10270 #endif
10271 
10272 #ifdef __LITTLE_ENDIAN__
10273 #define vld1q_s64_x4(__p0) __extension__ ({ \
10274   int64x2x4_t __ret; \
10275   __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
10276   __ret; \
10277 })
10278 #else
10279 #define vld1q_s64_x4(__p0) __extension__ ({ \
10280   int64x2x4_t __ret; \
10281   __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
10282  \
10283   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10284   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10285   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10286   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10287   __ret; \
10288 })
10289 #endif
10290 
10291 #ifdef __LITTLE_ENDIAN__
10292 #define vld1q_s16_x4(__p0) __extension__ ({ \
10293   int16x8x4_t __ret; \
10294   __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
10295   __ret; \
10296 })
10297 #else
10298 #define vld1q_s16_x4(__p0) __extension__ ({ \
10299   int16x8x4_t __ret; \
10300   __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
10301  \
10302   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10303   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10304   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10305   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10306   __ret; \
10307 })
10308 #endif
10309 
10310 #ifdef __LITTLE_ENDIAN__
10311 #define vld1_u8_x4(__p0) __extension__ ({ \
10312   uint8x8x4_t __ret; \
10313   __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
10314   __ret; \
10315 })
10316 #else
10317 #define vld1_u8_x4(__p0) __extension__ ({ \
10318   uint8x8x4_t __ret; \
10319   __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
10320  \
10321   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10322   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10323   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10324   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10325   __ret; \
10326 })
10327 #endif
10328 
10329 #ifdef __LITTLE_ENDIAN__
10330 #define vld1_u32_x4(__p0) __extension__ ({ \
10331   uint32x2x4_t __ret; \
10332   __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
10333   __ret; \
10334 })
10335 #else
10336 #define vld1_u32_x4(__p0) __extension__ ({ \
10337   uint32x2x4_t __ret; \
10338   __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
10339  \
10340   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10341   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10342   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10343   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10344   __ret; \
10345 })
10346 #endif
10347 
10348 #define vld1_u64_x4(__p0) __extension__ ({ \
10349   uint64x1x4_t __ret; \
10350   __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
10351   __ret; \
10352 })
10353 #ifdef __LITTLE_ENDIAN__
10354 #define vld1_u16_x4(__p0) __extension__ ({ \
10355   uint16x4x4_t __ret; \
10356   __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
10357   __ret; \
10358 })
10359 #else
10360 #define vld1_u16_x4(__p0) __extension__ ({ \
10361   uint16x4x4_t __ret; \
10362   __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
10363  \
10364   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10365   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10366   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10367   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10368   __ret; \
10369 })
10370 #endif
10371 
10372 #ifdef __LITTLE_ENDIAN__
10373 #define vld1_s8_x4(__p0) __extension__ ({ \
10374   int8x8x4_t __ret; \
10375   __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
10376   __ret; \
10377 })
10378 #else
10379 #define vld1_s8_x4(__p0) __extension__ ({ \
10380   int8x8x4_t __ret; \
10381   __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
10382  \
10383   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10384   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10385   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10386   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10387   __ret; \
10388 })
10389 #endif
10390 
10391 #ifdef __LITTLE_ENDIAN__
10392 #define vld1_f32_x4(__p0) __extension__ ({ \
10393   float32x2x4_t __ret; \
10394   __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
10395   __ret; \
10396 })
10397 #else
10398 #define vld1_f32_x4(__p0) __extension__ ({ \
10399   float32x2x4_t __ret; \
10400   __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
10401  \
10402   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10403   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10404   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10405   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10406   __ret; \
10407 })
10408 #endif
10409 
10410 #ifdef __LITTLE_ENDIAN__
10411 #define vld1_s32_x4(__p0) __extension__ ({ \
10412   int32x2x4_t __ret; \
10413   __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
10414   __ret; \
10415 })
10416 #else
10417 #define vld1_s32_x4(__p0) __extension__ ({ \
10418   int32x2x4_t __ret; \
10419   __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
10420  \
10421   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10422   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10423   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10424   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10425   __ret; \
10426 })
10427 #endif
10428 
10429 #define vld1_s64_x4(__p0) __extension__ ({ \
10430   int64x1x4_t __ret; \
10431   __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
10432   __ret; \
10433 })
10434 #ifdef __LITTLE_ENDIAN__
10435 #define vld1_s16_x4(__p0) __extension__ ({ \
10436   int16x4x4_t __ret; \
10437   __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
10438   __ret; \
10439 })
10440 #else
10441 #define vld1_s16_x4(__p0) __extension__ ({ \
10442   int16x4x4_t __ret; \
10443   __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
10444  \
10445   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10446   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10447   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10448   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10449   __ret; \
10450 })
10451 #endif
10452 
10453 #ifdef __LITTLE_ENDIAN__
10454 #define vld2_p8(__p0) __extension__ ({ \
10455   poly8x8x2_t __ret; \
10456   __builtin_neon_vld2_v(&__ret, __p0, 4); \
10457   __ret; \
10458 })
10459 #else
10460 #define vld2_p8(__p0) __extension__ ({ \
10461   poly8x8x2_t __ret; \
10462   __builtin_neon_vld2_v(&__ret, __p0, 4); \
10463  \
10464   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10465   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10466   __ret; \
10467 })
10468 #endif
10469 
10470 #ifdef __LITTLE_ENDIAN__
10471 #define vld2_p16(__p0) __extension__ ({ \
10472   poly16x4x2_t __ret; \
10473   __builtin_neon_vld2_v(&__ret, __p0, 5); \
10474   __ret; \
10475 })
10476 #else
10477 #define vld2_p16(__p0) __extension__ ({ \
10478   poly16x4x2_t __ret; \
10479   __builtin_neon_vld2_v(&__ret, __p0, 5); \
10480  \
10481   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10482   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10483   __ret; \
10484 })
10485 #endif
10486 
10487 #ifdef __LITTLE_ENDIAN__
10488 #define vld2q_p8(__p0) __extension__ ({ \
10489   poly8x16x2_t __ret; \
10490   __builtin_neon_vld2q_v(&__ret, __p0, 36); \
10491   __ret; \
10492 })
10493 #else
10494 #define vld2q_p8(__p0) __extension__ ({ \
10495   poly8x16x2_t __ret; \
10496   __builtin_neon_vld2q_v(&__ret, __p0, 36); \
10497  \
10498   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10499   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10500   __ret; \
10501 })
10502 #endif
10503 
10504 #ifdef __LITTLE_ENDIAN__
10505 #define vld2q_p16(__p0) __extension__ ({ \
10506   poly16x8x2_t __ret; \
10507   __builtin_neon_vld2q_v(&__ret, __p0, 37); \
10508   __ret; \
10509 })
10510 #else
10511 #define vld2q_p16(__p0) __extension__ ({ \
10512   poly16x8x2_t __ret; \
10513   __builtin_neon_vld2q_v(&__ret, __p0, 37); \
10514  \
10515   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10516   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10517   __ret; \
10518 })
10519 #endif
10520 
10521 #ifdef __LITTLE_ENDIAN__
10522 #define vld2q_u8(__p0) __extension__ ({ \
10523   uint8x16x2_t __ret; \
10524   __builtin_neon_vld2q_v(&__ret, __p0, 48); \
10525   __ret; \
10526 })
10527 #else
10528 #define vld2q_u8(__p0) __extension__ ({ \
10529   uint8x16x2_t __ret; \
10530   __builtin_neon_vld2q_v(&__ret, __p0, 48); \
10531  \
10532   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10533   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10534   __ret; \
10535 })
10536 #endif
10537 
10538 #ifdef __LITTLE_ENDIAN__
10539 #define vld2q_u32(__p0) __extension__ ({ \
10540   uint32x4x2_t __ret; \
10541   __builtin_neon_vld2q_v(&__ret, __p0, 50); \
10542   __ret; \
10543 })
10544 #else
10545 #define vld2q_u32(__p0) __extension__ ({ \
10546   uint32x4x2_t __ret; \
10547   __builtin_neon_vld2q_v(&__ret, __p0, 50); \
10548  \
10549   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10550   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10551   __ret; \
10552 })
10553 #endif
10554 
10555 #ifdef __LITTLE_ENDIAN__
10556 #define vld2q_u16(__p0) __extension__ ({ \
10557   uint16x8x2_t __ret; \
10558   __builtin_neon_vld2q_v(&__ret, __p0, 49); \
10559   __ret; \
10560 })
10561 #else
10562 #define vld2q_u16(__p0) __extension__ ({ \
10563   uint16x8x2_t __ret; \
10564   __builtin_neon_vld2q_v(&__ret, __p0, 49); \
10565  \
10566   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10567   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10568   __ret; \
10569 })
10570 #endif
10571 
10572 #ifdef __LITTLE_ENDIAN__
10573 #define vld2q_s8(__p0) __extension__ ({ \
10574   int8x16x2_t __ret; \
10575   __builtin_neon_vld2q_v(&__ret, __p0, 32); \
10576   __ret; \
10577 })
10578 #else
10579 #define vld2q_s8(__p0) __extension__ ({ \
10580   int8x16x2_t __ret; \
10581   __builtin_neon_vld2q_v(&__ret, __p0, 32); \
10582  \
10583   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10584   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10585   __ret; \
10586 })
10587 #endif
10588 
10589 #ifdef __LITTLE_ENDIAN__
10590 #define vld2q_f32(__p0) __extension__ ({ \
10591   float32x4x2_t __ret; \
10592   __builtin_neon_vld2q_v(&__ret, __p0, 41); \
10593   __ret; \
10594 })
10595 #else
10596 #define vld2q_f32(__p0) __extension__ ({ \
10597   float32x4x2_t __ret; \
10598   __builtin_neon_vld2q_v(&__ret, __p0, 41); \
10599  \
10600   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10601   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10602   __ret; \
10603 })
10604 #endif
10605 
10606 #ifdef __LITTLE_ENDIAN__
10607 #define vld2q_s32(__p0) __extension__ ({ \
10608   int32x4x2_t __ret; \
10609   __builtin_neon_vld2q_v(&__ret, __p0, 34); \
10610   __ret; \
10611 })
10612 #else
10613 #define vld2q_s32(__p0) __extension__ ({ \
10614   int32x4x2_t __ret; \
10615   __builtin_neon_vld2q_v(&__ret, __p0, 34); \
10616  \
10617   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10618   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10619   __ret; \
10620 })
10621 #endif
10622 
10623 #ifdef __LITTLE_ENDIAN__
10624 #define vld2q_s16(__p0) __extension__ ({ \
10625   int16x8x2_t __ret; \
10626   __builtin_neon_vld2q_v(&__ret, __p0, 33); \
10627   __ret; \
10628 })
10629 #else
10630 #define vld2q_s16(__p0) __extension__ ({ \
10631   int16x8x2_t __ret; \
10632   __builtin_neon_vld2q_v(&__ret, __p0, 33); \
10633  \
10634   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10635   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10636   __ret; \
10637 })
10638 #endif
10639 
10640 #ifdef __LITTLE_ENDIAN__
10641 #define vld2_u8(__p0) __extension__ ({ \
10642   uint8x8x2_t __ret; \
10643   __builtin_neon_vld2_v(&__ret, __p0, 16); \
10644   __ret; \
10645 })
10646 #else
10647 #define vld2_u8(__p0) __extension__ ({ \
10648   uint8x8x2_t __ret; \
10649   __builtin_neon_vld2_v(&__ret, __p0, 16); \
10650  \
10651   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10652   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10653   __ret; \
10654 })
10655 #endif
10656 
10657 #ifdef __LITTLE_ENDIAN__
10658 #define vld2_u32(__p0) __extension__ ({ \
10659   uint32x2x2_t __ret; \
10660   __builtin_neon_vld2_v(&__ret, __p0, 18); \
10661   __ret; \
10662 })
10663 #else
10664 #define vld2_u32(__p0) __extension__ ({ \
10665   uint32x2x2_t __ret; \
10666   __builtin_neon_vld2_v(&__ret, __p0, 18); \
10667  \
10668   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10669   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10670   __ret; \
10671 })
10672 #endif
10673 
10674 #define vld2_u64(__p0) __extension__ ({ \
10675   uint64x1x2_t __ret; \
10676   __builtin_neon_vld2_v(&__ret, __p0, 19); \
10677   __ret; \
10678 })
10679 #ifdef __LITTLE_ENDIAN__
10680 #define vld2_u16(__p0) __extension__ ({ \
10681   uint16x4x2_t __ret; \
10682   __builtin_neon_vld2_v(&__ret, __p0, 17); \
10683   __ret; \
10684 })
10685 #else
10686 #define vld2_u16(__p0) __extension__ ({ \
10687   uint16x4x2_t __ret; \
10688   __builtin_neon_vld2_v(&__ret, __p0, 17); \
10689  \
10690   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10691   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10692   __ret; \
10693 })
10694 #endif
10695 
10696 #ifdef __LITTLE_ENDIAN__
10697 #define vld2_s8(__p0) __extension__ ({ \
10698   int8x8x2_t __ret; \
10699   __builtin_neon_vld2_v(&__ret, __p0, 0); \
10700   __ret; \
10701 })
10702 #else
10703 #define vld2_s8(__p0) __extension__ ({ \
10704   int8x8x2_t __ret; \
10705   __builtin_neon_vld2_v(&__ret, __p0, 0); \
10706  \
10707   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10708   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10709   __ret; \
10710 })
10711 #endif
10712 
10713 #ifdef __LITTLE_ENDIAN__
10714 #define vld2_f32(__p0) __extension__ ({ \
10715   float32x2x2_t __ret; \
10716   __builtin_neon_vld2_v(&__ret, __p0, 9); \
10717   __ret; \
10718 })
10719 #else
10720 #define vld2_f32(__p0) __extension__ ({ \
10721   float32x2x2_t __ret; \
10722   __builtin_neon_vld2_v(&__ret, __p0, 9); \
10723  \
10724   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10725   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10726   __ret; \
10727 })
10728 #endif
10729 
10730 #ifdef __LITTLE_ENDIAN__
10731 #define vld2_s32(__p0) __extension__ ({ \
10732   int32x2x2_t __ret; \
10733   __builtin_neon_vld2_v(&__ret, __p0, 2); \
10734   __ret; \
10735 })
10736 #else
10737 #define vld2_s32(__p0) __extension__ ({ \
10738   int32x2x2_t __ret; \
10739   __builtin_neon_vld2_v(&__ret, __p0, 2); \
10740  \
10741   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10742   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10743   __ret; \
10744 })
10745 #endif
10746 
10747 #define vld2_s64(__p0) __extension__ ({ \
10748   int64x1x2_t __ret; \
10749   __builtin_neon_vld2_v(&__ret, __p0, 3); \
10750   __ret; \
10751 })
10752 #ifdef __LITTLE_ENDIAN__
10753 #define vld2_s16(__p0) __extension__ ({ \
10754   int16x4x2_t __ret; \
10755   __builtin_neon_vld2_v(&__ret, __p0, 1); \
10756   __ret; \
10757 })
10758 #else
10759 #define vld2_s16(__p0) __extension__ ({ \
10760   int16x4x2_t __ret; \
10761   __builtin_neon_vld2_v(&__ret, __p0, 1); \
10762  \
10763   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10764   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10765   __ret; \
10766 })
10767 #endif
10768 
10769 #ifdef __LITTLE_ENDIAN__
10770 #define vld2_dup_p8(__p0) __extension__ ({ \
10771   poly8x8x2_t __ret; \
10772   __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
10773   __ret; \
10774 })
10775 #else
10776 #define vld2_dup_p8(__p0) __extension__ ({ \
10777   poly8x8x2_t __ret; \
10778   __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
10779  \
10780   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10781   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10782   __ret; \
10783 })
10784 #endif
10785 
10786 #ifdef __LITTLE_ENDIAN__
10787 #define vld2_dup_p16(__p0) __extension__ ({ \
10788   poly16x4x2_t __ret; \
10789   __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
10790   __ret; \
10791 })
10792 #else
10793 #define vld2_dup_p16(__p0) __extension__ ({ \
10794   poly16x4x2_t __ret; \
10795   __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
10796  \
10797   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10798   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10799   __ret; \
10800 })
10801 #endif
10802 
10803 #ifdef __LITTLE_ENDIAN__
10804 #define vld2q_dup_p8(__p0) __extension__ ({ \
10805   poly8x16x2_t __ret; \
10806   __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
10807   __ret; \
10808 })
10809 #else
10810 #define vld2q_dup_p8(__p0) __extension__ ({ \
10811   poly8x16x2_t __ret; \
10812   __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
10813  \
10814   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10815   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10816   __ret; \
10817 })
10818 #endif
10819 
10820 #ifdef __LITTLE_ENDIAN__
10821 #define vld2q_dup_p16(__p0) __extension__ ({ \
10822   poly16x8x2_t __ret; \
10823   __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
10824   __ret; \
10825 })
10826 #else
10827 #define vld2q_dup_p16(__p0) __extension__ ({ \
10828   poly16x8x2_t __ret; \
10829   __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
10830  \
10831   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10832   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10833   __ret; \
10834 })
10835 #endif
10836 
10837 #ifdef __LITTLE_ENDIAN__
10838 #define vld2q_dup_u8(__p0) __extension__ ({ \
10839   uint8x16x2_t __ret; \
10840   __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
10841   __ret; \
10842 })
10843 #else
10844 #define vld2q_dup_u8(__p0) __extension__ ({ \
10845   uint8x16x2_t __ret; \
10846   __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
10847  \
10848   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10849   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10850   __ret; \
10851 })
10852 #endif
10853 
10854 #ifdef __LITTLE_ENDIAN__
10855 #define vld2q_dup_u32(__p0) __extension__ ({ \
10856   uint32x4x2_t __ret; \
10857   __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
10858   __ret; \
10859 })
10860 #else
10861 #define vld2q_dup_u32(__p0) __extension__ ({ \
10862   uint32x4x2_t __ret; \
10863   __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
10864  \
10865   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10866   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10867   __ret; \
10868 })
10869 #endif
10870 
10871 #ifdef __LITTLE_ENDIAN__
10872 #define vld2q_dup_u64(__p0) __extension__ ({ \
10873   uint64x2x2_t __ret; \
10874   __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
10875   __ret; \
10876 })
10877 #else
10878 #define vld2q_dup_u64(__p0) __extension__ ({ \
10879   uint64x2x2_t __ret; \
10880   __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
10881  \
10882   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10883   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10884   __ret; \
10885 })
10886 #endif
10887 
10888 #ifdef __LITTLE_ENDIAN__
10889 #define vld2q_dup_u16(__p0) __extension__ ({ \
10890   uint16x8x2_t __ret; \
10891   __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
10892   __ret; \
10893 })
10894 #else
10895 #define vld2q_dup_u16(__p0) __extension__ ({ \
10896   uint16x8x2_t __ret; \
10897   __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
10898  \
10899   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10900   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10901   __ret; \
10902 })
10903 #endif
10904 
10905 #ifdef __LITTLE_ENDIAN__
10906 #define vld2q_dup_s8(__p0) __extension__ ({ \
10907   int8x16x2_t __ret; \
10908   __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
10909   __ret; \
10910 })
10911 #else
10912 #define vld2q_dup_s8(__p0) __extension__ ({ \
10913   int8x16x2_t __ret; \
10914   __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
10915  \
10916   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10917   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10918   __ret; \
10919 })
10920 #endif
10921 
10922 #ifdef __LITTLE_ENDIAN__
10923 #define vld2q_dup_f32(__p0) __extension__ ({ \
10924   float32x4x2_t __ret; \
10925   __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
10926   __ret; \
10927 })
10928 #else
10929 #define vld2q_dup_f32(__p0) __extension__ ({ \
10930   float32x4x2_t __ret; \
10931   __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
10932  \
10933   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10934   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10935   __ret; \
10936 })
10937 #endif
10938 
10939 #ifdef __LITTLE_ENDIAN__
10940 #define vld2q_dup_s32(__p0) __extension__ ({ \
10941   int32x4x2_t __ret; \
10942   __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
10943   __ret; \
10944 })
10945 #else
10946 #define vld2q_dup_s32(__p0) __extension__ ({ \
10947   int32x4x2_t __ret; \
10948   __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
10949  \
10950   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10951   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10952   __ret; \
10953 })
10954 #endif
10955 
10956 #ifdef __LITTLE_ENDIAN__
10957 #define vld2q_dup_s64(__p0) __extension__ ({ \
10958   int64x2x2_t __ret; \
10959   __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
10960   __ret; \
10961 })
10962 #else
10963 #define vld2q_dup_s64(__p0) __extension__ ({ \
10964   int64x2x2_t __ret; \
10965   __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
10966  \
10967   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10968   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10969   __ret; \
10970 })
10971 #endif
10972 
10973 #ifdef __LITTLE_ENDIAN__
10974 #define vld2q_dup_s16(__p0) __extension__ ({ \
10975   int16x8x2_t __ret; \
10976   __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
10977   __ret; \
10978 })
10979 #else
10980 #define vld2q_dup_s16(__p0) __extension__ ({ \
10981   int16x8x2_t __ret; \
10982   __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
10983  \
10984   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10985   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10986   __ret; \
10987 })
10988 #endif
10989 
10990 #ifdef __LITTLE_ENDIAN__
10991 #define vld2_dup_u8(__p0) __extension__ ({ \
10992   uint8x8x2_t __ret; \
10993   __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
10994   __ret; \
10995 })
10996 #else
10997 #define vld2_dup_u8(__p0) __extension__ ({ \
10998   uint8x8x2_t __ret; \
10999   __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
11000  \
11001   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11002   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11003   __ret; \
11004 })
11005 #endif
11006 
11007 #ifdef __LITTLE_ENDIAN__
11008 #define vld2_dup_u32(__p0) __extension__ ({ \
11009   uint32x2x2_t __ret; \
11010   __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
11011   __ret; \
11012 })
11013 #else
11014 #define vld2_dup_u32(__p0) __extension__ ({ \
11015   uint32x2x2_t __ret; \
11016   __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
11017  \
11018   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11019   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11020   __ret; \
11021 })
11022 #endif
11023 
11024 #define vld2_dup_u64(__p0) __extension__ ({ \
11025   uint64x1x2_t __ret; \
11026   __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
11027   __ret; \
11028 })
11029 #ifdef __LITTLE_ENDIAN__
11030 #define vld2_dup_u16(__p0) __extension__ ({ \
11031   uint16x4x2_t __ret; \
11032   __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
11033   __ret; \
11034 })
11035 #else
11036 #define vld2_dup_u16(__p0) __extension__ ({ \
11037   uint16x4x2_t __ret; \
11038   __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
11039  \
11040   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11041   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11042   __ret; \
11043 })
11044 #endif
11045 
11046 #ifdef __LITTLE_ENDIAN__
11047 #define vld2_dup_s8(__p0) __extension__ ({ \
11048   int8x8x2_t __ret; \
11049   __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
11050   __ret; \
11051 })
11052 #else
11053 #define vld2_dup_s8(__p0) __extension__ ({ \
11054   int8x8x2_t __ret; \
11055   __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
11056  \
11057   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11058   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11059   __ret; \
11060 })
11061 #endif
11062 
11063 #ifdef __LITTLE_ENDIAN__
11064 #define vld2_dup_f32(__p0) __extension__ ({ \
11065   float32x2x2_t __ret; \
11066   __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
11067   __ret; \
11068 })
11069 #else
11070 #define vld2_dup_f32(__p0) __extension__ ({ \
11071   float32x2x2_t __ret; \
11072   __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
11073  \
11074   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11075   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11076   __ret; \
11077 })
11078 #endif
11079 
11080 #ifdef __LITTLE_ENDIAN__
11081 #define vld2_dup_s32(__p0) __extension__ ({ \
11082   int32x2x2_t __ret; \
11083   __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
11084   __ret; \
11085 })
11086 #else
11087 #define vld2_dup_s32(__p0) __extension__ ({ \
11088   int32x2x2_t __ret; \
11089   __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
11090  \
11091   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11092   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11093   __ret; \
11094 })
11095 #endif
11096 
11097 #define vld2_dup_s64(__p0) __extension__ ({ \
11098   int64x1x2_t __ret; \
11099   __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
11100   __ret; \
11101 })
11102 #ifdef __LITTLE_ENDIAN__
11103 #define vld2_dup_s16(__p0) __extension__ ({ \
11104   int16x4x2_t __ret; \
11105   __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
11106   __ret; \
11107 })
11108 #else
11109 #define vld2_dup_s16(__p0) __extension__ ({ \
11110   int16x4x2_t __ret; \
11111   __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
11112  \
11113   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11114   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11115   __ret; \
11116 })
11117 #endif
11118 
11119 #ifdef __LITTLE_ENDIAN__
11120 #define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11121   poly8x8x2_t __ret; \
11122   poly8x8x2_t __s1 = __p1; \
11123   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
11124   __ret; \
11125 })
11126 #else
11127 #define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11128   poly8x8x2_t __ret; \
11129   poly8x8x2_t __s1 = __p1; \
11130   poly8x8x2_t __rev1; \
11131   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11132   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11133   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
11134  \
11135   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11136   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11137   __ret; \
11138 })
11139 #endif
11140 
11141 #ifdef __LITTLE_ENDIAN__
11142 #define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11143   poly16x4x2_t __ret; \
11144   poly16x4x2_t __s1 = __p1; \
11145   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
11146   __ret; \
11147 })
11148 #else
11149 #define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11150   poly16x4x2_t __ret; \
11151   poly16x4x2_t __s1 = __p1; \
11152   poly16x4x2_t __rev1; \
11153   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11154   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11155   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
11156  \
11157   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11158   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11159   __ret; \
11160 })
11161 #endif
11162 
11163 #ifdef __LITTLE_ENDIAN__
11164 #define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11165   poly16x8x2_t __ret; \
11166   poly16x8x2_t __s1 = __p1; \
11167   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
11168   __ret; \
11169 })
11170 #else
11171 #define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11172   poly16x8x2_t __ret; \
11173   poly16x8x2_t __s1 = __p1; \
11174   poly16x8x2_t __rev1; \
11175   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11176   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11177   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
11178  \
11179   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11180   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11181   __ret; \
11182 })
11183 #endif
11184 
11185 #ifdef __LITTLE_ENDIAN__
11186 #define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11187   uint32x4x2_t __ret; \
11188   uint32x4x2_t __s1 = __p1; \
11189   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
11190   __ret; \
11191 })
11192 #else
11193 #define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11194   uint32x4x2_t __ret; \
11195   uint32x4x2_t __s1 = __p1; \
11196   uint32x4x2_t __rev1; \
11197   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11198   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11199   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
11200  \
11201   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11202   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11203   __ret; \
11204 })
11205 #endif
11206 
11207 #ifdef __LITTLE_ENDIAN__
11208 #define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11209   uint16x8x2_t __ret; \
11210   uint16x8x2_t __s1 = __p1; \
11211   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
11212   __ret; \
11213 })
11214 #else
11215 #define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11216   uint16x8x2_t __ret; \
11217   uint16x8x2_t __s1 = __p1; \
11218   uint16x8x2_t __rev1; \
11219   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11220   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11221   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
11222  \
11223   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11224   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11225   __ret; \
11226 })
11227 #endif
11228 
11229 #ifdef __LITTLE_ENDIAN__
11230 #define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11231   float32x4x2_t __ret; \
11232   float32x4x2_t __s1 = __p1; \
11233   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 41); \
11234   __ret; \
11235 })
11236 #else
11237 #define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11238   float32x4x2_t __ret; \
11239   float32x4x2_t __s1 = __p1; \
11240   float32x4x2_t __rev1; \
11241   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11242   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11243   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 41); \
11244  \
11245   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11246   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11247   __ret; \
11248 })
11249 #endif
11250 
11251 #ifdef __LITTLE_ENDIAN__
11252 #define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11253   int32x4x2_t __ret; \
11254   int32x4x2_t __s1 = __p1; \
11255   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 34); \
11256   __ret; \
11257 })
11258 #else
11259 #define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11260   int32x4x2_t __ret; \
11261   int32x4x2_t __s1 = __p1; \
11262   int32x4x2_t __rev1; \
11263   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11264   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11265   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 34); \
11266  \
11267   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11268   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11269   __ret; \
11270 })
11271 #endif
11272 
11273 #ifdef __LITTLE_ENDIAN__
11274 #define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11275   int16x8x2_t __ret; \
11276   int16x8x2_t __s1 = __p1; \
11277   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 33); \
11278   __ret; \
11279 })
11280 #else
11281 #define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11282   int16x8x2_t __ret; \
11283   int16x8x2_t __s1 = __p1; \
11284   int16x8x2_t __rev1; \
11285   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11286   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11287   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 33); \
11288  \
11289   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11290   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11291   __ret; \
11292 })
11293 #endif
11294 
11295 #ifdef __LITTLE_ENDIAN__
11296 #define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11297   uint8x8x2_t __ret; \
11298   uint8x8x2_t __s1 = __p1; \
11299   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
11300   __ret; \
11301 })
11302 #else
11303 #define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11304   uint8x8x2_t __ret; \
11305   uint8x8x2_t __s1 = __p1; \
11306   uint8x8x2_t __rev1; \
11307   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11308   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11309   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
11310  \
11311   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11312   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11313   __ret; \
11314 })
11315 #endif
11316 
11317 #ifdef __LITTLE_ENDIAN__
11318 #define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11319   uint32x2x2_t __ret; \
11320   uint32x2x2_t __s1 = __p1; \
11321   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
11322   __ret; \
11323 })
11324 #else
11325 #define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11326   uint32x2x2_t __ret; \
11327   uint32x2x2_t __s1 = __p1; \
11328   uint32x2x2_t __rev1; \
11329   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11330   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11331   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
11332  \
11333   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11334   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11335   __ret; \
11336 })
11337 #endif
11338 
11339 #ifdef __LITTLE_ENDIAN__
11340 #define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11341   uint16x4x2_t __ret; \
11342   uint16x4x2_t __s1 = __p1; \
11343   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
11344   __ret; \
11345 })
11346 #else
11347 #define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11348   uint16x4x2_t __ret; \
11349   uint16x4x2_t __s1 = __p1; \
11350   uint16x4x2_t __rev1; \
11351   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11352   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11353   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
11354  \
11355   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11356   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11357   __ret; \
11358 })
11359 #endif
11360 
11361 #ifdef __LITTLE_ENDIAN__
11362 #define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11363   int8x8x2_t __ret; \
11364   int8x8x2_t __s1 = __p1; \
11365   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
11366   __ret; \
11367 })
11368 #else
11369 #define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11370   int8x8x2_t __ret; \
11371   int8x8x2_t __s1 = __p1; \
11372   int8x8x2_t __rev1; \
11373   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11374   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11375   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
11376  \
11377   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11378   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11379   __ret; \
11380 })
11381 #endif
11382 
11383 #ifdef __LITTLE_ENDIAN__
11384 #define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11385   float32x2x2_t __ret; \
11386   float32x2x2_t __s1 = __p1; \
11387   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 9); \
11388   __ret; \
11389 })
11390 #else
11391 #define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11392   float32x2x2_t __ret; \
11393   float32x2x2_t __s1 = __p1; \
11394   float32x2x2_t __rev1; \
11395   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11396   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11397   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 9); \
11398  \
11399   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11400   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11401   __ret; \
11402 })
11403 #endif
11404 
11405 #ifdef __LITTLE_ENDIAN__
11406 #define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11407   int32x2x2_t __ret; \
11408   int32x2x2_t __s1 = __p1; \
11409   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 2); \
11410   __ret; \
11411 })
11412 #else
11413 #define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11414   int32x2x2_t __ret; \
11415   int32x2x2_t __s1 = __p1; \
11416   int32x2x2_t __rev1; \
11417   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11418   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11419   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 2); \
11420  \
11421   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11422   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11423   __ret; \
11424 })
11425 #endif
11426 
11427 #ifdef __LITTLE_ENDIAN__
11428 #define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11429   int16x4x2_t __ret; \
11430   int16x4x2_t __s1 = __p1; \
11431   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 1); \
11432   __ret; \
11433 })
11434 #else
11435 #define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11436   int16x4x2_t __ret; \
11437   int16x4x2_t __s1 = __p1; \
11438   int16x4x2_t __rev1; \
11439   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11440   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11441   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 1); \
11442  \
11443   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11444   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11445   __ret; \
11446 })
11447 #endif
11448 
11449 #ifdef __LITTLE_ENDIAN__
11450 #define vld3_p8(__p0) __extension__ ({ \
11451   poly8x8x3_t __ret; \
11452   __builtin_neon_vld3_v(&__ret, __p0, 4); \
11453   __ret; \
11454 })
11455 #else
11456 #define vld3_p8(__p0) __extension__ ({ \
11457   poly8x8x3_t __ret; \
11458   __builtin_neon_vld3_v(&__ret, __p0, 4); \
11459  \
11460   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11461   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11462   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11463   __ret; \
11464 })
11465 #endif
11466 
11467 #ifdef __LITTLE_ENDIAN__
11468 #define vld3_p16(__p0) __extension__ ({ \
11469   poly16x4x3_t __ret; \
11470   __builtin_neon_vld3_v(&__ret, __p0, 5); \
11471   __ret; \
11472 })
11473 #else
11474 #define vld3_p16(__p0) __extension__ ({ \
11475   poly16x4x3_t __ret; \
11476   __builtin_neon_vld3_v(&__ret, __p0, 5); \
11477  \
11478   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11479   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11480   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11481   __ret; \
11482 })
11483 #endif
11484 
11485 #ifdef __LITTLE_ENDIAN__
11486 #define vld3q_p8(__p0) __extension__ ({ \
11487   poly8x16x3_t __ret; \
11488   __builtin_neon_vld3q_v(&__ret, __p0, 36); \
11489   __ret; \
11490 })
11491 #else
11492 #define vld3q_p8(__p0) __extension__ ({ \
11493   poly8x16x3_t __ret; \
11494   __builtin_neon_vld3q_v(&__ret, __p0, 36); \
11495  \
11496   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11497   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11498   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11499   __ret; \
11500 })
11501 #endif
11502 
11503 #ifdef __LITTLE_ENDIAN__
11504 #define vld3q_p16(__p0) __extension__ ({ \
11505   poly16x8x3_t __ret; \
11506   __builtin_neon_vld3q_v(&__ret, __p0, 37); \
11507   __ret; \
11508 })
11509 #else
11510 #define vld3q_p16(__p0) __extension__ ({ \
11511   poly16x8x3_t __ret; \
11512   __builtin_neon_vld3q_v(&__ret, __p0, 37); \
11513  \
11514   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11515   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11516   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11517   __ret; \
11518 })
11519 #endif
11520 
11521 #ifdef __LITTLE_ENDIAN__
11522 #define vld3q_u8(__p0) __extension__ ({ \
11523   uint8x16x3_t __ret; \
11524   __builtin_neon_vld3q_v(&__ret, __p0, 48); \
11525   __ret; \
11526 })
11527 #else
11528 #define vld3q_u8(__p0) __extension__ ({ \
11529   uint8x16x3_t __ret; \
11530   __builtin_neon_vld3q_v(&__ret, __p0, 48); \
11531  \
11532   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11533   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11534   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11535   __ret; \
11536 })
11537 #endif
11538 
11539 #ifdef __LITTLE_ENDIAN__
11540 #define vld3q_u32(__p0) __extension__ ({ \
11541   uint32x4x3_t __ret; \
11542   __builtin_neon_vld3q_v(&__ret, __p0, 50); \
11543   __ret; \
11544 })
11545 #else
11546 #define vld3q_u32(__p0) __extension__ ({ \
11547   uint32x4x3_t __ret; \
11548   __builtin_neon_vld3q_v(&__ret, __p0, 50); \
11549  \
11550   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11551   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11552   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11553   __ret; \
11554 })
11555 #endif
11556 
11557 #ifdef __LITTLE_ENDIAN__
11558 #define vld3q_u16(__p0) __extension__ ({ \
11559   uint16x8x3_t __ret; \
11560   __builtin_neon_vld3q_v(&__ret, __p0, 49); \
11561   __ret; \
11562 })
11563 #else
11564 #define vld3q_u16(__p0) __extension__ ({ \
11565   uint16x8x3_t __ret; \
11566   __builtin_neon_vld3q_v(&__ret, __p0, 49); \
11567  \
11568   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11569   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11570   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11571   __ret; \
11572 })
11573 #endif
11574 
11575 #ifdef __LITTLE_ENDIAN__
11576 #define vld3q_s8(__p0) __extension__ ({ \
11577   int8x16x3_t __ret; \
11578   __builtin_neon_vld3q_v(&__ret, __p0, 32); \
11579   __ret; \
11580 })
11581 #else
11582 #define vld3q_s8(__p0) __extension__ ({ \
11583   int8x16x3_t __ret; \
11584   __builtin_neon_vld3q_v(&__ret, __p0, 32); \
11585  \
11586   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11587   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11588   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11589   __ret; \
11590 })
11591 #endif
11592 
11593 #ifdef __LITTLE_ENDIAN__
11594 #define vld3q_f32(__p0) __extension__ ({ \
11595   float32x4x3_t __ret; \
11596   __builtin_neon_vld3q_v(&__ret, __p0, 41); \
11597   __ret; \
11598 })
11599 #else
11600 #define vld3q_f32(__p0) __extension__ ({ \
11601   float32x4x3_t __ret; \
11602   __builtin_neon_vld3q_v(&__ret, __p0, 41); \
11603  \
11604   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11605   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11606   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11607   __ret; \
11608 })
11609 #endif
11610 
11611 #ifdef __LITTLE_ENDIAN__
11612 #define vld3q_s32(__p0) __extension__ ({ \
11613   int32x4x3_t __ret; \
11614   __builtin_neon_vld3q_v(&__ret, __p0, 34); \
11615   __ret; \
11616 })
11617 #else
11618 #define vld3q_s32(__p0) __extension__ ({ \
11619   int32x4x3_t __ret; \
11620   __builtin_neon_vld3q_v(&__ret, __p0, 34); \
11621  \
11622   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11623   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11624   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11625   __ret; \
11626 })
11627 #endif
11628 
11629 #ifdef __LITTLE_ENDIAN__
11630 #define vld3q_s16(__p0) __extension__ ({ \
11631   int16x8x3_t __ret; \
11632   __builtin_neon_vld3q_v(&__ret, __p0, 33); \
11633   __ret; \
11634 })
11635 #else
11636 #define vld3q_s16(__p0) __extension__ ({ \
11637   int16x8x3_t __ret; \
11638   __builtin_neon_vld3q_v(&__ret, __p0, 33); \
11639  \
11640   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11641   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11642   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11643   __ret; \
11644 })
11645 #endif
11646 
11647 #ifdef __LITTLE_ENDIAN__
11648 #define vld3_u8(__p0) __extension__ ({ \
11649   uint8x8x3_t __ret; \
11650   __builtin_neon_vld3_v(&__ret, __p0, 16); \
11651   __ret; \
11652 })
11653 #else
11654 #define vld3_u8(__p0) __extension__ ({ \
11655   uint8x8x3_t __ret; \
11656   __builtin_neon_vld3_v(&__ret, __p0, 16); \
11657  \
11658   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11659   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11660   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11661   __ret; \
11662 })
11663 #endif
11664 
11665 #ifdef __LITTLE_ENDIAN__
11666 #define vld3_u32(__p0) __extension__ ({ \
11667   uint32x2x3_t __ret; \
11668   __builtin_neon_vld3_v(&__ret, __p0, 18); \
11669   __ret; \
11670 })
11671 #else
11672 #define vld3_u32(__p0) __extension__ ({ \
11673   uint32x2x3_t __ret; \
11674   __builtin_neon_vld3_v(&__ret, __p0, 18); \
11675  \
11676   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11677   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11678   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11679   __ret; \
11680 })
11681 #endif
11682 
11683 #define vld3_u64(__p0) __extension__ ({ \
11684   uint64x1x3_t __ret; \
11685   __builtin_neon_vld3_v(&__ret, __p0, 19); \
11686   __ret; \
11687 })
11688 #ifdef __LITTLE_ENDIAN__
11689 #define vld3_u16(__p0) __extension__ ({ \
11690   uint16x4x3_t __ret; \
11691   __builtin_neon_vld3_v(&__ret, __p0, 17); \
11692   __ret; \
11693 })
11694 #else
11695 #define vld3_u16(__p0) __extension__ ({ \
11696   uint16x4x3_t __ret; \
11697   __builtin_neon_vld3_v(&__ret, __p0, 17); \
11698  \
11699   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11700   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11701   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11702   __ret; \
11703 })
11704 #endif
11705 
11706 #ifdef __LITTLE_ENDIAN__
11707 #define vld3_s8(__p0) __extension__ ({ \
11708   int8x8x3_t __ret; \
11709   __builtin_neon_vld3_v(&__ret, __p0, 0); \
11710   __ret; \
11711 })
11712 #else
11713 #define vld3_s8(__p0) __extension__ ({ \
11714   int8x8x3_t __ret; \
11715   __builtin_neon_vld3_v(&__ret, __p0, 0); \
11716  \
11717   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11718   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11719   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11720   __ret; \
11721 })
11722 #endif
11723 
11724 #ifdef __LITTLE_ENDIAN__
11725 #define vld3_f32(__p0) __extension__ ({ \
11726   float32x2x3_t __ret; \
11727   __builtin_neon_vld3_v(&__ret, __p0, 9); \
11728   __ret; \
11729 })
11730 #else
11731 #define vld3_f32(__p0) __extension__ ({ \
11732   float32x2x3_t __ret; \
11733   __builtin_neon_vld3_v(&__ret, __p0, 9); \
11734  \
11735   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11736   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11737   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11738   __ret; \
11739 })
11740 #endif
11741 
11742 #ifdef __LITTLE_ENDIAN__
11743 #define vld3_s32(__p0) __extension__ ({ \
11744   int32x2x3_t __ret; \
11745   __builtin_neon_vld3_v(&__ret, __p0, 2); \
11746   __ret; \
11747 })
11748 #else
11749 #define vld3_s32(__p0) __extension__ ({ \
11750   int32x2x3_t __ret; \
11751   __builtin_neon_vld3_v(&__ret, __p0, 2); \
11752  \
11753   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11754   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11755   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11756   __ret; \
11757 })
11758 #endif
11759 
11760 #define vld3_s64(__p0) __extension__ ({ \
11761   int64x1x3_t __ret; \
11762   __builtin_neon_vld3_v(&__ret, __p0, 3); \
11763   __ret; \
11764 })
11765 #ifdef __LITTLE_ENDIAN__
11766 #define vld3_s16(__p0) __extension__ ({ \
11767   int16x4x3_t __ret; \
11768   __builtin_neon_vld3_v(&__ret, __p0, 1); \
11769   __ret; \
11770 })
11771 #else
11772 #define vld3_s16(__p0) __extension__ ({ \
11773   int16x4x3_t __ret; \
11774   __builtin_neon_vld3_v(&__ret, __p0, 1); \
11775  \
11776   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11777   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11778   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11779   __ret; \
11780 })
11781 #endif
11782 
11783 #ifdef __LITTLE_ENDIAN__
11784 #define vld3_dup_p8(__p0) __extension__ ({ \
11785   poly8x8x3_t __ret; \
11786   __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
11787   __ret; \
11788 })
11789 #else
11790 #define vld3_dup_p8(__p0) __extension__ ({ \
11791   poly8x8x3_t __ret; \
11792   __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
11793  \
11794   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11795   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11796   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11797   __ret; \
11798 })
11799 #endif
11800 
11801 #ifdef __LITTLE_ENDIAN__
11802 #define vld3_dup_p16(__p0) __extension__ ({ \
11803   poly16x4x3_t __ret; \
11804   __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
11805   __ret; \
11806 })
11807 #else
11808 #define vld3_dup_p16(__p0) __extension__ ({ \
11809   poly16x4x3_t __ret; \
11810   __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
11811  \
11812   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11813   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11814   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11815   __ret; \
11816 })
11817 #endif
11818 
11819 #ifdef __LITTLE_ENDIAN__
11820 #define vld3q_dup_p8(__p0) __extension__ ({ \
11821   poly8x16x3_t __ret; \
11822   __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
11823   __ret; \
11824 })
11825 #else
11826 #define vld3q_dup_p8(__p0) __extension__ ({ \
11827   poly8x16x3_t __ret; \
11828   __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
11829  \
11830   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11831   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11832   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11833   __ret; \
11834 })
11835 #endif
11836 
11837 #ifdef __LITTLE_ENDIAN__
11838 #define vld3q_dup_p16(__p0) __extension__ ({ \
11839   poly16x8x3_t __ret; \
11840   __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
11841   __ret; \
11842 })
11843 #else
11844 #define vld3q_dup_p16(__p0) __extension__ ({ \
11845   poly16x8x3_t __ret; \
11846   __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
11847  \
11848   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11849   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11850   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11851   __ret; \
11852 })
11853 #endif
11854 
11855 #ifdef __LITTLE_ENDIAN__
11856 #define vld3q_dup_u8(__p0) __extension__ ({ \
11857   uint8x16x3_t __ret; \
11858   __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
11859   __ret; \
11860 })
11861 #else
11862 #define vld3q_dup_u8(__p0) __extension__ ({ \
11863   uint8x16x3_t __ret; \
11864   __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
11865  \
11866   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11867   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11868   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11869   __ret; \
11870 })
11871 #endif
11872 
11873 #ifdef __LITTLE_ENDIAN__
11874 #define vld3q_dup_u32(__p0) __extension__ ({ \
11875   uint32x4x3_t __ret; \
11876   __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
11877   __ret; \
11878 })
11879 #else
11880 #define vld3q_dup_u32(__p0) __extension__ ({ \
11881   uint32x4x3_t __ret; \
11882   __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
11883  \
11884   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11885   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11886   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11887   __ret; \
11888 })
11889 #endif
11890 
11891 #ifdef __LITTLE_ENDIAN__
11892 #define vld3q_dup_u64(__p0) __extension__ ({ \
11893   uint64x2x3_t __ret; \
11894   __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
11895   __ret; \
11896 })
11897 #else
11898 #define vld3q_dup_u64(__p0) __extension__ ({ \
11899   uint64x2x3_t __ret; \
11900   __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
11901  \
11902   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11903   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11904   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11905   __ret; \
11906 })
11907 #endif
11908 
11909 #ifdef __LITTLE_ENDIAN__
11910 #define vld3q_dup_u16(__p0) __extension__ ({ \
11911   uint16x8x3_t __ret; \
11912   __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
11913   __ret; \
11914 })
11915 #else
11916 #define vld3q_dup_u16(__p0) __extension__ ({ \
11917   uint16x8x3_t __ret; \
11918   __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
11919  \
11920   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11921   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11922   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11923   __ret; \
11924 })
11925 #endif
11926 
11927 #ifdef __LITTLE_ENDIAN__
11928 #define vld3q_dup_s8(__p0) __extension__ ({ \
11929   int8x16x3_t __ret; \
11930   __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
11931   __ret; \
11932 })
11933 #else
11934 #define vld3q_dup_s8(__p0) __extension__ ({ \
11935   int8x16x3_t __ret; \
11936   __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
11937  \
11938   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11939   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11940   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11941   __ret; \
11942 })
11943 #endif
11944 
11945 #ifdef __LITTLE_ENDIAN__
11946 #define vld3q_dup_f32(__p0) __extension__ ({ \
11947   float32x4x3_t __ret; \
11948   __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
11949   __ret; \
11950 })
11951 #else
11952 #define vld3q_dup_f32(__p0) __extension__ ({ \
11953   float32x4x3_t __ret; \
11954   __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
11955  \
11956   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11957   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11958   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11959   __ret; \
11960 })
11961 #endif
11962 
11963 #ifdef __LITTLE_ENDIAN__
11964 #define vld3q_dup_s32(__p0) __extension__ ({ \
11965   int32x4x3_t __ret; \
11966   __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
11967   __ret; \
11968 })
11969 #else
11970 #define vld3q_dup_s32(__p0) __extension__ ({ \
11971   int32x4x3_t __ret; \
11972   __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
11973  \
11974   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11975   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11976   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11977   __ret; \
11978 })
11979 #endif
11980 
11981 #ifdef __LITTLE_ENDIAN__
11982 #define vld3q_dup_s64(__p0) __extension__ ({ \
11983   int64x2x3_t __ret; \
11984   __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
11985   __ret; \
11986 })
11987 #else
11988 #define vld3q_dup_s64(__p0) __extension__ ({ \
11989   int64x2x3_t __ret; \
11990   __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
11991  \
11992   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11993   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11994   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11995   __ret; \
11996 })
11997 #endif
11998 
11999 #ifdef __LITTLE_ENDIAN__
12000 #define vld3q_dup_s16(__p0) __extension__ ({ \
12001   int16x8x3_t __ret; \
12002   __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
12003   __ret; \
12004 })
12005 #else
12006 #define vld3q_dup_s16(__p0) __extension__ ({ \
12007   int16x8x3_t __ret; \
12008   __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
12009  \
12010   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12011   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12012   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12013   __ret; \
12014 })
12015 #endif
12016 
12017 #ifdef __LITTLE_ENDIAN__
12018 #define vld3_dup_u8(__p0) __extension__ ({ \
12019   uint8x8x3_t __ret; \
12020   __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
12021   __ret; \
12022 })
12023 #else
12024 #define vld3_dup_u8(__p0) __extension__ ({ \
12025   uint8x8x3_t __ret; \
12026   __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
12027  \
12028   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12029   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12030   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12031   __ret; \
12032 })
12033 #endif
12034 
12035 #ifdef __LITTLE_ENDIAN__
12036 #define vld3_dup_u32(__p0) __extension__ ({ \
12037   uint32x2x3_t __ret; \
12038   __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
12039   __ret; \
12040 })
12041 #else
12042 #define vld3_dup_u32(__p0) __extension__ ({ \
12043   uint32x2x3_t __ret; \
12044   __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
12045  \
12046   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12047   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12048   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12049   __ret; \
12050 })
12051 #endif
12052 
12053 #define vld3_dup_u64(__p0) __extension__ ({ \
12054   uint64x1x3_t __ret; \
12055   __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
12056   __ret; \
12057 })
12058 #ifdef __LITTLE_ENDIAN__
12059 #define vld3_dup_u16(__p0) __extension__ ({ \
12060   uint16x4x3_t __ret; \
12061   __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
12062   __ret; \
12063 })
12064 #else
12065 #define vld3_dup_u16(__p0) __extension__ ({ \
12066   uint16x4x3_t __ret; \
12067   __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
12068  \
12069   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12070   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12071   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12072   __ret; \
12073 })
12074 #endif
12075 
12076 #ifdef __LITTLE_ENDIAN__
12077 #define vld3_dup_s8(__p0) __extension__ ({ \
12078   int8x8x3_t __ret; \
12079   __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
12080   __ret; \
12081 })
12082 #else
12083 #define vld3_dup_s8(__p0) __extension__ ({ \
12084   int8x8x3_t __ret; \
12085   __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
12086  \
12087   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12088   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12089   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12090   __ret; \
12091 })
12092 #endif
12093 
12094 #ifdef __LITTLE_ENDIAN__
12095 #define vld3_dup_f32(__p0) __extension__ ({ \
12096   float32x2x3_t __ret; \
12097   __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
12098   __ret; \
12099 })
12100 #else
12101 #define vld3_dup_f32(__p0) __extension__ ({ \
12102   float32x2x3_t __ret; \
12103   __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
12104  \
12105   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12106   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12107   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12108   __ret; \
12109 })
12110 #endif
12111 
12112 #ifdef __LITTLE_ENDIAN__
12113 #define vld3_dup_s32(__p0) __extension__ ({ \
12114   int32x2x3_t __ret; \
12115   __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
12116   __ret; \
12117 })
12118 #else
12119 #define vld3_dup_s32(__p0) __extension__ ({ \
12120   int32x2x3_t __ret; \
12121   __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
12122  \
12123   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12124   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12125   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12126   __ret; \
12127 })
12128 #endif
12129 
12130 #define vld3_dup_s64(__p0) __extension__ ({ \
12131   int64x1x3_t __ret; \
12132   __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
12133   __ret; \
12134 })
12135 #ifdef __LITTLE_ENDIAN__
12136 #define vld3_dup_s16(__p0) __extension__ ({ \
12137   int16x4x3_t __ret; \
12138   __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
12139   __ret; \
12140 })
12141 #else
12142 #define vld3_dup_s16(__p0) __extension__ ({ \
12143   int16x4x3_t __ret; \
12144   __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
12145  \
12146   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12147   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12148   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12149   __ret; \
12150 })
12151 #endif
12152 
12153 #ifdef __LITTLE_ENDIAN__
12154 #define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
12155   poly8x8x3_t __ret; \
12156   poly8x8x3_t __s1 = __p1; \
12157   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
12158   __ret; \
12159 })
12160 #else
12161 #define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
12162   poly8x8x3_t __ret; \
12163   poly8x8x3_t __s1 = __p1; \
12164   poly8x8x3_t __rev1; \
12165   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12166   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12167   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12168   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
12169  \
12170   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12171   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12172   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12173   __ret; \
12174 })
12175 #endif
12176 
12177 #ifdef __LITTLE_ENDIAN__
12178 #define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
12179   poly16x4x3_t __ret; \
12180   poly16x4x3_t __s1 = __p1; \
12181   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
12182   __ret; \
12183 })
12184 #else
12185 #define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
12186   poly16x4x3_t __ret; \
12187   poly16x4x3_t __s1 = __p1; \
12188   poly16x4x3_t __rev1; \
12189   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12190   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12191   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12192   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
12193  \
12194   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12195   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12196   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12197   __ret; \
12198 })
12199 #endif
12200 
12201 #ifdef __LITTLE_ENDIAN__
12202 #define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
12203   poly16x8x3_t __ret; \
12204   poly16x8x3_t __s1 = __p1; \
12205   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
12206   __ret; \
12207 })
12208 #else
12209 #define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
12210   poly16x8x3_t __ret; \
12211   poly16x8x3_t __s1 = __p1; \
12212   poly16x8x3_t __rev1; \
12213   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12214   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12215   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12216   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
12217  \
12218   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12219   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12220   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12221   __ret; \
12222 })
12223 #endif
12224 
12225 #ifdef __LITTLE_ENDIAN__
12226 #define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
12227   uint32x4x3_t __ret; \
12228   uint32x4x3_t __s1 = __p1; \
12229   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
12230   __ret; \
12231 })
12232 #else
12233 #define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
12234   uint32x4x3_t __ret; \
12235   uint32x4x3_t __s1 = __p1; \
12236   uint32x4x3_t __rev1; \
12237   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12238   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12239   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12240   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
12241  \
12242   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12243   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12244   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12245   __ret; \
12246 })
12247 #endif
12248 
12249 #ifdef __LITTLE_ENDIAN__
12250 #define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
12251   uint16x8x3_t __ret; \
12252   uint16x8x3_t __s1 = __p1; \
12253   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
12254   __ret; \
12255 })
12256 #else
12257 #define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
12258   uint16x8x3_t __ret; \
12259   uint16x8x3_t __s1 = __p1; \
12260   uint16x8x3_t __rev1; \
12261   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12262   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12263   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12264   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
12265  \
12266   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12267   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12268   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12269   __ret; \
12270 })
12271 #endif
12272 
12273 #ifdef __LITTLE_ENDIAN__
12274 #define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
12275   float32x4x3_t __ret; \
12276   float32x4x3_t __s1 = __p1; \
12277   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 41); \
12278   __ret; \
12279 })
12280 #else
12281 #define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
12282   float32x4x3_t __ret; \
12283   float32x4x3_t __s1 = __p1; \
12284   float32x4x3_t __rev1; \
12285   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12286   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12287   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12288   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 41); \
12289  \
12290   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12291   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12292   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12293   __ret; \
12294 })
12295 #endif
12296 
12297 #ifdef __LITTLE_ENDIAN__
12298 #define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
12299   int32x4x3_t __ret; \
12300   int32x4x3_t __s1 = __p1; \
12301   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 34); \
12302   __ret; \
12303 })
12304 #else
12305 #define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
12306   int32x4x3_t __ret; \
12307   int32x4x3_t __s1 = __p1; \
12308   int32x4x3_t __rev1; \
12309   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12310   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12311   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12312   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 34); \
12313  \
12314   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12315   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12316   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12317   __ret; \
12318 })
12319 #endif
12320 
12321 #ifdef __LITTLE_ENDIAN__
12322 #define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
12323   int16x8x3_t __ret; \
12324   int16x8x3_t __s1 = __p1; \
12325   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 33); \
12326   __ret; \
12327 })
12328 #else
12329 #define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
12330   int16x8x3_t __ret; \
12331   int16x8x3_t __s1 = __p1; \
12332   int16x8x3_t __rev1; \
12333   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12334   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12335   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12336   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 33); \
12337  \
12338   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12339   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12340   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12341   __ret; \
12342 })
12343 #endif
12344 
12345 #ifdef __LITTLE_ENDIAN__
12346 #define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
12347   uint8x8x3_t __ret; \
12348   uint8x8x3_t __s1 = __p1; \
12349   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
12350   __ret; \
12351 })
12352 #else
12353 #define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
12354   uint8x8x3_t __ret; \
12355   uint8x8x3_t __s1 = __p1; \
12356   uint8x8x3_t __rev1; \
12357   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12358   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12359   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12360   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
12361  \
12362   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12363   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12364   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12365   __ret; \
12366 })
12367 #endif
12368 
12369 #ifdef __LITTLE_ENDIAN__
12370 #define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
12371   uint32x2x3_t __ret; \
12372   uint32x2x3_t __s1 = __p1; \
12373   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
12374   __ret; \
12375 })
12376 #else
12377 #define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
12378   uint32x2x3_t __ret; \
12379   uint32x2x3_t __s1 = __p1; \
12380   uint32x2x3_t __rev1; \
12381   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
12382   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
12383   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
12384   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
12385  \
12386   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12387   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12388   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12389   __ret; \
12390 })
12391 #endif
12392 
12393 #ifdef __LITTLE_ENDIAN__
12394 #define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
12395   uint16x4x3_t __ret; \
12396   uint16x4x3_t __s1 = __p1; \
12397   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
12398   __ret; \
12399 })
12400 #else
12401 #define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
12402   uint16x4x3_t __ret; \
12403   uint16x4x3_t __s1 = __p1; \
12404   uint16x4x3_t __rev1; \
12405   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12406   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12407   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12408   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
12409  \
12410   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12411   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12412   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12413   __ret; \
12414 })
12415 #endif
12416 
12417 #ifdef __LITTLE_ENDIAN__
12418 #define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
12419   int8x8x3_t __ret; \
12420   int8x8x3_t __s1 = __p1; \
12421   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
12422   __ret; \
12423 })
12424 #else
12425 #define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
12426   int8x8x3_t __ret; \
12427   int8x8x3_t __s1 = __p1; \
12428   int8x8x3_t __rev1; \
12429   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12430   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12431   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12432   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
12433  \
12434   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12435   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12436   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12437   __ret; \
12438 })
12439 #endif
12440 
12441 #ifdef __LITTLE_ENDIAN__
12442 #define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
12443   float32x2x3_t __ret; \
12444   float32x2x3_t __s1 = __p1; \
12445   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 9); \
12446   __ret; \
12447 })
12448 #else
12449 #define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
12450   float32x2x3_t __ret; \
12451   float32x2x3_t __s1 = __p1; \
12452   float32x2x3_t __rev1; \
12453   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
12454   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
12455   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
12456   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 9); \
12457  \
12458   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12459   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12460   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12461   __ret; \
12462 })
12463 #endif
12464 
12465 #ifdef __LITTLE_ENDIAN__
12466 #define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
12467   int32x2x3_t __ret; \
12468   int32x2x3_t __s1 = __p1; \
12469   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 2); \
12470   __ret; \
12471 })
12472 #else
12473 #define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
12474   int32x2x3_t __ret; \
12475   int32x2x3_t __s1 = __p1; \
12476   int32x2x3_t __rev1; \
12477   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
12478   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
12479   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
12480   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 2); \
12481  \
12482   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12483   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12484   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12485   __ret; \
12486 })
12487 #endif
12488 
12489 #ifdef __LITTLE_ENDIAN__
12490 #define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
12491   int16x4x3_t __ret; \
12492   int16x4x3_t __s1 = __p1; \
12493   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 1); \
12494   __ret; \
12495 })
12496 #else
12497 #define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
12498   int16x4x3_t __ret; \
12499   int16x4x3_t __s1 = __p1; \
12500   int16x4x3_t __rev1; \
12501   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12502   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12503   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12504   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 1); \
12505  \
12506   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12507   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12508   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12509   __ret; \
12510 })
12511 #endif
12512 
12513 #ifdef __LITTLE_ENDIAN__
12514 #define vld4_p8(__p0) __extension__ ({ \
12515   poly8x8x4_t __ret; \
12516   __builtin_neon_vld4_v(&__ret, __p0, 4); \
12517   __ret; \
12518 })
12519 #else
12520 #define vld4_p8(__p0) __extension__ ({ \
12521   poly8x8x4_t __ret; \
12522   __builtin_neon_vld4_v(&__ret, __p0, 4); \
12523  \
12524   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12525   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12526   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12527   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12528   __ret; \
12529 })
12530 #endif
12531 
12532 #ifdef __LITTLE_ENDIAN__
12533 #define vld4_p16(__p0) __extension__ ({ \
12534   poly16x4x4_t __ret; \
12535   __builtin_neon_vld4_v(&__ret, __p0, 5); \
12536   __ret; \
12537 })
12538 #else
12539 #define vld4_p16(__p0) __extension__ ({ \
12540   poly16x4x4_t __ret; \
12541   __builtin_neon_vld4_v(&__ret, __p0, 5); \
12542  \
12543   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12544   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12545   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12546   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12547   __ret; \
12548 })
12549 #endif
12550 
12551 #ifdef __LITTLE_ENDIAN__
12552 #define vld4q_p8(__p0) __extension__ ({ \
12553   poly8x16x4_t __ret; \
12554   __builtin_neon_vld4q_v(&__ret, __p0, 36); \
12555   __ret; \
12556 })
12557 #else
12558 #define vld4q_p8(__p0) __extension__ ({ \
12559   poly8x16x4_t __ret; \
12560   __builtin_neon_vld4q_v(&__ret, __p0, 36); \
12561  \
12562   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12563   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12564   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12565   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12566   __ret; \
12567 })
12568 #endif
12569 
12570 #ifdef __LITTLE_ENDIAN__
12571 #define vld4q_p16(__p0) __extension__ ({ \
12572   poly16x8x4_t __ret; \
12573   __builtin_neon_vld4q_v(&__ret, __p0, 37); \
12574   __ret; \
12575 })
12576 #else
12577 #define vld4q_p16(__p0) __extension__ ({ \
12578   poly16x8x4_t __ret; \
12579   __builtin_neon_vld4q_v(&__ret, __p0, 37); \
12580  \
12581   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12582   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12583   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12584   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12585   __ret; \
12586 })
12587 #endif
12588 
12589 #ifdef __LITTLE_ENDIAN__
12590 #define vld4q_u8(__p0) __extension__ ({ \
12591   uint8x16x4_t __ret; \
12592   __builtin_neon_vld4q_v(&__ret, __p0, 48); \
12593   __ret; \
12594 })
12595 #else
12596 #define vld4q_u8(__p0) __extension__ ({ \
12597   uint8x16x4_t __ret; \
12598   __builtin_neon_vld4q_v(&__ret, __p0, 48); \
12599  \
12600   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12601   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12602   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12603   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12604   __ret; \
12605 })
12606 #endif
12607 
12608 #ifdef __LITTLE_ENDIAN__
12609 #define vld4q_u32(__p0) __extension__ ({ \
12610   uint32x4x4_t __ret; \
12611   __builtin_neon_vld4q_v(&__ret, __p0, 50); \
12612   __ret; \
12613 })
12614 #else
12615 #define vld4q_u32(__p0) __extension__ ({ \
12616   uint32x4x4_t __ret; \
12617   __builtin_neon_vld4q_v(&__ret, __p0, 50); \
12618  \
12619   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12620   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12621   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12622   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12623   __ret; \
12624 })
12625 #endif
12626 
12627 #ifdef __LITTLE_ENDIAN__
12628 #define vld4q_u16(__p0) __extension__ ({ \
12629   uint16x8x4_t __ret; \
12630   __builtin_neon_vld4q_v(&__ret, __p0, 49); \
12631   __ret; \
12632 })
12633 #else
12634 #define vld4q_u16(__p0) __extension__ ({ \
12635   uint16x8x4_t __ret; \
12636   __builtin_neon_vld4q_v(&__ret, __p0, 49); \
12637  \
12638   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12639   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12640   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12641   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12642   __ret; \
12643 })
12644 #endif
12645 
12646 #ifdef __LITTLE_ENDIAN__
12647 #define vld4q_s8(__p0) __extension__ ({ \
12648   int8x16x4_t __ret; \
12649   __builtin_neon_vld4q_v(&__ret, __p0, 32); \
12650   __ret; \
12651 })
12652 #else
12653 #define vld4q_s8(__p0) __extension__ ({ \
12654   int8x16x4_t __ret; \
12655   __builtin_neon_vld4q_v(&__ret, __p0, 32); \
12656  \
12657   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12658   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12659   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12660   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12661   __ret; \
12662 })
12663 #endif
12664 
12665 #ifdef __LITTLE_ENDIAN__
12666 #define vld4q_f32(__p0) __extension__ ({ \
12667   float32x4x4_t __ret; \
12668   __builtin_neon_vld4q_v(&__ret, __p0, 41); \
12669   __ret; \
12670 })
12671 #else
12672 #define vld4q_f32(__p0) __extension__ ({ \
12673   float32x4x4_t __ret; \
12674   __builtin_neon_vld4q_v(&__ret, __p0, 41); \
12675  \
12676   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12677   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12678   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12679   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12680   __ret; \
12681 })
12682 #endif
12683 
12684 #ifdef __LITTLE_ENDIAN__
12685 #define vld4q_s32(__p0) __extension__ ({ \
12686   int32x4x4_t __ret; \
12687   __builtin_neon_vld4q_v(&__ret, __p0, 34); \
12688   __ret; \
12689 })
12690 #else
12691 #define vld4q_s32(__p0) __extension__ ({ \
12692   int32x4x4_t __ret; \
12693   __builtin_neon_vld4q_v(&__ret, __p0, 34); \
12694  \
12695   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12696   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12697   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12698   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12699   __ret; \
12700 })
12701 #endif
12702 
12703 #ifdef __LITTLE_ENDIAN__
12704 #define vld4q_s16(__p0) __extension__ ({ \
12705   int16x8x4_t __ret; \
12706   __builtin_neon_vld4q_v(&__ret, __p0, 33); \
12707   __ret; \
12708 })
12709 #else
12710 #define vld4q_s16(__p0) __extension__ ({ \
12711   int16x8x4_t __ret; \
12712   __builtin_neon_vld4q_v(&__ret, __p0, 33); \
12713  \
12714   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12715   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12716   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12717   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12718   __ret; \
12719 })
12720 #endif
12721 
12722 #ifdef __LITTLE_ENDIAN__
12723 #define vld4_u8(__p0) __extension__ ({ \
12724   uint8x8x4_t __ret; \
12725   __builtin_neon_vld4_v(&__ret, __p0, 16); \
12726   __ret; \
12727 })
12728 #else
12729 #define vld4_u8(__p0) __extension__ ({ \
12730   uint8x8x4_t __ret; \
12731   __builtin_neon_vld4_v(&__ret, __p0, 16); \
12732  \
12733   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12734   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12735   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12736   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12737   __ret; \
12738 })
12739 #endif
12740 
12741 #ifdef __LITTLE_ENDIAN__
12742 #define vld4_u32(__p0) __extension__ ({ \
12743   uint32x2x4_t __ret; \
12744   __builtin_neon_vld4_v(&__ret, __p0, 18); \
12745   __ret; \
12746 })
12747 #else
12748 #define vld4_u32(__p0) __extension__ ({ \
12749   uint32x2x4_t __ret; \
12750   __builtin_neon_vld4_v(&__ret, __p0, 18); \
12751  \
12752   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12753   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12754   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12755   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
12756   __ret; \
12757 })
12758 #endif
12759 
12760 #define vld4_u64(__p0) __extension__ ({ \
12761   uint64x1x4_t __ret; \
12762   __builtin_neon_vld4_v(&__ret, __p0, 19); \
12763   __ret; \
12764 })
12765 #ifdef __LITTLE_ENDIAN__
12766 #define vld4_u16(__p0) __extension__ ({ \
12767   uint16x4x4_t __ret; \
12768   __builtin_neon_vld4_v(&__ret, __p0, 17); \
12769   __ret; \
12770 })
12771 #else
12772 #define vld4_u16(__p0) __extension__ ({ \
12773   uint16x4x4_t __ret; \
12774   __builtin_neon_vld4_v(&__ret, __p0, 17); \
12775  \
12776   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12777   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12778   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12779   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12780   __ret; \
12781 })
12782 #endif
12783 
12784 #ifdef __LITTLE_ENDIAN__
12785 #define vld4_s8(__p0) __extension__ ({ \
12786   int8x8x4_t __ret; \
12787   __builtin_neon_vld4_v(&__ret, __p0, 0); \
12788   __ret; \
12789 })
12790 #else
12791 #define vld4_s8(__p0) __extension__ ({ \
12792   int8x8x4_t __ret; \
12793   __builtin_neon_vld4_v(&__ret, __p0, 0); \
12794  \
12795   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12796   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12797   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12798   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12799   __ret; \
12800 })
12801 #endif
12802 
12803 #ifdef __LITTLE_ENDIAN__
12804 #define vld4_f32(__p0) __extension__ ({ \
12805   float32x2x4_t __ret; \
12806   __builtin_neon_vld4_v(&__ret, __p0, 9); \
12807   __ret; \
12808 })
12809 #else
12810 #define vld4_f32(__p0) __extension__ ({ \
12811   float32x2x4_t __ret; \
12812   __builtin_neon_vld4_v(&__ret, __p0, 9); \
12813  \
12814   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12815   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12816   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12817   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
12818   __ret; \
12819 })
12820 #endif
12821 
12822 #ifdef __LITTLE_ENDIAN__
12823 #define vld4_s32(__p0) __extension__ ({ \
12824   int32x2x4_t __ret; \
12825   __builtin_neon_vld4_v(&__ret, __p0, 2); \
12826   __ret; \
12827 })
12828 #else
12829 #define vld4_s32(__p0) __extension__ ({ \
12830   int32x2x4_t __ret; \
12831   __builtin_neon_vld4_v(&__ret, __p0, 2); \
12832  \
12833   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12834   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12835   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12836   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
12837   __ret; \
12838 })
12839 #endif
12840 
12841 #define vld4_s64(__p0) __extension__ ({ \
12842   int64x1x4_t __ret; \
12843   __builtin_neon_vld4_v(&__ret, __p0, 3); \
12844   __ret; \
12845 })
12846 #ifdef __LITTLE_ENDIAN__
12847 #define vld4_s16(__p0) __extension__ ({ \
12848   int16x4x4_t __ret; \
12849   __builtin_neon_vld4_v(&__ret, __p0, 1); \
12850   __ret; \
12851 })
12852 #else
12853 #define vld4_s16(__p0) __extension__ ({ \
12854   int16x4x4_t __ret; \
12855   __builtin_neon_vld4_v(&__ret, __p0, 1); \
12856  \
12857   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12858   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12859   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12860   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12861   __ret; \
12862 })
12863 #endif
12864 
12865 #ifdef __LITTLE_ENDIAN__
12866 #define vld4_dup_p8(__p0) __extension__ ({ \
12867   poly8x8x4_t __ret; \
12868   __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
12869   __ret; \
12870 })
12871 #else
12872 #define vld4_dup_p8(__p0) __extension__ ({ \
12873   poly8x8x4_t __ret; \
12874   __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
12875  \
12876   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12877   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12878   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12879   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12880   __ret; \
12881 })
12882 #endif
12883 
12884 #ifdef __LITTLE_ENDIAN__
12885 #define vld4_dup_p16(__p0) __extension__ ({ \
12886   poly16x4x4_t __ret; \
12887   __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
12888   __ret; \
12889 })
12890 #else
12891 #define vld4_dup_p16(__p0) __extension__ ({ \
12892   poly16x4x4_t __ret; \
12893   __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
12894  \
12895   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12896   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12897   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12898   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12899   __ret; \
12900 })
12901 #endif
12902 
12903 #ifdef __LITTLE_ENDIAN__
12904 #define vld4q_dup_p8(__p0) __extension__ ({ \
12905   poly8x16x4_t __ret; \
12906   __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
12907   __ret; \
12908 })
12909 #else
12910 #define vld4q_dup_p8(__p0) __extension__ ({ \
12911   poly8x16x4_t __ret; \
12912   __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
12913  \
12914   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12915   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12916   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12917   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12918   __ret; \
12919 })
12920 #endif
12921 
12922 #ifdef __LITTLE_ENDIAN__
12923 #define vld4q_dup_p16(__p0) __extension__ ({ \
12924   poly16x8x4_t __ret; \
12925   __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
12926   __ret; \
12927 })
12928 #else
12929 #define vld4q_dup_p16(__p0) __extension__ ({ \
12930   poly16x8x4_t __ret; \
12931   __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
12932  \
12933   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12934   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12935   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12936   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12937   __ret; \
12938 })
12939 #endif
12940 
12941 #ifdef __LITTLE_ENDIAN__
12942 #define vld4q_dup_u8(__p0) __extension__ ({ \
12943   uint8x16x4_t __ret; \
12944   __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
12945   __ret; \
12946 })
12947 #else
12948 #define vld4q_dup_u8(__p0) __extension__ ({ \
12949   uint8x16x4_t __ret; \
12950   __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
12951  \
12952   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12953   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12954   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12955   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12956   __ret; \
12957 })
12958 #endif
12959 
12960 #ifdef __LITTLE_ENDIAN__
12961 #define vld4q_dup_u32(__p0) __extension__ ({ \
12962   uint32x4x4_t __ret; \
12963   __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
12964   __ret; \
12965 })
12966 #else
12967 #define vld4q_dup_u32(__p0) __extension__ ({ \
12968   uint32x4x4_t __ret; \
12969   __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
12970  \
12971   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12972   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12973   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12974   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12975   __ret; \
12976 })
12977 #endif
12978 
12979 #ifdef __LITTLE_ENDIAN__
12980 #define vld4q_dup_u64(__p0) __extension__ ({ \
12981   uint64x2x4_t __ret; \
12982   __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
12983   __ret; \
12984 })
12985 #else
12986 #define vld4q_dup_u64(__p0) __extension__ ({ \
12987   uint64x2x4_t __ret; \
12988   __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
12989  \
12990   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12991   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12992   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12993   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
12994   __ret; \
12995 })
12996 #endif
12997 
12998 #ifdef __LITTLE_ENDIAN__
12999 #define vld4q_dup_u16(__p0) __extension__ ({ \
13000   uint16x8x4_t __ret; \
13001   __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
13002   __ret; \
13003 })
13004 #else
13005 #define vld4q_dup_u16(__p0) __extension__ ({ \
13006   uint16x8x4_t __ret; \
13007   __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
13008  \
13009   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13010   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13011   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13012   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13013   __ret; \
13014 })
13015 #endif
13016 
13017 #ifdef __LITTLE_ENDIAN__
13018 #define vld4q_dup_s8(__p0) __extension__ ({ \
13019   int8x16x4_t __ret; \
13020   __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
13021   __ret; \
13022 })
13023 #else
13024 #define vld4q_dup_s8(__p0) __extension__ ({ \
13025   int8x16x4_t __ret; \
13026   __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
13027  \
13028   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13029   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13030   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13031   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13032   __ret; \
13033 })
13034 #endif
13035 
13036 #ifdef __LITTLE_ENDIAN__
13037 #define vld4q_dup_f32(__p0) __extension__ ({ \
13038   float32x4x4_t __ret; \
13039   __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
13040   __ret; \
13041 })
13042 #else
13043 #define vld4q_dup_f32(__p0) __extension__ ({ \
13044   float32x4x4_t __ret; \
13045   __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
13046  \
13047   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13048   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13049   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13050   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13051   __ret; \
13052 })
13053 #endif
13054 
13055 #ifdef __LITTLE_ENDIAN__
13056 #define vld4q_dup_s32(__p0) __extension__ ({ \
13057   int32x4x4_t __ret; \
13058   __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
13059   __ret; \
13060 })
13061 #else
13062 #define vld4q_dup_s32(__p0) __extension__ ({ \
13063   int32x4x4_t __ret; \
13064   __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
13065  \
13066   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13067   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13068   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13069   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13070   __ret; \
13071 })
13072 #endif
13073 
13074 #ifdef __LITTLE_ENDIAN__
13075 #define vld4q_dup_s64(__p0) __extension__ ({ \
13076   int64x2x4_t __ret; \
13077   __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
13078   __ret; \
13079 })
13080 #else
13081 #define vld4q_dup_s64(__p0) __extension__ ({ \
13082   int64x2x4_t __ret; \
13083   __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
13084  \
13085   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13086   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13087   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13088   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13089   __ret; \
13090 })
13091 #endif
13092 
13093 #ifdef __LITTLE_ENDIAN__
13094 #define vld4q_dup_s16(__p0) __extension__ ({ \
13095   int16x8x4_t __ret; \
13096   __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
13097   __ret; \
13098 })
13099 #else
13100 #define vld4q_dup_s16(__p0) __extension__ ({ \
13101   int16x8x4_t __ret; \
13102   __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
13103  \
13104   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13105   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13106   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13107   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13108   __ret; \
13109 })
13110 #endif
13111 
13112 #ifdef __LITTLE_ENDIAN__
13113 #define vld4_dup_u8(__p0) __extension__ ({ \
13114   uint8x8x4_t __ret; \
13115   __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
13116   __ret; \
13117 })
13118 #else
13119 #define vld4_dup_u8(__p0) __extension__ ({ \
13120   uint8x8x4_t __ret; \
13121   __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
13122  \
13123   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13124   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13125   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13126   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13127   __ret; \
13128 })
13129 #endif
13130 
13131 #ifdef __LITTLE_ENDIAN__
13132 #define vld4_dup_u32(__p0) __extension__ ({ \
13133   uint32x2x4_t __ret; \
13134   __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
13135   __ret; \
13136 })
13137 #else
13138 #define vld4_dup_u32(__p0) __extension__ ({ \
13139   uint32x2x4_t __ret; \
13140   __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
13141  \
13142   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13143   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13144   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13145   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13146   __ret; \
13147 })
13148 #endif
13149 
13150 #define vld4_dup_u64(__p0) __extension__ ({ \
13151   uint64x1x4_t __ret; \
13152   __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
13153   __ret; \
13154 })
13155 #ifdef __LITTLE_ENDIAN__
13156 #define vld4_dup_u16(__p0) __extension__ ({ \
13157   uint16x4x4_t __ret; \
13158   __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
13159   __ret; \
13160 })
13161 #else
13162 #define vld4_dup_u16(__p0) __extension__ ({ \
13163   uint16x4x4_t __ret; \
13164   __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
13165  \
13166   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13167   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13168   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13169   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13170   __ret; \
13171 })
13172 #endif
13173 
13174 #ifdef __LITTLE_ENDIAN__
13175 #define vld4_dup_s8(__p0) __extension__ ({ \
13176   int8x8x4_t __ret; \
13177   __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
13178   __ret; \
13179 })
13180 #else
13181 #define vld4_dup_s8(__p0) __extension__ ({ \
13182   int8x8x4_t __ret; \
13183   __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
13184  \
13185   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13186   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13187   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13188   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13189   __ret; \
13190 })
13191 #endif
13192 
13193 #ifdef __LITTLE_ENDIAN__
13194 #define vld4_dup_f32(__p0) __extension__ ({ \
13195   float32x2x4_t __ret; \
13196   __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
13197   __ret; \
13198 })
13199 #else
13200 #define vld4_dup_f32(__p0) __extension__ ({ \
13201   float32x2x4_t __ret; \
13202   __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
13203  \
13204   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13205   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13206   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13207   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13208   __ret; \
13209 })
13210 #endif
13211 
13212 #ifdef __LITTLE_ENDIAN__
13213 #define vld4_dup_s32(__p0) __extension__ ({ \
13214   int32x2x4_t __ret; \
13215   __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
13216   __ret; \
13217 })
13218 #else
13219 #define vld4_dup_s32(__p0) __extension__ ({ \
13220   int32x2x4_t __ret; \
13221   __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
13222  \
13223   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13224   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13225   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13226   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13227   __ret; \
13228 })
13229 #endif
13230 
13231 #define vld4_dup_s64(__p0) __extension__ ({ \
13232   int64x1x4_t __ret; \
13233   __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
13234   __ret; \
13235 })
13236 #ifdef __LITTLE_ENDIAN__
13237 #define vld4_dup_s16(__p0) __extension__ ({ \
13238   int16x4x4_t __ret; \
13239   __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
13240   __ret; \
13241 })
13242 #else
13243 #define vld4_dup_s16(__p0) __extension__ ({ \
13244   int16x4x4_t __ret; \
13245   __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
13246  \
13247   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13248   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13249   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13250   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13251   __ret; \
13252 })
13253 #endif
13254 
13255 #ifdef __LITTLE_ENDIAN__
13256 #define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
13257   poly8x8x4_t __ret; \
13258   poly8x8x4_t __s1 = __p1; \
13259   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
13260   __ret; \
13261 })
13262 #else
13263 #define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
13264   poly8x8x4_t __ret; \
13265   poly8x8x4_t __s1 = __p1; \
13266   poly8x8x4_t __rev1; \
13267   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13268   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13269   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13270   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13271   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
13272  \
13273   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13274   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13275   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13276   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13277   __ret; \
13278 })
13279 #endif
13280 
13281 #ifdef __LITTLE_ENDIAN__
13282 #define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
13283   poly16x4x4_t __ret; \
13284   poly16x4x4_t __s1 = __p1; \
13285   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
13286   __ret; \
13287 })
13288 #else
13289 #define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
13290   poly16x4x4_t __ret; \
13291   poly16x4x4_t __s1 = __p1; \
13292   poly16x4x4_t __rev1; \
13293   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13294   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13295   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13296   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13297   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
13298  \
13299   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13300   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13301   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13302   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13303   __ret; \
13304 })
13305 #endif
13306 
13307 #ifdef __LITTLE_ENDIAN__
13308 #define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
13309   poly16x8x4_t __ret; \
13310   poly16x8x4_t __s1 = __p1; \
13311   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
13312   __ret; \
13313 })
13314 #else
13315 #define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
13316   poly16x8x4_t __ret; \
13317   poly16x8x4_t __s1 = __p1; \
13318   poly16x8x4_t __rev1; \
13319   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13320   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13321   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13322   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13323   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
13324  \
13325   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13326   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13327   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13328   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13329   __ret; \
13330 })
13331 #endif
13332 
13333 #ifdef __LITTLE_ENDIAN__
13334 #define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
13335   uint32x4x4_t __ret; \
13336   uint32x4x4_t __s1 = __p1; \
13337   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
13338   __ret; \
13339 })
13340 #else
13341 #define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
13342   uint32x4x4_t __ret; \
13343   uint32x4x4_t __s1 = __p1; \
13344   uint32x4x4_t __rev1; \
13345   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13346   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13347   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13348   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13349   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
13350  \
13351   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13352   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13353   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13354   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13355   __ret; \
13356 })
13357 #endif
13358 
13359 #ifdef __LITTLE_ENDIAN__
13360 #define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
13361   uint16x8x4_t __ret; \
13362   uint16x8x4_t __s1 = __p1; \
13363   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
13364   __ret; \
13365 })
13366 #else
13367 #define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
13368   uint16x8x4_t __ret; \
13369   uint16x8x4_t __s1 = __p1; \
13370   uint16x8x4_t __rev1; \
13371   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13372   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13373   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13374   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13375   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
13376  \
13377   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13378   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13379   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13380   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13381   __ret; \
13382 })
13383 #endif
13384 
13385 #ifdef __LITTLE_ENDIAN__
13386 #define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
13387   float32x4x4_t __ret; \
13388   float32x4x4_t __s1 = __p1; \
13389   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 41); \
13390   __ret; \
13391 })
13392 #else
13393 #define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
13394   float32x4x4_t __ret; \
13395   float32x4x4_t __s1 = __p1; \
13396   float32x4x4_t __rev1; \
13397   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13398   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13399   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13400   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13401   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 41); \
13402  \
13403   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13404   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13405   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13406   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13407   __ret; \
13408 })
13409 #endif
13410 
13411 #ifdef __LITTLE_ENDIAN__
13412 #define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
13413   int32x4x4_t __ret; \
13414   int32x4x4_t __s1 = __p1; \
13415   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 34); \
13416   __ret; \
13417 })
13418 #else
13419 #define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
13420   int32x4x4_t __ret; \
13421   int32x4x4_t __s1 = __p1; \
13422   int32x4x4_t __rev1; \
13423   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13424   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13425   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13426   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13427   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 34); \
13428  \
13429   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13430   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13431   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13432   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13433   __ret; \
13434 })
13435 #endif
13436 
13437 #ifdef __LITTLE_ENDIAN__
13438 #define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
13439   int16x8x4_t __ret; \
13440   int16x8x4_t __s1 = __p1; \
13441   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 33); \
13442   __ret; \
13443 })
13444 #else
13445 #define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
13446   int16x8x4_t __ret; \
13447   int16x8x4_t __s1 = __p1; \
13448   int16x8x4_t __rev1; \
13449   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13450   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13451   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13452   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13453   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 33); \
13454  \
13455   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13456   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13457   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13458   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13459   __ret; \
13460 })
13461 #endif
13462 
13463 #ifdef __LITTLE_ENDIAN__
13464 #define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
13465   uint8x8x4_t __ret; \
13466   uint8x8x4_t __s1 = __p1; \
13467   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
13468   __ret; \
13469 })
13470 #else
13471 #define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
13472   uint8x8x4_t __ret; \
13473   uint8x8x4_t __s1 = __p1; \
13474   uint8x8x4_t __rev1; \
13475   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13476   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13477   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13478   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13479   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
13480  \
13481   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13482   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13483   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13484   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13485   __ret; \
13486 })
13487 #endif
13488 
13489 #ifdef __LITTLE_ENDIAN__
13490 #define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
13491   uint32x2x4_t __ret; \
13492   uint32x2x4_t __s1 = __p1; \
13493   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
13494   __ret; \
13495 })
13496 #else
13497 #define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
13498   uint32x2x4_t __ret; \
13499   uint32x2x4_t __s1 = __p1; \
13500   uint32x2x4_t __rev1; \
13501   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
13502   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
13503   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
13504   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
13505   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
13506  \
13507   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13508   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13509   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13510   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13511   __ret; \
13512 })
13513 #endif
13514 
13515 #ifdef __LITTLE_ENDIAN__
13516 #define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
13517   uint16x4x4_t __ret; \
13518   uint16x4x4_t __s1 = __p1; \
13519   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
13520   __ret; \
13521 })
13522 #else
13523 #define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
13524   uint16x4x4_t __ret; \
13525   uint16x4x4_t __s1 = __p1; \
13526   uint16x4x4_t __rev1; \
13527   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13528   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13529   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13530   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13531   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
13532  \
13533   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13534   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13535   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13536   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13537   __ret; \
13538 })
13539 #endif
13540 
13541 #ifdef __LITTLE_ENDIAN__
13542 #define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
13543   int8x8x4_t __ret; \
13544   int8x8x4_t __s1 = __p1; \
13545   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
13546   __ret; \
13547 })
13548 #else
13549 #define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
13550   int8x8x4_t __ret; \
13551   int8x8x4_t __s1 = __p1; \
13552   int8x8x4_t __rev1; \
13553   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13554   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13555   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13556   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13557   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
13558  \
13559   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13560   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13561   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13562   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13563   __ret; \
13564 })
13565 #endif
13566 
13567 #ifdef __LITTLE_ENDIAN__
13568 #define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
13569   float32x2x4_t __ret; \
13570   float32x2x4_t __s1 = __p1; \
13571   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 9); \
13572   __ret; \
13573 })
13574 #else
13575 #define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
13576   float32x2x4_t __ret; \
13577   float32x2x4_t __s1 = __p1; \
13578   float32x2x4_t __rev1; \
13579   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
13580   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
13581   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
13582   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
13583   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 9); \
13584  \
13585   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13586   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13587   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13588   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13589   __ret; \
13590 })
13591 #endif
13592 
13593 #ifdef __LITTLE_ENDIAN__
13594 #define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
13595   int32x2x4_t __ret; \
13596   int32x2x4_t __s1 = __p1; \
13597   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 2); \
13598   __ret; \
13599 })
13600 #else
13601 #define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
13602   int32x2x4_t __ret; \
13603   int32x2x4_t __s1 = __p1; \
13604   int32x2x4_t __rev1; \
13605   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
13606   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
13607   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
13608   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
13609   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 2); \
13610  \
13611   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13612   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13613   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13614   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13615   __ret; \
13616 })
13617 #endif
13618 
13619 #ifdef __LITTLE_ENDIAN__
13620 #define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
13621   int16x4x4_t __ret; \
13622   int16x4x4_t __s1 = __p1; \
13623   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 1); \
13624   __ret; \
13625 })
13626 #else
13627 #define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
13628   int16x4x4_t __ret; \
13629   int16x4x4_t __s1 = __p1; \
13630   int16x4x4_t __rev1; \
13631   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13632   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13633   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13634   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13635   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 1); \
13636  \
13637   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13638   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13639   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13640   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13641   __ret; \
13642 })
13643 #endif
13644 
13645 #ifdef __LITTLE_ENDIAN__
vmaxq_u8(uint8x16_t __p0,uint8x16_t __p1)13646 __ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
13647   uint8x16_t __ret;
13648   __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
13649   return __ret;
13650 }
13651 #else
vmaxq_u8(uint8x16_t __p0,uint8x16_t __p1)13652 __ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
13653   uint8x16_t __ret;
13654   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13655   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13656   __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
13657   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13658   return __ret;
13659 }
13660 #endif
13661 
13662 #ifdef __LITTLE_ENDIAN__
vmaxq_u32(uint32x4_t __p0,uint32x4_t __p1)13663 __ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
13664   uint32x4_t __ret;
13665   __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
13666   return __ret;
13667 }
13668 #else
vmaxq_u32(uint32x4_t __p0,uint32x4_t __p1)13669 __ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
13670   uint32x4_t __ret;
13671   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13672   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13673   __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
13674   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13675   return __ret;
13676 }
13677 #endif
13678 
13679 #ifdef __LITTLE_ENDIAN__
vmaxq_u16(uint16x8_t __p0,uint16x8_t __p1)13680 __ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
13681   uint16x8_t __ret;
13682   __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
13683   return __ret;
13684 }
13685 #else
vmaxq_u16(uint16x8_t __p0,uint16x8_t __p1)13686 __ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
13687   uint16x8_t __ret;
13688   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13689   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13690   __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
13691   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13692   return __ret;
13693 }
13694 #endif
13695 
13696 #ifdef __LITTLE_ENDIAN__
vmaxq_s8(int8x16_t __p0,int8x16_t __p1)13697 __ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
13698   int8x16_t __ret;
13699   __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
13700   return __ret;
13701 }
13702 #else
vmaxq_s8(int8x16_t __p0,int8x16_t __p1)13703 __ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
13704   int8x16_t __ret;
13705   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13706   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13707   __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
13708   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13709   return __ret;
13710 }
13711 #endif
13712 
13713 #ifdef __LITTLE_ENDIAN__
vmaxq_f32(float32x4_t __p0,float32x4_t __p1)13714 __ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
13715   float32x4_t __ret;
13716   __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
13717   return __ret;
13718 }
13719 #else
vmaxq_f32(float32x4_t __p0,float32x4_t __p1)13720 __ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
13721   float32x4_t __ret;
13722   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13723   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13724   __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
13725   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13726   return __ret;
13727 }
13728 #endif
13729 
13730 #ifdef __LITTLE_ENDIAN__
vmaxq_s32(int32x4_t __p0,int32x4_t __p1)13731 __ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
13732   int32x4_t __ret;
13733   __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
13734   return __ret;
13735 }
13736 #else
vmaxq_s32(int32x4_t __p0,int32x4_t __p1)13737 __ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
13738   int32x4_t __ret;
13739   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13740   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13741   __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
13742   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13743   return __ret;
13744 }
13745 #endif
13746 
13747 #ifdef __LITTLE_ENDIAN__
vmaxq_s16(int16x8_t __p0,int16x8_t __p1)13748 __ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
13749   int16x8_t __ret;
13750   __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
13751   return __ret;
13752 }
13753 #else
vmaxq_s16(int16x8_t __p0,int16x8_t __p1)13754 __ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
13755   int16x8_t __ret;
13756   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13757   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13758   __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
13759   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13760   return __ret;
13761 }
13762 #endif
13763 
13764 #ifdef __LITTLE_ENDIAN__
vmax_u8(uint8x8_t __p0,uint8x8_t __p1)13765 __ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
13766   uint8x8_t __ret;
13767   __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
13768   return __ret;
13769 }
13770 #else
vmax_u8(uint8x8_t __p0,uint8x8_t __p1)13771 __ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
13772   uint8x8_t __ret;
13773   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13774   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13775   __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
13776   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13777   return __ret;
13778 }
13779 #endif
13780 
13781 #ifdef __LITTLE_ENDIAN__
vmax_u32(uint32x2_t __p0,uint32x2_t __p1)13782 __ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
13783   uint32x2_t __ret;
13784   __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
13785   return __ret;
13786 }
13787 #else
vmax_u32(uint32x2_t __p0,uint32x2_t __p1)13788 __ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
13789   uint32x2_t __ret;
13790   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13791   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13792   __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
13793   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13794   return __ret;
13795 }
13796 #endif
13797 
13798 #ifdef __LITTLE_ENDIAN__
vmax_u16(uint16x4_t __p0,uint16x4_t __p1)13799 __ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
13800   uint16x4_t __ret;
13801   __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
13802   return __ret;
13803 }
13804 #else
vmax_u16(uint16x4_t __p0,uint16x4_t __p1)13805 __ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
13806   uint16x4_t __ret;
13807   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13808   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13809   __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
13810   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13811   return __ret;
13812 }
13813 #endif
13814 
13815 #ifdef __LITTLE_ENDIAN__
vmax_s8(int8x8_t __p0,int8x8_t __p1)13816 __ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
13817   int8x8_t __ret;
13818   __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
13819   return __ret;
13820 }
13821 #else
vmax_s8(int8x8_t __p0,int8x8_t __p1)13822 __ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
13823   int8x8_t __ret;
13824   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13825   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13826   __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
13827   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13828   return __ret;
13829 }
13830 #endif
13831 
13832 #ifdef __LITTLE_ENDIAN__
vmax_f32(float32x2_t __p0,float32x2_t __p1)13833 __ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
13834   float32x2_t __ret;
13835   __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
13836   return __ret;
13837 }
13838 #else
vmax_f32(float32x2_t __p0,float32x2_t __p1)13839 __ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
13840   float32x2_t __ret;
13841   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13842   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13843   __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
13844   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13845   return __ret;
13846 }
13847 #endif
13848 
13849 #ifdef __LITTLE_ENDIAN__
vmax_s32(int32x2_t __p0,int32x2_t __p1)13850 __ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
13851   int32x2_t __ret;
13852   __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
13853   return __ret;
13854 }
13855 #else
vmax_s32(int32x2_t __p0,int32x2_t __p1)13856 __ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
13857   int32x2_t __ret;
13858   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13859   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13860   __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
13861   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13862   return __ret;
13863 }
13864 #endif
13865 
13866 #ifdef __LITTLE_ENDIAN__
vmax_s16(int16x4_t __p0,int16x4_t __p1)13867 __ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
13868   int16x4_t __ret;
13869   __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
13870   return __ret;
13871 }
13872 #else
vmax_s16(int16x4_t __p0,int16x4_t __p1)13873 __ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
13874   int16x4_t __ret;
13875   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13876   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13877   __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
13878   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13879   return __ret;
13880 }
13881 #endif
13882 
13883 #ifdef __LITTLE_ENDIAN__
vminq_u8(uint8x16_t __p0,uint8x16_t __p1)13884 __ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
13885   uint8x16_t __ret;
13886   __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
13887   return __ret;
13888 }
13889 #else
vminq_u8(uint8x16_t __p0,uint8x16_t __p1)13890 __ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
13891   uint8x16_t __ret;
13892   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13893   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13894   __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
13895   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13896   return __ret;
13897 }
13898 #endif
13899 
13900 #ifdef __LITTLE_ENDIAN__
vminq_u32(uint32x4_t __p0,uint32x4_t __p1)13901 __ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
13902   uint32x4_t __ret;
13903   __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
13904   return __ret;
13905 }
13906 #else
vminq_u32(uint32x4_t __p0,uint32x4_t __p1)13907 __ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
13908   uint32x4_t __ret;
13909   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13910   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13911   __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
13912   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13913   return __ret;
13914 }
13915 #endif
13916 
13917 #ifdef __LITTLE_ENDIAN__
vminq_u16(uint16x8_t __p0,uint16x8_t __p1)13918 __ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
13919   uint16x8_t __ret;
13920   __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
13921   return __ret;
13922 }
13923 #else
vminq_u16(uint16x8_t __p0,uint16x8_t __p1)13924 __ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
13925   uint16x8_t __ret;
13926   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13927   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13928   __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
13929   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13930   return __ret;
13931 }
13932 #endif
13933 
13934 #ifdef __LITTLE_ENDIAN__
vminq_s8(int8x16_t __p0,int8x16_t __p1)13935 __ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
13936   int8x16_t __ret;
13937   __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
13938   return __ret;
13939 }
13940 #else
vminq_s8(int8x16_t __p0,int8x16_t __p1)13941 __ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
13942   int8x16_t __ret;
13943   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13944   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13945   __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
13946   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13947   return __ret;
13948 }
13949 #endif
13950 
13951 #ifdef __LITTLE_ENDIAN__
vminq_f32(float32x4_t __p0,float32x4_t __p1)13952 __ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
13953   float32x4_t __ret;
13954   __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
13955   return __ret;
13956 }
13957 #else
vminq_f32(float32x4_t __p0,float32x4_t __p1)13958 __ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
13959   float32x4_t __ret;
13960   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13961   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13962   __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
13963   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13964   return __ret;
13965 }
13966 #endif
13967 
13968 #ifdef __LITTLE_ENDIAN__
vminq_s32(int32x4_t __p0,int32x4_t __p1)13969 __ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
13970   int32x4_t __ret;
13971   __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
13972   return __ret;
13973 }
13974 #else
vminq_s32(int32x4_t __p0,int32x4_t __p1)13975 __ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
13976   int32x4_t __ret;
13977   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13978   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13979   __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
13980   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13981   return __ret;
13982 }
13983 #endif
13984 
13985 #ifdef __LITTLE_ENDIAN__
vminq_s16(int16x8_t __p0,int16x8_t __p1)13986 __ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
13987   int16x8_t __ret;
13988   __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
13989   return __ret;
13990 }
13991 #else
vminq_s16(int16x8_t __p0,int16x8_t __p1)13992 __ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
13993   int16x8_t __ret;
13994   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13995   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13996   __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
13997   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13998   return __ret;
13999 }
14000 #endif
14001 
14002 #ifdef __LITTLE_ENDIAN__
vmin_u8(uint8x8_t __p0,uint8x8_t __p1)14003 __ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
14004   uint8x8_t __ret;
14005   __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
14006   return __ret;
14007 }
14008 #else
vmin_u8(uint8x8_t __p0,uint8x8_t __p1)14009 __ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
14010   uint8x8_t __ret;
14011   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14012   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14013   __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
14014   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14015   return __ret;
14016 }
14017 #endif
14018 
14019 #ifdef __LITTLE_ENDIAN__
vmin_u32(uint32x2_t __p0,uint32x2_t __p1)14020 __ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
14021   uint32x2_t __ret;
14022   __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
14023   return __ret;
14024 }
14025 #else
vmin_u32(uint32x2_t __p0,uint32x2_t __p1)14026 __ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
14027   uint32x2_t __ret;
14028   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14029   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14030   __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
14031   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14032   return __ret;
14033 }
14034 #endif
14035 
14036 #ifdef __LITTLE_ENDIAN__
vmin_u16(uint16x4_t __p0,uint16x4_t __p1)14037 __ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
14038   uint16x4_t __ret;
14039   __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
14040   return __ret;
14041 }
14042 #else
vmin_u16(uint16x4_t __p0,uint16x4_t __p1)14043 __ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
14044   uint16x4_t __ret;
14045   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14046   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14047   __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
14048   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14049   return __ret;
14050 }
14051 #endif
14052 
14053 #ifdef __LITTLE_ENDIAN__
vmin_s8(int8x8_t __p0,int8x8_t __p1)14054 __ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
14055   int8x8_t __ret;
14056   __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
14057   return __ret;
14058 }
14059 #else
vmin_s8(int8x8_t __p0,int8x8_t __p1)14060 __ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
14061   int8x8_t __ret;
14062   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14063   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14064   __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
14065   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14066   return __ret;
14067 }
14068 #endif
14069 
14070 #ifdef __LITTLE_ENDIAN__
vmin_f32(float32x2_t __p0,float32x2_t __p1)14071 __ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
14072   float32x2_t __ret;
14073   __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
14074   return __ret;
14075 }
14076 #else
vmin_f32(float32x2_t __p0,float32x2_t __p1)14077 __ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
14078   float32x2_t __ret;
14079   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14080   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14081   __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
14082   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14083   return __ret;
14084 }
14085 #endif
14086 
14087 #ifdef __LITTLE_ENDIAN__
vmin_s32(int32x2_t __p0,int32x2_t __p1)14088 __ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
14089   int32x2_t __ret;
14090   __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
14091   return __ret;
14092 }
14093 #else
vmin_s32(int32x2_t __p0,int32x2_t __p1)14094 __ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
14095   int32x2_t __ret;
14096   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14097   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14098   __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
14099   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14100   return __ret;
14101 }
14102 #endif
14103 
14104 #ifdef __LITTLE_ENDIAN__
vmin_s16(int16x4_t __p0,int16x4_t __p1)14105 __ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
14106   int16x4_t __ret;
14107   __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
14108   return __ret;
14109 }
14110 #else
vmin_s16(int16x4_t __p0,int16x4_t __p1)14111 __ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
14112   int16x4_t __ret;
14113   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14114   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14115   __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
14116   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14117   return __ret;
14118 }
14119 #endif
14120 
14121 #ifdef __LITTLE_ENDIAN__
vmlaq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)14122 __ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
14123   uint8x16_t __ret;
14124   __ret = __p0 + __p1 * __p2;
14125   return __ret;
14126 }
14127 #else
vmlaq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)14128 __ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
14129   uint8x16_t __ret;
14130   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14131   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14132   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14133   __ret = __rev0 + __rev1 * __rev2;
14134   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14135   return __ret;
14136 }
14137 #endif
14138 
14139 #ifdef __LITTLE_ENDIAN__
vmlaq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)14140 __ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
14141   uint32x4_t __ret;
14142   __ret = __p0 + __p1 * __p2;
14143   return __ret;
14144 }
14145 #else
vmlaq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)14146 __ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
14147   uint32x4_t __ret;
14148   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14149   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14150   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14151   __ret = __rev0 + __rev1 * __rev2;
14152   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14153   return __ret;
14154 }
14155 #endif
14156 
14157 #ifdef __LITTLE_ENDIAN__
vmlaq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)14158 __ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
14159   uint16x8_t __ret;
14160   __ret = __p0 + __p1 * __p2;
14161   return __ret;
14162 }
14163 #else
vmlaq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)14164 __ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
14165   uint16x8_t __ret;
14166   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14167   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14168   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14169   __ret = __rev0 + __rev1 * __rev2;
14170   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14171   return __ret;
14172 }
14173 #endif
14174 
14175 #ifdef __LITTLE_ENDIAN__
vmlaq_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)14176 __ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
14177   int8x16_t __ret;
14178   __ret = __p0 + __p1 * __p2;
14179   return __ret;
14180 }
14181 #else
vmlaq_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)14182 __ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
14183   int8x16_t __ret;
14184   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14185   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14186   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14187   __ret = __rev0 + __rev1 * __rev2;
14188   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14189   return __ret;
14190 }
14191 #endif
14192 
14193 #ifdef __LITTLE_ENDIAN__
vmlaq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)14194 __ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
14195   float32x4_t __ret;
14196   __ret = __p0 + __p1 * __p2;
14197   return __ret;
14198 }
14199 #else
vmlaq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)14200 __ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
14201   float32x4_t __ret;
14202   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14203   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14204   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14205   __ret = __rev0 + __rev1 * __rev2;
14206   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14207   return __ret;
14208 }
14209 #endif
14210 
14211 #ifdef __LITTLE_ENDIAN__
vmlaq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)14212 __ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
14213   int32x4_t __ret;
14214   __ret = __p0 + __p1 * __p2;
14215   return __ret;
14216 }
14217 #else
vmlaq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)14218 __ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
14219   int32x4_t __ret;
14220   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14221   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14222   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14223   __ret = __rev0 + __rev1 * __rev2;
14224   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14225   return __ret;
14226 }
14227 #endif
14228 
14229 #ifdef __LITTLE_ENDIAN__
vmlaq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)14230 __ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
14231   int16x8_t __ret;
14232   __ret = __p0 + __p1 * __p2;
14233   return __ret;
14234 }
14235 #else
vmlaq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)14236 __ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
14237   int16x8_t __ret;
14238   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14239   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14240   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14241   __ret = __rev0 + __rev1 * __rev2;
14242   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14243   return __ret;
14244 }
14245 #endif
14246 
14247 #ifdef __LITTLE_ENDIAN__
vmla_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)14248 __ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
14249   uint8x8_t __ret;
14250   __ret = __p0 + __p1 * __p2;
14251   return __ret;
14252 }
14253 #else
vmla_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)14254 __ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
14255   uint8x8_t __ret;
14256   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14257   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14258   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14259   __ret = __rev0 + __rev1 * __rev2;
14260   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14261   return __ret;
14262 }
14263 #endif
14264 
14265 #ifdef __LITTLE_ENDIAN__
vmla_u32(uint32x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)14266 __ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
14267   uint32x2_t __ret;
14268   __ret = __p0 + __p1 * __p2;
14269   return __ret;
14270 }
14271 #else
vmla_u32(uint32x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)14272 __ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
14273   uint32x2_t __ret;
14274   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14275   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14276   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
14277   __ret = __rev0 + __rev1 * __rev2;
14278   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14279   return __ret;
14280 }
14281 #endif
14282 
14283 #ifdef __LITTLE_ENDIAN__
vmla_u16(uint16x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)14284 __ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
14285   uint16x4_t __ret;
14286   __ret = __p0 + __p1 * __p2;
14287   return __ret;
14288 }
14289 #else
vmla_u16(uint16x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)14290 __ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
14291   uint16x4_t __ret;
14292   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14293   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14294   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14295   __ret = __rev0 + __rev1 * __rev2;
14296   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14297   return __ret;
14298 }
14299 #endif
14300 
14301 #ifdef __LITTLE_ENDIAN__
vmla_s8(int8x8_t __p0,int8x8_t __p1,int8x8_t __p2)14302 __ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
14303   int8x8_t __ret;
14304   __ret = __p0 + __p1 * __p2;
14305   return __ret;
14306 }
14307 #else
vmla_s8(int8x8_t __p0,int8x8_t __p1,int8x8_t __p2)14308 __ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
14309   int8x8_t __ret;
14310   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14311   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14312   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14313   __ret = __rev0 + __rev1 * __rev2;
14314   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14315   return __ret;
14316 }
14317 #endif
14318 
14319 #ifdef __LITTLE_ENDIAN__
vmla_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)14320 __ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
14321   float32x2_t __ret;
14322   __ret = __p0 + __p1 * __p2;
14323   return __ret;
14324 }
14325 #else
vmla_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)14326 __ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
14327   float32x2_t __ret;
14328   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14329   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14330   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
14331   __ret = __rev0 + __rev1 * __rev2;
14332   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14333   return __ret;
14334 }
14335 #endif
14336 
14337 #ifdef __LITTLE_ENDIAN__
vmla_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)14338 __ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
14339   int32x2_t __ret;
14340   __ret = __p0 + __p1 * __p2;
14341   return __ret;
14342 }
14343 #else
vmla_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)14344 __ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
14345   int32x2_t __ret;
14346   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14347   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14348   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
14349   __ret = __rev0 + __rev1 * __rev2;
14350   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14351   return __ret;
14352 }
14353 #endif
14354 
14355 #ifdef __LITTLE_ENDIAN__
vmla_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)14356 __ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
14357   int16x4_t __ret;
14358   __ret = __p0 + __p1 * __p2;
14359   return __ret;
14360 }
14361 #else
vmla_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)14362 __ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
14363   int16x4_t __ret;
14364   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14365   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14366   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14367   __ret = __rev0 + __rev1 * __rev2;
14368   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14369   return __ret;
14370 }
14371 #endif
14372 
14373 #ifdef __LITTLE_ENDIAN__
14374 #define vmlaq_lane_u32(__p0_46, __p1_46, __p2_46, __p3_46) __extension__ ({ \
14375   uint32x4_t __ret_46; \
14376   uint32x4_t __s0_46 = __p0_46; \
14377   uint32x4_t __s1_46 = __p1_46; \
14378   uint32x2_t __s2_46 = __p2_46; \
14379   __ret_46 = __s0_46 + __s1_46 * splatq_lane_u32(__s2_46, __p3_46); \
14380   __ret_46; \
14381 })
14382 #else
14383 #define vmlaq_lane_u32(__p0_47, __p1_47, __p2_47, __p3_47) __extension__ ({ \
14384   uint32x4_t __ret_47; \
14385   uint32x4_t __s0_47 = __p0_47; \
14386   uint32x4_t __s1_47 = __p1_47; \
14387   uint32x2_t __s2_47 = __p2_47; \
14388   uint32x4_t __rev0_47;  __rev0_47 = __builtin_shufflevector(__s0_47, __s0_47, 3, 2, 1, 0); \
14389   uint32x4_t __rev1_47;  __rev1_47 = __builtin_shufflevector(__s1_47, __s1_47, 3, 2, 1, 0); \
14390   uint32x2_t __rev2_47;  __rev2_47 = __builtin_shufflevector(__s2_47, __s2_47, 1, 0); \
14391   __ret_47 = __rev0_47 + __rev1_47 * __noswap_splatq_lane_u32(__rev2_47, __p3_47); \
14392   __ret_47 = __builtin_shufflevector(__ret_47, __ret_47, 3, 2, 1, 0); \
14393   __ret_47; \
14394 })
14395 #endif
14396 
14397 #ifdef __LITTLE_ENDIAN__
14398 #define vmlaq_lane_u16(__p0_48, __p1_48, __p2_48, __p3_48) __extension__ ({ \
14399   uint16x8_t __ret_48; \
14400   uint16x8_t __s0_48 = __p0_48; \
14401   uint16x8_t __s1_48 = __p1_48; \
14402   uint16x4_t __s2_48 = __p2_48; \
14403   __ret_48 = __s0_48 + __s1_48 * splatq_lane_u16(__s2_48, __p3_48); \
14404   __ret_48; \
14405 })
14406 #else
14407 #define vmlaq_lane_u16(__p0_49, __p1_49, __p2_49, __p3_49) __extension__ ({ \
14408   uint16x8_t __ret_49; \
14409   uint16x8_t __s0_49 = __p0_49; \
14410   uint16x8_t __s1_49 = __p1_49; \
14411   uint16x4_t __s2_49 = __p2_49; \
14412   uint16x8_t __rev0_49;  __rev0_49 = __builtin_shufflevector(__s0_49, __s0_49, 7, 6, 5, 4, 3, 2, 1, 0); \
14413   uint16x8_t __rev1_49;  __rev1_49 = __builtin_shufflevector(__s1_49, __s1_49, 7, 6, 5, 4, 3, 2, 1, 0); \
14414   uint16x4_t __rev2_49;  __rev2_49 = __builtin_shufflevector(__s2_49, __s2_49, 3, 2, 1, 0); \
14415   __ret_49 = __rev0_49 + __rev1_49 * __noswap_splatq_lane_u16(__rev2_49, __p3_49); \
14416   __ret_49 = __builtin_shufflevector(__ret_49, __ret_49, 7, 6, 5, 4, 3, 2, 1, 0); \
14417   __ret_49; \
14418 })
14419 #endif
14420 
14421 #ifdef __LITTLE_ENDIAN__
14422 #define vmlaq_lane_f32(__p0_50, __p1_50, __p2_50, __p3_50) __extension__ ({ \
14423   float32x4_t __ret_50; \
14424   float32x4_t __s0_50 = __p0_50; \
14425   float32x4_t __s1_50 = __p1_50; \
14426   float32x2_t __s2_50 = __p2_50; \
14427   __ret_50 = __s0_50 + __s1_50 * splatq_lane_f32(__s2_50, __p3_50); \
14428   __ret_50; \
14429 })
14430 #else
14431 #define vmlaq_lane_f32(__p0_51, __p1_51, __p2_51, __p3_51) __extension__ ({ \
14432   float32x4_t __ret_51; \
14433   float32x4_t __s0_51 = __p0_51; \
14434   float32x4_t __s1_51 = __p1_51; \
14435   float32x2_t __s2_51 = __p2_51; \
14436   float32x4_t __rev0_51;  __rev0_51 = __builtin_shufflevector(__s0_51, __s0_51, 3, 2, 1, 0); \
14437   float32x4_t __rev1_51;  __rev1_51 = __builtin_shufflevector(__s1_51, __s1_51, 3, 2, 1, 0); \
14438   float32x2_t __rev2_51;  __rev2_51 = __builtin_shufflevector(__s2_51, __s2_51, 1, 0); \
14439   __ret_51 = __rev0_51 + __rev1_51 * __noswap_splatq_lane_f32(__rev2_51, __p3_51); \
14440   __ret_51 = __builtin_shufflevector(__ret_51, __ret_51, 3, 2, 1, 0); \
14441   __ret_51; \
14442 })
14443 #endif
14444 
14445 #ifdef __LITTLE_ENDIAN__
14446 #define vmlaq_lane_s32(__p0_52, __p1_52, __p2_52, __p3_52) __extension__ ({ \
14447   int32x4_t __ret_52; \
14448   int32x4_t __s0_52 = __p0_52; \
14449   int32x4_t __s1_52 = __p1_52; \
14450   int32x2_t __s2_52 = __p2_52; \
14451   __ret_52 = __s0_52 + __s1_52 * splatq_lane_s32(__s2_52, __p3_52); \
14452   __ret_52; \
14453 })
14454 #else
14455 #define vmlaq_lane_s32(__p0_53, __p1_53, __p2_53, __p3_53) __extension__ ({ \
14456   int32x4_t __ret_53; \
14457   int32x4_t __s0_53 = __p0_53; \
14458   int32x4_t __s1_53 = __p1_53; \
14459   int32x2_t __s2_53 = __p2_53; \
14460   int32x4_t __rev0_53;  __rev0_53 = __builtin_shufflevector(__s0_53, __s0_53, 3, 2, 1, 0); \
14461   int32x4_t __rev1_53;  __rev1_53 = __builtin_shufflevector(__s1_53, __s1_53, 3, 2, 1, 0); \
14462   int32x2_t __rev2_53;  __rev2_53 = __builtin_shufflevector(__s2_53, __s2_53, 1, 0); \
14463   __ret_53 = __rev0_53 + __rev1_53 * __noswap_splatq_lane_s32(__rev2_53, __p3_53); \
14464   __ret_53 = __builtin_shufflevector(__ret_53, __ret_53, 3, 2, 1, 0); \
14465   __ret_53; \
14466 })
14467 #endif
14468 
14469 #ifdef __LITTLE_ENDIAN__
14470 #define vmlaq_lane_s16(__p0_54, __p1_54, __p2_54, __p3_54) __extension__ ({ \
14471   int16x8_t __ret_54; \
14472   int16x8_t __s0_54 = __p0_54; \
14473   int16x8_t __s1_54 = __p1_54; \
14474   int16x4_t __s2_54 = __p2_54; \
14475   __ret_54 = __s0_54 + __s1_54 * splatq_lane_s16(__s2_54, __p3_54); \
14476   __ret_54; \
14477 })
14478 #else
14479 #define vmlaq_lane_s16(__p0_55, __p1_55, __p2_55, __p3_55) __extension__ ({ \
14480   int16x8_t __ret_55; \
14481   int16x8_t __s0_55 = __p0_55; \
14482   int16x8_t __s1_55 = __p1_55; \
14483   int16x4_t __s2_55 = __p2_55; \
14484   int16x8_t __rev0_55;  __rev0_55 = __builtin_shufflevector(__s0_55, __s0_55, 7, 6, 5, 4, 3, 2, 1, 0); \
14485   int16x8_t __rev1_55;  __rev1_55 = __builtin_shufflevector(__s1_55, __s1_55, 7, 6, 5, 4, 3, 2, 1, 0); \
14486   int16x4_t __rev2_55;  __rev2_55 = __builtin_shufflevector(__s2_55, __s2_55, 3, 2, 1, 0); \
14487   __ret_55 = __rev0_55 + __rev1_55 * __noswap_splatq_lane_s16(__rev2_55, __p3_55); \
14488   __ret_55 = __builtin_shufflevector(__ret_55, __ret_55, 7, 6, 5, 4, 3, 2, 1, 0); \
14489   __ret_55; \
14490 })
14491 #endif
14492 
14493 #ifdef __LITTLE_ENDIAN__
14494 #define vmla_lane_u32(__p0_56, __p1_56, __p2_56, __p3_56) __extension__ ({ \
14495   uint32x2_t __ret_56; \
14496   uint32x2_t __s0_56 = __p0_56; \
14497   uint32x2_t __s1_56 = __p1_56; \
14498   uint32x2_t __s2_56 = __p2_56; \
14499   __ret_56 = __s0_56 + __s1_56 * splat_lane_u32(__s2_56, __p3_56); \
14500   __ret_56; \
14501 })
14502 #else
14503 #define vmla_lane_u32(__p0_57, __p1_57, __p2_57, __p3_57) __extension__ ({ \
14504   uint32x2_t __ret_57; \
14505   uint32x2_t __s0_57 = __p0_57; \
14506   uint32x2_t __s1_57 = __p1_57; \
14507   uint32x2_t __s2_57 = __p2_57; \
14508   uint32x2_t __rev0_57;  __rev0_57 = __builtin_shufflevector(__s0_57, __s0_57, 1, 0); \
14509   uint32x2_t __rev1_57;  __rev1_57 = __builtin_shufflevector(__s1_57, __s1_57, 1, 0); \
14510   uint32x2_t __rev2_57;  __rev2_57 = __builtin_shufflevector(__s2_57, __s2_57, 1, 0); \
14511   __ret_57 = __rev0_57 + __rev1_57 * __noswap_splat_lane_u32(__rev2_57, __p3_57); \
14512   __ret_57 = __builtin_shufflevector(__ret_57, __ret_57, 1, 0); \
14513   __ret_57; \
14514 })
14515 #endif
14516 
14517 #ifdef __LITTLE_ENDIAN__
14518 #define vmla_lane_u16(__p0_58, __p1_58, __p2_58, __p3_58) __extension__ ({ \
14519   uint16x4_t __ret_58; \
14520   uint16x4_t __s0_58 = __p0_58; \
14521   uint16x4_t __s1_58 = __p1_58; \
14522   uint16x4_t __s2_58 = __p2_58; \
14523   __ret_58 = __s0_58 + __s1_58 * splat_lane_u16(__s2_58, __p3_58); \
14524   __ret_58; \
14525 })
14526 #else
14527 #define vmla_lane_u16(__p0_59, __p1_59, __p2_59, __p3_59) __extension__ ({ \
14528   uint16x4_t __ret_59; \
14529   uint16x4_t __s0_59 = __p0_59; \
14530   uint16x4_t __s1_59 = __p1_59; \
14531   uint16x4_t __s2_59 = __p2_59; \
14532   uint16x4_t __rev0_59;  __rev0_59 = __builtin_shufflevector(__s0_59, __s0_59, 3, 2, 1, 0); \
14533   uint16x4_t __rev1_59;  __rev1_59 = __builtin_shufflevector(__s1_59, __s1_59, 3, 2, 1, 0); \
14534   uint16x4_t __rev2_59;  __rev2_59 = __builtin_shufflevector(__s2_59, __s2_59, 3, 2, 1, 0); \
14535   __ret_59 = __rev0_59 + __rev1_59 * __noswap_splat_lane_u16(__rev2_59, __p3_59); \
14536   __ret_59 = __builtin_shufflevector(__ret_59, __ret_59, 3, 2, 1, 0); \
14537   __ret_59; \
14538 })
14539 #endif
14540 
14541 #ifdef __LITTLE_ENDIAN__
14542 #define vmla_lane_f32(__p0_60, __p1_60, __p2_60, __p3_60) __extension__ ({ \
14543   float32x2_t __ret_60; \
14544   float32x2_t __s0_60 = __p0_60; \
14545   float32x2_t __s1_60 = __p1_60; \
14546   float32x2_t __s2_60 = __p2_60; \
14547   __ret_60 = __s0_60 + __s1_60 * splat_lane_f32(__s2_60, __p3_60); \
14548   __ret_60; \
14549 })
14550 #else
14551 #define vmla_lane_f32(__p0_61, __p1_61, __p2_61, __p3_61) __extension__ ({ \
14552   float32x2_t __ret_61; \
14553   float32x2_t __s0_61 = __p0_61; \
14554   float32x2_t __s1_61 = __p1_61; \
14555   float32x2_t __s2_61 = __p2_61; \
14556   float32x2_t __rev0_61;  __rev0_61 = __builtin_shufflevector(__s0_61, __s0_61, 1, 0); \
14557   float32x2_t __rev1_61;  __rev1_61 = __builtin_shufflevector(__s1_61, __s1_61, 1, 0); \
14558   float32x2_t __rev2_61;  __rev2_61 = __builtin_shufflevector(__s2_61, __s2_61, 1, 0); \
14559   __ret_61 = __rev0_61 + __rev1_61 * __noswap_splat_lane_f32(__rev2_61, __p3_61); \
14560   __ret_61 = __builtin_shufflevector(__ret_61, __ret_61, 1, 0); \
14561   __ret_61; \
14562 })
14563 #endif
14564 
14565 #ifdef __LITTLE_ENDIAN__
14566 #define vmla_lane_s32(__p0_62, __p1_62, __p2_62, __p3_62) __extension__ ({ \
14567   int32x2_t __ret_62; \
14568   int32x2_t __s0_62 = __p0_62; \
14569   int32x2_t __s1_62 = __p1_62; \
14570   int32x2_t __s2_62 = __p2_62; \
14571   __ret_62 = __s0_62 + __s1_62 * splat_lane_s32(__s2_62, __p3_62); \
14572   __ret_62; \
14573 })
14574 #else
14575 #define vmla_lane_s32(__p0_63, __p1_63, __p2_63, __p3_63) __extension__ ({ \
14576   int32x2_t __ret_63; \
14577   int32x2_t __s0_63 = __p0_63; \
14578   int32x2_t __s1_63 = __p1_63; \
14579   int32x2_t __s2_63 = __p2_63; \
14580   int32x2_t __rev0_63;  __rev0_63 = __builtin_shufflevector(__s0_63, __s0_63, 1, 0); \
14581   int32x2_t __rev1_63;  __rev1_63 = __builtin_shufflevector(__s1_63, __s1_63, 1, 0); \
14582   int32x2_t __rev2_63;  __rev2_63 = __builtin_shufflevector(__s2_63, __s2_63, 1, 0); \
14583   __ret_63 = __rev0_63 + __rev1_63 * __noswap_splat_lane_s32(__rev2_63, __p3_63); \
14584   __ret_63 = __builtin_shufflevector(__ret_63, __ret_63, 1, 0); \
14585   __ret_63; \
14586 })
14587 #endif
14588 
14589 #ifdef __LITTLE_ENDIAN__
14590 #define vmla_lane_s16(__p0_64, __p1_64, __p2_64, __p3_64) __extension__ ({ \
14591   int16x4_t __ret_64; \
14592   int16x4_t __s0_64 = __p0_64; \
14593   int16x4_t __s1_64 = __p1_64; \
14594   int16x4_t __s2_64 = __p2_64; \
14595   __ret_64 = __s0_64 + __s1_64 * splat_lane_s16(__s2_64, __p3_64); \
14596   __ret_64; \
14597 })
14598 #else
14599 #define vmla_lane_s16(__p0_65, __p1_65, __p2_65, __p3_65) __extension__ ({ \
14600   int16x4_t __ret_65; \
14601   int16x4_t __s0_65 = __p0_65; \
14602   int16x4_t __s1_65 = __p1_65; \
14603   int16x4_t __s2_65 = __p2_65; \
14604   int16x4_t __rev0_65;  __rev0_65 = __builtin_shufflevector(__s0_65, __s0_65, 3, 2, 1, 0); \
14605   int16x4_t __rev1_65;  __rev1_65 = __builtin_shufflevector(__s1_65, __s1_65, 3, 2, 1, 0); \
14606   int16x4_t __rev2_65;  __rev2_65 = __builtin_shufflevector(__s2_65, __s2_65, 3, 2, 1, 0); \
14607   __ret_65 = __rev0_65 + __rev1_65 * __noswap_splat_lane_s16(__rev2_65, __p3_65); \
14608   __ret_65 = __builtin_shufflevector(__ret_65, __ret_65, 3, 2, 1, 0); \
14609   __ret_65; \
14610 })
14611 #endif
14612 
14613 #ifdef __LITTLE_ENDIAN__
vmlaq_n_u32(uint32x4_t __p0,uint32x4_t __p1,uint32_t __p2)14614 __ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
14615   uint32x4_t __ret;
14616   __ret = __p0 + __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
14617   return __ret;
14618 }
14619 #else
vmlaq_n_u32(uint32x4_t __p0,uint32x4_t __p1,uint32_t __p2)14620 __ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
14621   uint32x4_t __ret;
14622   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14623   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14624   __ret = __rev0 + __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
14625   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14626   return __ret;
14627 }
14628 #endif
14629 
14630 #ifdef __LITTLE_ENDIAN__
vmlaq_n_u16(uint16x8_t __p0,uint16x8_t __p1,uint16_t __p2)14631 __ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
14632   uint16x8_t __ret;
14633   __ret = __p0 + __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
14634   return __ret;
14635 }
14636 #else
vmlaq_n_u16(uint16x8_t __p0,uint16x8_t __p1,uint16_t __p2)14637 __ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
14638   uint16x8_t __ret;
14639   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14640   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14641   __ret = __rev0 + __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
14642   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14643   return __ret;
14644 }
14645 #endif
14646 
14647 #ifdef __LITTLE_ENDIAN__
vmlaq_n_f32(float32x4_t __p0,float32x4_t __p1,float32_t __p2)14648 __ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
14649   float32x4_t __ret;
14650   __ret = __p0 + __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
14651   return __ret;
14652 }
14653 #else
vmlaq_n_f32(float32x4_t __p0,float32x4_t __p1,float32_t __p2)14654 __ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
14655   float32x4_t __ret;
14656   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14657   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14658   __ret = __rev0 + __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
14659   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14660   return __ret;
14661 }
14662 #endif
14663 
14664 #ifdef __LITTLE_ENDIAN__
vmlaq_n_s32(int32x4_t __p0,int32x4_t __p1,int32_t __p2)14665 __ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
14666   int32x4_t __ret;
14667   __ret = __p0 + __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
14668   return __ret;
14669 }
14670 #else
vmlaq_n_s32(int32x4_t __p0,int32x4_t __p1,int32_t __p2)14671 __ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
14672   int32x4_t __ret;
14673   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14674   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14675   __ret = __rev0 + __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
14676   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14677   return __ret;
14678 }
14679 #endif
14680 
14681 #ifdef __LITTLE_ENDIAN__
vmlaq_n_s16(int16x8_t __p0,int16x8_t __p1,int16_t __p2)14682 __ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
14683   int16x8_t __ret;
14684   __ret = __p0 + __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
14685   return __ret;
14686 }
14687 #else
vmlaq_n_s16(int16x8_t __p0,int16x8_t __p1,int16_t __p2)14688 __ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
14689   int16x8_t __ret;
14690   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14691   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14692   __ret = __rev0 + __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
14693   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14694   return __ret;
14695 }
14696 #endif
14697 
14698 #ifdef __LITTLE_ENDIAN__
vmla_n_u32(uint32x2_t __p0,uint32x2_t __p1,uint32_t __p2)14699 __ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
14700   uint32x2_t __ret;
14701   __ret = __p0 + __p1 * (uint32x2_t) {__p2, __p2};
14702   return __ret;
14703 }
14704 #else
vmla_n_u32(uint32x2_t __p0,uint32x2_t __p1,uint32_t __p2)14705 __ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
14706   uint32x2_t __ret;
14707   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14708   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14709   __ret = __rev0 + __rev1 * (uint32x2_t) {__p2, __p2};
14710   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14711   return __ret;
14712 }
14713 #endif
14714 
14715 #ifdef __LITTLE_ENDIAN__
vmla_n_u16(uint16x4_t __p0,uint16x4_t __p1,uint16_t __p2)14716 __ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
14717   uint16x4_t __ret;
14718   __ret = __p0 + __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
14719   return __ret;
14720 }
14721 #else
vmla_n_u16(uint16x4_t __p0,uint16x4_t __p1,uint16_t __p2)14722 __ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
14723   uint16x4_t __ret;
14724   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14725   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14726   __ret = __rev0 + __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
14727   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14728   return __ret;
14729 }
14730 #endif
14731 
14732 #ifdef __LITTLE_ENDIAN__
vmla_n_f32(float32x2_t __p0,float32x2_t __p1,float32_t __p2)14733 __ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
14734   float32x2_t __ret;
14735   __ret = __p0 + __p1 * (float32x2_t) {__p2, __p2};
14736   return __ret;
14737 }
14738 #else
vmla_n_f32(float32x2_t __p0,float32x2_t __p1,float32_t __p2)14739 __ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
14740   float32x2_t __ret;
14741   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14742   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14743   __ret = __rev0 + __rev1 * (float32x2_t) {__p2, __p2};
14744   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14745   return __ret;
14746 }
14747 #endif
14748 
14749 #ifdef __LITTLE_ENDIAN__
vmla_n_s32(int32x2_t __p0,int32x2_t __p1,int32_t __p2)14750 __ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
14751   int32x2_t __ret;
14752   __ret = __p0 + __p1 * (int32x2_t) {__p2, __p2};
14753   return __ret;
14754 }
14755 #else
vmla_n_s32(int32x2_t __p0,int32x2_t __p1,int32_t __p2)14756 __ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
14757   int32x2_t __ret;
14758   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14759   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14760   __ret = __rev0 + __rev1 * (int32x2_t) {__p2, __p2};
14761   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14762   return __ret;
14763 }
14764 #endif
14765 
14766 #ifdef __LITTLE_ENDIAN__
vmla_n_s16(int16x4_t __p0,int16x4_t __p1,int16_t __p2)14767 __ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
14768   int16x4_t __ret;
14769   __ret = __p0 + __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
14770   return __ret;
14771 }
14772 #else
vmla_n_s16(int16x4_t __p0,int16x4_t __p1,int16_t __p2)14773 __ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
14774   int16x4_t __ret;
14775   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14776   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14777   __ret = __rev0 + __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
14778   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14779   return __ret;
14780 }
14781 #endif
14782 
14783 #ifdef __LITTLE_ENDIAN__
vmlsq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)14784 __ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
14785   uint8x16_t __ret;
14786   __ret = __p0 - __p1 * __p2;
14787   return __ret;
14788 }
14789 #else
vmlsq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)14790 __ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
14791   uint8x16_t __ret;
14792   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14793   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14794   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14795   __ret = __rev0 - __rev1 * __rev2;
14796   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14797   return __ret;
14798 }
14799 #endif
14800 
14801 #ifdef __LITTLE_ENDIAN__
vmlsq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)14802 __ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
14803   uint32x4_t __ret;
14804   __ret = __p0 - __p1 * __p2;
14805   return __ret;
14806 }
14807 #else
vmlsq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)14808 __ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
14809   uint32x4_t __ret;
14810   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14811   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14812   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14813   __ret = __rev0 - __rev1 * __rev2;
14814   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14815   return __ret;
14816 }
14817 #endif
14818 
14819 #ifdef __LITTLE_ENDIAN__
vmlsq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)14820 __ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
14821   uint16x8_t __ret;
14822   __ret = __p0 - __p1 * __p2;
14823   return __ret;
14824 }
14825 #else
vmlsq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)14826 __ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
14827   uint16x8_t __ret;
14828   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14829   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14830   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14831   __ret = __rev0 - __rev1 * __rev2;
14832   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14833   return __ret;
14834 }
14835 #endif
14836 
14837 #ifdef __LITTLE_ENDIAN__
vmlsq_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)14838 __ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
14839   int8x16_t __ret;
14840   __ret = __p0 - __p1 * __p2;
14841   return __ret;
14842 }
14843 #else
vmlsq_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)14844 __ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
14845   int8x16_t __ret;
14846   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14847   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14848   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14849   __ret = __rev0 - __rev1 * __rev2;
14850   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14851   return __ret;
14852 }
14853 #endif
14854 
14855 #ifdef __LITTLE_ENDIAN__
vmlsq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)14856 __ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
14857   float32x4_t __ret;
14858   __ret = __p0 - __p1 * __p2;
14859   return __ret;
14860 }
14861 #else
vmlsq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)14862 __ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
14863   float32x4_t __ret;
14864   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14865   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14866   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14867   __ret = __rev0 - __rev1 * __rev2;
14868   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14869   return __ret;
14870 }
14871 #endif
14872 
14873 #ifdef __LITTLE_ENDIAN__
vmlsq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)14874 __ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
14875   int32x4_t __ret;
14876   __ret = __p0 - __p1 * __p2;
14877   return __ret;
14878 }
14879 #else
vmlsq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)14880 __ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
14881   int32x4_t __ret;
14882   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14883   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14884   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14885   __ret = __rev0 - __rev1 * __rev2;
14886   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14887   return __ret;
14888 }
14889 #endif
14890 
14891 #ifdef __LITTLE_ENDIAN__
vmlsq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)14892 __ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
14893   int16x8_t __ret;
14894   __ret = __p0 - __p1 * __p2;
14895   return __ret;
14896 }
14897 #else
vmlsq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)14898 __ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
14899   int16x8_t __ret;
14900   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14901   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14902   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14903   __ret = __rev0 - __rev1 * __rev2;
14904   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14905   return __ret;
14906 }
14907 #endif
14908 
14909 #ifdef __LITTLE_ENDIAN__
vmls_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)14910 __ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
14911   uint8x8_t __ret;
14912   __ret = __p0 - __p1 * __p2;
14913   return __ret;
14914 }
14915 #else
vmls_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)14916 __ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
14917   uint8x8_t __ret;
14918   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14919   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14920   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14921   __ret = __rev0 - __rev1 * __rev2;
14922   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14923   return __ret;
14924 }
14925 #endif
14926 
14927 #ifdef __LITTLE_ENDIAN__
vmls_u32(uint32x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)14928 __ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
14929   uint32x2_t __ret;
14930   __ret = __p0 - __p1 * __p2;
14931   return __ret;
14932 }
14933 #else
vmls_u32(uint32x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)14934 __ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
14935   uint32x2_t __ret;
14936   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14937   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14938   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
14939   __ret = __rev0 - __rev1 * __rev2;
14940   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14941   return __ret;
14942 }
14943 #endif
14944 
14945 #ifdef __LITTLE_ENDIAN__
vmls_u16(uint16x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)14946 __ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
14947   uint16x4_t __ret;
14948   __ret = __p0 - __p1 * __p2;
14949   return __ret;
14950 }
14951 #else
vmls_u16(uint16x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)14952 __ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
14953   uint16x4_t __ret;
14954   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14955   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14956   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14957   __ret = __rev0 - __rev1 * __rev2;
14958   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14959   return __ret;
14960 }
14961 #endif
14962 
14963 #ifdef __LITTLE_ENDIAN__
vmls_s8(int8x8_t __p0,int8x8_t __p1,int8x8_t __p2)14964 __ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
14965   int8x8_t __ret;
14966   __ret = __p0 - __p1 * __p2;
14967   return __ret;
14968 }
14969 #else
vmls_s8(int8x8_t __p0,int8x8_t __p1,int8x8_t __p2)14970 __ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
14971   int8x8_t __ret;
14972   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14973   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14974   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14975   __ret = __rev0 - __rev1 * __rev2;
14976   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14977   return __ret;
14978 }
14979 #endif
14980 
14981 #ifdef __LITTLE_ENDIAN__
vmls_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)14982 __ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
14983   float32x2_t __ret;
14984   __ret = __p0 - __p1 * __p2;
14985   return __ret;
14986 }
14987 #else
vmls_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)14988 __ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
14989   float32x2_t __ret;
14990   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14991   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14992   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
14993   __ret = __rev0 - __rev1 * __rev2;
14994   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14995   return __ret;
14996 }
14997 #endif
14998 
14999 #ifdef __LITTLE_ENDIAN__
vmls_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)15000 __ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
15001   int32x2_t __ret;
15002   __ret = __p0 - __p1 * __p2;
15003   return __ret;
15004 }
15005 #else
vmls_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)15006 __ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
15007   int32x2_t __ret;
15008   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15009   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15010   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
15011   __ret = __rev0 - __rev1 * __rev2;
15012   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15013   return __ret;
15014 }
15015 #endif
15016 
15017 #ifdef __LITTLE_ENDIAN__
vmls_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)15018 __ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
15019   int16x4_t __ret;
15020   __ret = __p0 - __p1 * __p2;
15021   return __ret;
15022 }
15023 #else
vmls_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)15024 __ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
15025   int16x4_t __ret;
15026   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15027   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15028   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
15029   __ret = __rev0 - __rev1 * __rev2;
15030   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15031   return __ret;
15032 }
15033 #endif
15034 
15035 #ifdef __LITTLE_ENDIAN__
15036 #define vmlsq_lane_u32(__p0_66, __p1_66, __p2_66, __p3_66) __extension__ ({ \
15037   uint32x4_t __ret_66; \
15038   uint32x4_t __s0_66 = __p0_66; \
15039   uint32x4_t __s1_66 = __p1_66; \
15040   uint32x2_t __s2_66 = __p2_66; \
15041   __ret_66 = __s0_66 - __s1_66 * splatq_lane_u32(__s2_66, __p3_66); \
15042   __ret_66; \
15043 })
15044 #else
15045 #define vmlsq_lane_u32(__p0_67, __p1_67, __p2_67, __p3_67) __extension__ ({ \
15046   uint32x4_t __ret_67; \
15047   uint32x4_t __s0_67 = __p0_67; \
15048   uint32x4_t __s1_67 = __p1_67; \
15049   uint32x2_t __s2_67 = __p2_67; \
15050   uint32x4_t __rev0_67;  __rev0_67 = __builtin_shufflevector(__s0_67, __s0_67, 3, 2, 1, 0); \
15051   uint32x4_t __rev1_67;  __rev1_67 = __builtin_shufflevector(__s1_67, __s1_67, 3, 2, 1, 0); \
15052   uint32x2_t __rev2_67;  __rev2_67 = __builtin_shufflevector(__s2_67, __s2_67, 1, 0); \
15053   __ret_67 = __rev0_67 - __rev1_67 * __noswap_splatq_lane_u32(__rev2_67, __p3_67); \
15054   __ret_67 = __builtin_shufflevector(__ret_67, __ret_67, 3, 2, 1, 0); \
15055   __ret_67; \
15056 })
15057 #endif
15058 
15059 #ifdef __LITTLE_ENDIAN__
15060 #define vmlsq_lane_u16(__p0_68, __p1_68, __p2_68, __p3_68) __extension__ ({ \
15061   uint16x8_t __ret_68; \
15062   uint16x8_t __s0_68 = __p0_68; \
15063   uint16x8_t __s1_68 = __p1_68; \
15064   uint16x4_t __s2_68 = __p2_68; \
15065   __ret_68 = __s0_68 - __s1_68 * splatq_lane_u16(__s2_68, __p3_68); \
15066   __ret_68; \
15067 })
15068 #else
15069 #define vmlsq_lane_u16(__p0_69, __p1_69, __p2_69, __p3_69) __extension__ ({ \
15070   uint16x8_t __ret_69; \
15071   uint16x8_t __s0_69 = __p0_69; \
15072   uint16x8_t __s1_69 = __p1_69; \
15073   uint16x4_t __s2_69 = __p2_69; \
15074   uint16x8_t __rev0_69;  __rev0_69 = __builtin_shufflevector(__s0_69, __s0_69, 7, 6, 5, 4, 3, 2, 1, 0); \
15075   uint16x8_t __rev1_69;  __rev1_69 = __builtin_shufflevector(__s1_69, __s1_69, 7, 6, 5, 4, 3, 2, 1, 0); \
15076   uint16x4_t __rev2_69;  __rev2_69 = __builtin_shufflevector(__s2_69, __s2_69, 3, 2, 1, 0); \
15077   __ret_69 = __rev0_69 - __rev1_69 * __noswap_splatq_lane_u16(__rev2_69, __p3_69); \
15078   __ret_69 = __builtin_shufflevector(__ret_69, __ret_69, 7, 6, 5, 4, 3, 2, 1, 0); \
15079   __ret_69; \
15080 })
15081 #endif
15082 
15083 #ifdef __LITTLE_ENDIAN__
15084 #define vmlsq_lane_f32(__p0_70, __p1_70, __p2_70, __p3_70) __extension__ ({ \
15085   float32x4_t __ret_70; \
15086   float32x4_t __s0_70 = __p0_70; \
15087   float32x4_t __s1_70 = __p1_70; \
15088   float32x2_t __s2_70 = __p2_70; \
15089   __ret_70 = __s0_70 - __s1_70 * splatq_lane_f32(__s2_70, __p3_70); \
15090   __ret_70; \
15091 })
15092 #else
15093 #define vmlsq_lane_f32(__p0_71, __p1_71, __p2_71, __p3_71) __extension__ ({ \
15094   float32x4_t __ret_71; \
15095   float32x4_t __s0_71 = __p0_71; \
15096   float32x4_t __s1_71 = __p1_71; \
15097   float32x2_t __s2_71 = __p2_71; \
15098   float32x4_t __rev0_71;  __rev0_71 = __builtin_shufflevector(__s0_71, __s0_71, 3, 2, 1, 0); \
15099   float32x4_t __rev1_71;  __rev1_71 = __builtin_shufflevector(__s1_71, __s1_71, 3, 2, 1, 0); \
15100   float32x2_t __rev2_71;  __rev2_71 = __builtin_shufflevector(__s2_71, __s2_71, 1, 0); \
15101   __ret_71 = __rev0_71 - __rev1_71 * __noswap_splatq_lane_f32(__rev2_71, __p3_71); \
15102   __ret_71 = __builtin_shufflevector(__ret_71, __ret_71, 3, 2, 1, 0); \
15103   __ret_71; \
15104 })
15105 #endif
15106 
15107 #ifdef __LITTLE_ENDIAN__
15108 #define vmlsq_lane_s32(__p0_72, __p1_72, __p2_72, __p3_72) __extension__ ({ \
15109   int32x4_t __ret_72; \
15110   int32x4_t __s0_72 = __p0_72; \
15111   int32x4_t __s1_72 = __p1_72; \
15112   int32x2_t __s2_72 = __p2_72; \
15113   __ret_72 = __s0_72 - __s1_72 * splatq_lane_s32(__s2_72, __p3_72); \
15114   __ret_72; \
15115 })
15116 #else
15117 #define vmlsq_lane_s32(__p0_73, __p1_73, __p2_73, __p3_73) __extension__ ({ \
15118   int32x4_t __ret_73; \
15119   int32x4_t __s0_73 = __p0_73; \
15120   int32x4_t __s1_73 = __p1_73; \
15121   int32x2_t __s2_73 = __p2_73; \
15122   int32x4_t __rev0_73;  __rev0_73 = __builtin_shufflevector(__s0_73, __s0_73, 3, 2, 1, 0); \
15123   int32x4_t __rev1_73;  __rev1_73 = __builtin_shufflevector(__s1_73, __s1_73, 3, 2, 1, 0); \
15124   int32x2_t __rev2_73;  __rev2_73 = __builtin_shufflevector(__s2_73, __s2_73, 1, 0); \
15125   __ret_73 = __rev0_73 - __rev1_73 * __noswap_splatq_lane_s32(__rev2_73, __p3_73); \
15126   __ret_73 = __builtin_shufflevector(__ret_73, __ret_73, 3, 2, 1, 0); \
15127   __ret_73; \
15128 })
15129 #endif
15130 
15131 #ifdef __LITTLE_ENDIAN__
15132 #define vmlsq_lane_s16(__p0_74, __p1_74, __p2_74, __p3_74) __extension__ ({ \
15133   int16x8_t __ret_74; \
15134   int16x8_t __s0_74 = __p0_74; \
15135   int16x8_t __s1_74 = __p1_74; \
15136   int16x4_t __s2_74 = __p2_74; \
15137   __ret_74 = __s0_74 - __s1_74 * splatq_lane_s16(__s2_74, __p3_74); \
15138   __ret_74; \
15139 })
15140 #else
15141 #define vmlsq_lane_s16(__p0_75, __p1_75, __p2_75, __p3_75) __extension__ ({ \
15142   int16x8_t __ret_75; \
15143   int16x8_t __s0_75 = __p0_75; \
15144   int16x8_t __s1_75 = __p1_75; \
15145   int16x4_t __s2_75 = __p2_75; \
15146   int16x8_t __rev0_75;  __rev0_75 = __builtin_shufflevector(__s0_75, __s0_75, 7, 6, 5, 4, 3, 2, 1, 0); \
15147   int16x8_t __rev1_75;  __rev1_75 = __builtin_shufflevector(__s1_75, __s1_75, 7, 6, 5, 4, 3, 2, 1, 0); \
15148   int16x4_t __rev2_75;  __rev2_75 = __builtin_shufflevector(__s2_75, __s2_75, 3, 2, 1, 0); \
15149   __ret_75 = __rev0_75 - __rev1_75 * __noswap_splatq_lane_s16(__rev2_75, __p3_75); \
15150   __ret_75 = __builtin_shufflevector(__ret_75, __ret_75, 7, 6, 5, 4, 3, 2, 1, 0); \
15151   __ret_75; \
15152 })
15153 #endif
15154 
15155 #ifdef __LITTLE_ENDIAN__
15156 #define vmls_lane_u32(__p0_76, __p1_76, __p2_76, __p3_76) __extension__ ({ \
15157   uint32x2_t __ret_76; \
15158   uint32x2_t __s0_76 = __p0_76; \
15159   uint32x2_t __s1_76 = __p1_76; \
15160   uint32x2_t __s2_76 = __p2_76; \
15161   __ret_76 = __s0_76 - __s1_76 * splat_lane_u32(__s2_76, __p3_76); \
15162   __ret_76; \
15163 })
15164 #else
15165 #define vmls_lane_u32(__p0_77, __p1_77, __p2_77, __p3_77) __extension__ ({ \
15166   uint32x2_t __ret_77; \
15167   uint32x2_t __s0_77 = __p0_77; \
15168   uint32x2_t __s1_77 = __p1_77; \
15169   uint32x2_t __s2_77 = __p2_77; \
15170   uint32x2_t __rev0_77;  __rev0_77 = __builtin_shufflevector(__s0_77, __s0_77, 1, 0); \
15171   uint32x2_t __rev1_77;  __rev1_77 = __builtin_shufflevector(__s1_77, __s1_77, 1, 0); \
15172   uint32x2_t __rev2_77;  __rev2_77 = __builtin_shufflevector(__s2_77, __s2_77, 1, 0); \
15173   __ret_77 = __rev0_77 - __rev1_77 * __noswap_splat_lane_u32(__rev2_77, __p3_77); \
15174   __ret_77 = __builtin_shufflevector(__ret_77, __ret_77, 1, 0); \
15175   __ret_77; \
15176 })
15177 #endif
15178 
15179 #ifdef __LITTLE_ENDIAN__
15180 #define vmls_lane_u16(__p0_78, __p1_78, __p2_78, __p3_78) __extension__ ({ \
15181   uint16x4_t __ret_78; \
15182   uint16x4_t __s0_78 = __p0_78; \
15183   uint16x4_t __s1_78 = __p1_78; \
15184   uint16x4_t __s2_78 = __p2_78; \
15185   __ret_78 = __s0_78 - __s1_78 * splat_lane_u16(__s2_78, __p3_78); \
15186   __ret_78; \
15187 })
15188 #else
15189 #define vmls_lane_u16(__p0_79, __p1_79, __p2_79, __p3_79) __extension__ ({ \
15190   uint16x4_t __ret_79; \
15191   uint16x4_t __s0_79 = __p0_79; \
15192   uint16x4_t __s1_79 = __p1_79; \
15193   uint16x4_t __s2_79 = __p2_79; \
15194   uint16x4_t __rev0_79;  __rev0_79 = __builtin_shufflevector(__s0_79, __s0_79, 3, 2, 1, 0); \
15195   uint16x4_t __rev1_79;  __rev1_79 = __builtin_shufflevector(__s1_79, __s1_79, 3, 2, 1, 0); \
15196   uint16x4_t __rev2_79;  __rev2_79 = __builtin_shufflevector(__s2_79, __s2_79, 3, 2, 1, 0); \
15197   __ret_79 = __rev0_79 - __rev1_79 * __noswap_splat_lane_u16(__rev2_79, __p3_79); \
15198   __ret_79 = __builtin_shufflevector(__ret_79, __ret_79, 3, 2, 1, 0); \
15199   __ret_79; \
15200 })
15201 #endif
15202 
15203 #ifdef __LITTLE_ENDIAN__
15204 #define vmls_lane_f32(__p0_80, __p1_80, __p2_80, __p3_80) __extension__ ({ \
15205   float32x2_t __ret_80; \
15206   float32x2_t __s0_80 = __p0_80; \
15207   float32x2_t __s1_80 = __p1_80; \
15208   float32x2_t __s2_80 = __p2_80; \
15209   __ret_80 = __s0_80 - __s1_80 * splat_lane_f32(__s2_80, __p3_80); \
15210   __ret_80; \
15211 })
15212 #else
15213 #define vmls_lane_f32(__p0_81, __p1_81, __p2_81, __p3_81) __extension__ ({ \
15214   float32x2_t __ret_81; \
15215   float32x2_t __s0_81 = __p0_81; \
15216   float32x2_t __s1_81 = __p1_81; \
15217   float32x2_t __s2_81 = __p2_81; \
15218   float32x2_t __rev0_81;  __rev0_81 = __builtin_shufflevector(__s0_81, __s0_81, 1, 0); \
15219   float32x2_t __rev1_81;  __rev1_81 = __builtin_shufflevector(__s1_81, __s1_81, 1, 0); \
15220   float32x2_t __rev2_81;  __rev2_81 = __builtin_shufflevector(__s2_81, __s2_81, 1, 0); \
15221   __ret_81 = __rev0_81 - __rev1_81 * __noswap_splat_lane_f32(__rev2_81, __p3_81); \
15222   __ret_81 = __builtin_shufflevector(__ret_81, __ret_81, 1, 0); \
15223   __ret_81; \
15224 })
15225 #endif
15226 
15227 #ifdef __LITTLE_ENDIAN__
15228 #define vmls_lane_s32(__p0_82, __p1_82, __p2_82, __p3_82) __extension__ ({ \
15229   int32x2_t __ret_82; \
15230   int32x2_t __s0_82 = __p0_82; \
15231   int32x2_t __s1_82 = __p1_82; \
15232   int32x2_t __s2_82 = __p2_82; \
15233   __ret_82 = __s0_82 - __s1_82 * splat_lane_s32(__s2_82, __p3_82); \
15234   __ret_82; \
15235 })
15236 #else
15237 #define vmls_lane_s32(__p0_83, __p1_83, __p2_83, __p3_83) __extension__ ({ \
15238   int32x2_t __ret_83; \
15239   int32x2_t __s0_83 = __p0_83; \
15240   int32x2_t __s1_83 = __p1_83; \
15241   int32x2_t __s2_83 = __p2_83; \
15242   int32x2_t __rev0_83;  __rev0_83 = __builtin_shufflevector(__s0_83, __s0_83, 1, 0); \
15243   int32x2_t __rev1_83;  __rev1_83 = __builtin_shufflevector(__s1_83, __s1_83, 1, 0); \
15244   int32x2_t __rev2_83;  __rev2_83 = __builtin_shufflevector(__s2_83, __s2_83, 1, 0); \
15245   __ret_83 = __rev0_83 - __rev1_83 * __noswap_splat_lane_s32(__rev2_83, __p3_83); \
15246   __ret_83 = __builtin_shufflevector(__ret_83, __ret_83, 1, 0); \
15247   __ret_83; \
15248 })
15249 #endif
15250 
15251 #ifdef __LITTLE_ENDIAN__
15252 #define vmls_lane_s16(__p0_84, __p1_84, __p2_84, __p3_84) __extension__ ({ \
15253   int16x4_t __ret_84; \
15254   int16x4_t __s0_84 = __p0_84; \
15255   int16x4_t __s1_84 = __p1_84; \
15256   int16x4_t __s2_84 = __p2_84; \
15257   __ret_84 = __s0_84 - __s1_84 * splat_lane_s16(__s2_84, __p3_84); \
15258   __ret_84; \
15259 })
15260 #else
15261 #define vmls_lane_s16(__p0_85, __p1_85, __p2_85, __p3_85) __extension__ ({ \
15262   int16x4_t __ret_85; \
15263   int16x4_t __s0_85 = __p0_85; \
15264   int16x4_t __s1_85 = __p1_85; \
15265   int16x4_t __s2_85 = __p2_85; \
15266   int16x4_t __rev0_85;  __rev0_85 = __builtin_shufflevector(__s0_85, __s0_85, 3, 2, 1, 0); \
15267   int16x4_t __rev1_85;  __rev1_85 = __builtin_shufflevector(__s1_85, __s1_85, 3, 2, 1, 0); \
15268   int16x4_t __rev2_85;  __rev2_85 = __builtin_shufflevector(__s2_85, __s2_85, 3, 2, 1, 0); \
15269   __ret_85 = __rev0_85 - __rev1_85 * __noswap_splat_lane_s16(__rev2_85, __p3_85); \
15270   __ret_85 = __builtin_shufflevector(__ret_85, __ret_85, 3, 2, 1, 0); \
15271   __ret_85; \
15272 })
15273 #endif
15274 
15275 #ifdef __LITTLE_ENDIAN__
vmlsq_n_u32(uint32x4_t __p0,uint32x4_t __p1,uint32_t __p2)15276 __ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
15277   uint32x4_t __ret;
15278   __ret = __p0 - __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
15279   return __ret;
15280 }
15281 #else
vmlsq_n_u32(uint32x4_t __p0,uint32x4_t __p1,uint32_t __p2)15282 __ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
15283   uint32x4_t __ret;
15284   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15285   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15286   __ret = __rev0 - __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
15287   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15288   return __ret;
15289 }
15290 #endif
15291 
15292 #ifdef __LITTLE_ENDIAN__
vmlsq_n_u16(uint16x8_t __p0,uint16x8_t __p1,uint16_t __p2)15293 __ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
15294   uint16x8_t __ret;
15295   __ret = __p0 - __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
15296   return __ret;
15297 }
15298 #else
vmlsq_n_u16(uint16x8_t __p0,uint16x8_t __p1,uint16_t __p2)15299 __ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
15300   uint16x8_t __ret;
15301   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15302   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15303   __ret = __rev0 - __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
15304   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15305   return __ret;
15306 }
15307 #endif
15308 
15309 #ifdef __LITTLE_ENDIAN__
vmlsq_n_f32(float32x4_t __p0,float32x4_t __p1,float32_t __p2)15310 __ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
15311   float32x4_t __ret;
15312   __ret = __p0 - __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
15313   return __ret;
15314 }
15315 #else
vmlsq_n_f32(float32x4_t __p0,float32x4_t __p1,float32_t __p2)15316 __ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
15317   float32x4_t __ret;
15318   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15319   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15320   __ret = __rev0 - __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
15321   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15322   return __ret;
15323 }
15324 #endif
15325 
15326 #ifdef __LITTLE_ENDIAN__
vmlsq_n_s32(int32x4_t __p0,int32x4_t __p1,int32_t __p2)15327 __ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
15328   int32x4_t __ret;
15329   __ret = __p0 - __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
15330   return __ret;
15331 }
15332 #else
vmlsq_n_s32(int32x4_t __p0,int32x4_t __p1,int32_t __p2)15333 __ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
15334   int32x4_t __ret;
15335   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15336   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15337   __ret = __rev0 - __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
15338   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15339   return __ret;
15340 }
15341 #endif
15342 
15343 #ifdef __LITTLE_ENDIAN__
vmlsq_n_s16(int16x8_t __p0,int16x8_t __p1,int16_t __p2)15344 __ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
15345   int16x8_t __ret;
15346   __ret = __p0 - __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
15347   return __ret;
15348 }
15349 #else
vmlsq_n_s16(int16x8_t __p0,int16x8_t __p1,int16_t __p2)15350 __ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
15351   int16x8_t __ret;
15352   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15353   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15354   __ret = __rev0 - __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
15355   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15356   return __ret;
15357 }
15358 #endif
15359 
15360 #ifdef __LITTLE_ENDIAN__
vmls_n_u32(uint32x2_t __p0,uint32x2_t __p1,uint32_t __p2)15361 __ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
15362   uint32x2_t __ret;
15363   __ret = __p0 - __p1 * (uint32x2_t) {__p2, __p2};
15364   return __ret;
15365 }
15366 #else
vmls_n_u32(uint32x2_t __p0,uint32x2_t __p1,uint32_t __p2)15367 __ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
15368   uint32x2_t __ret;
15369   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15370   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15371   __ret = __rev0 - __rev1 * (uint32x2_t) {__p2, __p2};
15372   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15373   return __ret;
15374 }
15375 #endif
15376 
15377 #ifdef __LITTLE_ENDIAN__
vmls_n_u16(uint16x4_t __p0,uint16x4_t __p1,uint16_t __p2)15378 __ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
15379   uint16x4_t __ret;
15380   __ret = __p0 - __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
15381   return __ret;
15382 }
15383 #else
vmls_n_u16(uint16x4_t __p0,uint16x4_t __p1,uint16_t __p2)15384 __ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
15385   uint16x4_t __ret;
15386   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15387   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15388   __ret = __rev0 - __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
15389   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15390   return __ret;
15391 }
15392 #endif
15393 
15394 #ifdef __LITTLE_ENDIAN__
vmls_n_f32(float32x2_t __p0,float32x2_t __p1,float32_t __p2)15395 __ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
15396   float32x2_t __ret;
15397   __ret = __p0 - __p1 * (float32x2_t) {__p2, __p2};
15398   return __ret;
15399 }
15400 #else
vmls_n_f32(float32x2_t __p0,float32x2_t __p1,float32_t __p2)15401 __ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
15402   float32x2_t __ret;
15403   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15404   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15405   __ret = __rev0 - __rev1 * (float32x2_t) {__p2, __p2};
15406   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15407   return __ret;
15408 }
15409 #endif
15410 
15411 #ifdef __LITTLE_ENDIAN__
vmls_n_s32(int32x2_t __p0,int32x2_t __p1,int32_t __p2)15412 __ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
15413   int32x2_t __ret;
15414   __ret = __p0 - __p1 * (int32x2_t) {__p2, __p2};
15415   return __ret;
15416 }
15417 #else
vmls_n_s32(int32x2_t __p0,int32x2_t __p1,int32_t __p2)15418 __ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
15419   int32x2_t __ret;
15420   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15421   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15422   __ret = __rev0 - __rev1 * (int32x2_t) {__p2, __p2};
15423   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15424   return __ret;
15425 }
15426 #endif
15427 
15428 #ifdef __LITTLE_ENDIAN__
vmls_n_s16(int16x4_t __p0,int16x4_t __p1,int16_t __p2)15429 __ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
15430   int16x4_t __ret;
15431   __ret = __p0 - __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
15432   return __ret;
15433 }
15434 #else
vmls_n_s16(int16x4_t __p0,int16x4_t __p1,int16_t __p2)15435 __ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
15436   int16x4_t __ret;
15437   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15438   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15439   __ret = __rev0 - __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
15440   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15441   return __ret;
15442 }
15443 #endif
15444 
15445 #ifdef __LITTLE_ENDIAN__
vmov_n_p8(poly8_t __p0)15446 __ai poly8x8_t vmov_n_p8(poly8_t __p0) {
15447   poly8x8_t __ret;
15448   __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15449   return __ret;
15450 }
15451 #else
vmov_n_p8(poly8_t __p0)15452 __ai poly8x8_t vmov_n_p8(poly8_t __p0) {
15453   poly8x8_t __ret;
15454   __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15455   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15456   return __ret;
15457 }
15458 #endif
15459 
15460 #ifdef __LITTLE_ENDIAN__
vmov_n_p16(poly16_t __p0)15461 __ai poly16x4_t vmov_n_p16(poly16_t __p0) {
15462   poly16x4_t __ret;
15463   __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
15464   return __ret;
15465 }
15466 #else
vmov_n_p16(poly16_t __p0)15467 __ai poly16x4_t vmov_n_p16(poly16_t __p0) {
15468   poly16x4_t __ret;
15469   __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
15470   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15471   return __ret;
15472 }
15473 #endif
15474 
15475 #ifdef __LITTLE_ENDIAN__
vmovq_n_p8(poly8_t __p0)15476 __ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
15477   poly8x16_t __ret;
15478   __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15479   return __ret;
15480 }
15481 #else
vmovq_n_p8(poly8_t __p0)15482 __ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
15483   poly8x16_t __ret;
15484   __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15485   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15486   return __ret;
15487 }
15488 #endif
15489 
15490 #ifdef __LITTLE_ENDIAN__
vmovq_n_p16(poly16_t __p0)15491 __ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
15492   poly16x8_t __ret;
15493   __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15494   return __ret;
15495 }
15496 #else
vmovq_n_p16(poly16_t __p0)15497 __ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
15498   poly16x8_t __ret;
15499   __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15500   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15501   return __ret;
15502 }
15503 #endif
15504 
15505 #ifdef __LITTLE_ENDIAN__
vmovq_n_u8(uint8_t __p0)15506 __ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
15507   uint8x16_t __ret;
15508   __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15509   return __ret;
15510 }
15511 #else
vmovq_n_u8(uint8_t __p0)15512 __ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
15513   uint8x16_t __ret;
15514   __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15515   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15516   return __ret;
15517 }
15518 #endif
15519 
15520 #ifdef __LITTLE_ENDIAN__
vmovq_n_u32(uint32_t __p0)15521 __ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
15522   uint32x4_t __ret;
15523   __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
15524   return __ret;
15525 }
15526 #else
vmovq_n_u32(uint32_t __p0)15527 __ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
15528   uint32x4_t __ret;
15529   __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
15530   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15531   return __ret;
15532 }
15533 #endif
15534 
15535 #ifdef __LITTLE_ENDIAN__
vmovq_n_u64(uint64_t __p0)15536 __ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
15537   uint64x2_t __ret;
15538   __ret = (uint64x2_t) {__p0, __p0};
15539   return __ret;
15540 }
15541 #else
vmovq_n_u64(uint64_t __p0)15542 __ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
15543   uint64x2_t __ret;
15544   __ret = (uint64x2_t) {__p0, __p0};
15545   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15546   return __ret;
15547 }
15548 #endif
15549 
15550 #ifdef __LITTLE_ENDIAN__
vmovq_n_u16(uint16_t __p0)15551 __ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
15552   uint16x8_t __ret;
15553   __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15554   return __ret;
15555 }
15556 #else
vmovq_n_u16(uint16_t __p0)15557 __ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
15558   uint16x8_t __ret;
15559   __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15560   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15561   return __ret;
15562 }
15563 #endif
15564 
15565 #ifdef __LITTLE_ENDIAN__
vmovq_n_s8(int8_t __p0)15566 __ai int8x16_t vmovq_n_s8(int8_t __p0) {
15567   int8x16_t __ret;
15568   __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15569   return __ret;
15570 }
15571 #else
vmovq_n_s8(int8_t __p0)15572 __ai int8x16_t vmovq_n_s8(int8_t __p0) {
15573   int8x16_t __ret;
15574   __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15575   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15576   return __ret;
15577 }
15578 #endif
15579 
15580 #ifdef __LITTLE_ENDIAN__
vmovq_n_f32(float32_t __p0)15581 __ai float32x4_t vmovq_n_f32(float32_t __p0) {
15582   float32x4_t __ret;
15583   __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
15584   return __ret;
15585 }
15586 #else
vmovq_n_f32(float32_t __p0)15587 __ai float32x4_t vmovq_n_f32(float32_t __p0) {
15588   float32x4_t __ret;
15589   __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
15590   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15591   return __ret;
15592 }
15593 #endif
15594 
15595 #ifdef __LITTLE_ENDIAN__
15596 #define vmovq_n_f16(__p0) __extension__ ({ \
15597   float16x8_t __ret; \
15598   float16_t __s0 = __p0; \
15599   __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
15600   __ret; \
15601 })
15602 #else
15603 #define vmovq_n_f16(__p0) __extension__ ({ \
15604   float16x8_t __ret; \
15605   float16_t __s0 = __p0; \
15606   __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
15607   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
15608   __ret; \
15609 })
15610 #endif
15611 
15612 #ifdef __LITTLE_ENDIAN__
vmovq_n_s32(int32_t __p0)15613 __ai int32x4_t vmovq_n_s32(int32_t __p0) {
15614   int32x4_t __ret;
15615   __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
15616   return __ret;
15617 }
15618 #else
vmovq_n_s32(int32_t __p0)15619 __ai int32x4_t vmovq_n_s32(int32_t __p0) {
15620   int32x4_t __ret;
15621   __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
15622   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15623   return __ret;
15624 }
15625 #endif
15626 
15627 #ifdef __LITTLE_ENDIAN__
vmovq_n_s64(int64_t __p0)15628 __ai int64x2_t vmovq_n_s64(int64_t __p0) {
15629   int64x2_t __ret;
15630   __ret = (int64x2_t) {__p0, __p0};
15631   return __ret;
15632 }
15633 #else
vmovq_n_s64(int64_t __p0)15634 __ai int64x2_t vmovq_n_s64(int64_t __p0) {
15635   int64x2_t __ret;
15636   __ret = (int64x2_t) {__p0, __p0};
15637   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15638   return __ret;
15639 }
15640 #endif
15641 
15642 #ifdef __LITTLE_ENDIAN__
vmovq_n_s16(int16_t __p0)15643 __ai int16x8_t vmovq_n_s16(int16_t __p0) {
15644   int16x8_t __ret;
15645   __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15646   return __ret;
15647 }
15648 #else
vmovq_n_s16(int16_t __p0)15649 __ai int16x8_t vmovq_n_s16(int16_t __p0) {
15650   int16x8_t __ret;
15651   __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15652   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15653   return __ret;
15654 }
15655 #endif
15656 
15657 #ifdef __LITTLE_ENDIAN__
vmov_n_u8(uint8_t __p0)15658 __ai uint8x8_t vmov_n_u8(uint8_t __p0) {
15659   uint8x8_t __ret;
15660   __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15661   return __ret;
15662 }
15663 #else
vmov_n_u8(uint8_t __p0)15664 __ai uint8x8_t vmov_n_u8(uint8_t __p0) {
15665   uint8x8_t __ret;
15666   __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15667   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15668   return __ret;
15669 }
15670 #endif
15671 
15672 #ifdef __LITTLE_ENDIAN__
vmov_n_u32(uint32_t __p0)15673 __ai uint32x2_t vmov_n_u32(uint32_t __p0) {
15674   uint32x2_t __ret;
15675   __ret = (uint32x2_t) {__p0, __p0};
15676   return __ret;
15677 }
15678 #else
vmov_n_u32(uint32_t __p0)15679 __ai uint32x2_t vmov_n_u32(uint32_t __p0) {
15680   uint32x2_t __ret;
15681   __ret = (uint32x2_t) {__p0, __p0};
15682   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15683   return __ret;
15684 }
15685 #endif
15686 
vmov_n_u64(uint64_t __p0)15687 __ai uint64x1_t vmov_n_u64(uint64_t __p0) {
15688   uint64x1_t __ret;
15689   __ret = (uint64x1_t) {__p0};
15690   return __ret;
15691 }
15692 #ifdef __LITTLE_ENDIAN__
vmov_n_u16(uint16_t __p0)15693 __ai uint16x4_t vmov_n_u16(uint16_t __p0) {
15694   uint16x4_t __ret;
15695   __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
15696   return __ret;
15697 }
15698 #else
vmov_n_u16(uint16_t __p0)15699 __ai uint16x4_t vmov_n_u16(uint16_t __p0) {
15700   uint16x4_t __ret;
15701   __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
15702   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15703   return __ret;
15704 }
15705 #endif
15706 
15707 #ifdef __LITTLE_ENDIAN__
vmov_n_s8(int8_t __p0)15708 __ai int8x8_t vmov_n_s8(int8_t __p0) {
15709   int8x8_t __ret;
15710   __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15711   return __ret;
15712 }
15713 #else
vmov_n_s8(int8_t __p0)15714 __ai int8x8_t vmov_n_s8(int8_t __p0) {
15715   int8x8_t __ret;
15716   __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15717   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15718   return __ret;
15719 }
15720 #endif
15721 
15722 #ifdef __LITTLE_ENDIAN__
vmov_n_f32(float32_t __p0)15723 __ai float32x2_t vmov_n_f32(float32_t __p0) {
15724   float32x2_t __ret;
15725   __ret = (float32x2_t) {__p0, __p0};
15726   return __ret;
15727 }
15728 #else
vmov_n_f32(float32_t __p0)15729 __ai float32x2_t vmov_n_f32(float32_t __p0) {
15730   float32x2_t __ret;
15731   __ret = (float32x2_t) {__p0, __p0};
15732   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15733   return __ret;
15734 }
15735 #endif
15736 
15737 #ifdef __LITTLE_ENDIAN__
15738 #define vmov_n_f16(__p0) __extension__ ({ \
15739   float16x4_t __ret; \
15740   float16_t __s0 = __p0; \
15741   __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
15742   __ret; \
15743 })
15744 #else
15745 #define vmov_n_f16(__p0) __extension__ ({ \
15746   float16x4_t __ret; \
15747   float16_t __s0 = __p0; \
15748   __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
15749   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
15750   __ret; \
15751 })
15752 #endif
15753 
15754 #ifdef __LITTLE_ENDIAN__
vmov_n_s32(int32_t __p0)15755 __ai int32x2_t vmov_n_s32(int32_t __p0) {
15756   int32x2_t __ret;
15757   __ret = (int32x2_t) {__p0, __p0};
15758   return __ret;
15759 }
15760 #else
vmov_n_s32(int32_t __p0)15761 __ai int32x2_t vmov_n_s32(int32_t __p0) {
15762   int32x2_t __ret;
15763   __ret = (int32x2_t) {__p0, __p0};
15764   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15765   return __ret;
15766 }
15767 #endif
15768 
vmov_n_s64(int64_t __p0)15769 __ai int64x1_t vmov_n_s64(int64_t __p0) {
15770   int64x1_t __ret;
15771   __ret = (int64x1_t) {__p0};
15772   return __ret;
15773 }
15774 #ifdef __LITTLE_ENDIAN__
vmov_n_s16(int16_t __p0)15775 __ai int16x4_t vmov_n_s16(int16_t __p0) {
15776   int16x4_t __ret;
15777   __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
15778   return __ret;
15779 }
15780 #else
vmov_n_s16(int16_t __p0)15781 __ai int16x4_t vmov_n_s16(int16_t __p0) {
15782   int16x4_t __ret;
15783   __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
15784   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15785   return __ret;
15786 }
15787 #endif
15788 
15789 #ifdef __LITTLE_ENDIAN__
vmovl_u8(uint8x8_t __p0)15790 __ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
15791   uint16x8_t __ret;
15792   __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
15793   return __ret;
15794 }
15795 #else
vmovl_u8(uint8x8_t __p0)15796 __ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
15797   uint16x8_t __ret;
15798   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15799   __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 49);
15800   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15801   return __ret;
15802 }
__noswap_vmovl_u8(uint8x8_t __p0)15803 __ai uint16x8_t __noswap_vmovl_u8(uint8x8_t __p0) {
15804   uint16x8_t __ret;
15805   __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
15806   return __ret;
15807 }
15808 #endif
15809 
15810 #ifdef __LITTLE_ENDIAN__
vmovl_u32(uint32x2_t __p0)15811 __ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
15812   uint64x2_t __ret;
15813   __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
15814   return __ret;
15815 }
15816 #else
vmovl_u32(uint32x2_t __p0)15817 __ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
15818   uint64x2_t __ret;
15819   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15820   __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 51);
15821   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15822   return __ret;
15823 }
__noswap_vmovl_u32(uint32x2_t __p0)15824 __ai uint64x2_t __noswap_vmovl_u32(uint32x2_t __p0) {
15825   uint64x2_t __ret;
15826   __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
15827   return __ret;
15828 }
15829 #endif
15830 
15831 #ifdef __LITTLE_ENDIAN__
vmovl_u16(uint16x4_t __p0)15832 __ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
15833   uint32x4_t __ret;
15834   __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
15835   return __ret;
15836 }
15837 #else
vmovl_u16(uint16x4_t __p0)15838 __ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
15839   uint32x4_t __ret;
15840   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15841   __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 50);
15842   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15843   return __ret;
15844 }
__noswap_vmovl_u16(uint16x4_t __p0)15845 __ai uint32x4_t __noswap_vmovl_u16(uint16x4_t __p0) {
15846   uint32x4_t __ret;
15847   __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
15848   return __ret;
15849 }
15850 #endif
15851 
15852 #ifdef __LITTLE_ENDIAN__
vmovl_s8(int8x8_t __p0)15853 __ai int16x8_t vmovl_s8(int8x8_t __p0) {
15854   int16x8_t __ret;
15855   __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
15856   return __ret;
15857 }
15858 #else
vmovl_s8(int8x8_t __p0)15859 __ai int16x8_t vmovl_s8(int8x8_t __p0) {
15860   int16x8_t __ret;
15861   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15862   __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 33);
15863   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15864   return __ret;
15865 }
__noswap_vmovl_s8(int8x8_t __p0)15866 __ai int16x8_t __noswap_vmovl_s8(int8x8_t __p0) {
15867   int16x8_t __ret;
15868   __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
15869   return __ret;
15870 }
15871 #endif
15872 
15873 #ifdef __LITTLE_ENDIAN__
vmovl_s32(int32x2_t __p0)15874 __ai int64x2_t vmovl_s32(int32x2_t __p0) {
15875   int64x2_t __ret;
15876   __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
15877   return __ret;
15878 }
15879 #else
vmovl_s32(int32x2_t __p0)15880 __ai int64x2_t vmovl_s32(int32x2_t __p0) {
15881   int64x2_t __ret;
15882   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15883   __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 35);
15884   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15885   return __ret;
15886 }
__noswap_vmovl_s32(int32x2_t __p0)15887 __ai int64x2_t __noswap_vmovl_s32(int32x2_t __p0) {
15888   int64x2_t __ret;
15889   __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
15890   return __ret;
15891 }
15892 #endif
15893 
15894 #ifdef __LITTLE_ENDIAN__
vmovl_s16(int16x4_t __p0)15895 __ai int32x4_t vmovl_s16(int16x4_t __p0) {
15896   int32x4_t __ret;
15897   __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
15898   return __ret;
15899 }
15900 #else
vmovl_s16(int16x4_t __p0)15901 __ai int32x4_t vmovl_s16(int16x4_t __p0) {
15902   int32x4_t __ret;
15903   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15904   __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 34);
15905   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15906   return __ret;
15907 }
__noswap_vmovl_s16(int16x4_t __p0)15908 __ai int32x4_t __noswap_vmovl_s16(int16x4_t __p0) {
15909   int32x4_t __ret;
15910   __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
15911   return __ret;
15912 }
15913 #endif
15914 
15915 #ifdef __LITTLE_ENDIAN__
vmovn_u32(uint32x4_t __p0)15916 __ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
15917   uint16x4_t __ret;
15918   __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
15919   return __ret;
15920 }
15921 #else
vmovn_u32(uint32x4_t __p0)15922 __ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
15923   uint16x4_t __ret;
15924   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15925   __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 17);
15926   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15927   return __ret;
15928 }
__noswap_vmovn_u32(uint32x4_t __p0)15929 __ai uint16x4_t __noswap_vmovn_u32(uint32x4_t __p0) {
15930   uint16x4_t __ret;
15931   __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
15932   return __ret;
15933 }
15934 #endif
15935 
15936 #ifdef __LITTLE_ENDIAN__
vmovn_u64(uint64x2_t __p0)15937 __ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
15938   uint32x2_t __ret;
15939   __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
15940   return __ret;
15941 }
15942 #else
vmovn_u64(uint64x2_t __p0)15943 __ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
15944   uint32x2_t __ret;
15945   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15946   __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 18);
15947   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15948   return __ret;
15949 }
__noswap_vmovn_u64(uint64x2_t __p0)15950 __ai uint32x2_t __noswap_vmovn_u64(uint64x2_t __p0) {
15951   uint32x2_t __ret;
15952   __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
15953   return __ret;
15954 }
15955 #endif
15956 
15957 #ifdef __LITTLE_ENDIAN__
vmovn_u16(uint16x8_t __p0)15958 __ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
15959   uint8x8_t __ret;
15960   __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
15961   return __ret;
15962 }
15963 #else
vmovn_u16(uint16x8_t __p0)15964 __ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
15965   uint8x8_t __ret;
15966   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15967   __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 16);
15968   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15969   return __ret;
15970 }
__noswap_vmovn_u16(uint16x8_t __p0)15971 __ai uint8x8_t __noswap_vmovn_u16(uint16x8_t __p0) {
15972   uint8x8_t __ret;
15973   __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
15974   return __ret;
15975 }
15976 #endif
15977 
15978 #ifdef __LITTLE_ENDIAN__
vmovn_s32(int32x4_t __p0)15979 __ai int16x4_t vmovn_s32(int32x4_t __p0) {
15980   int16x4_t __ret;
15981   __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
15982   return __ret;
15983 }
15984 #else
vmovn_s32(int32x4_t __p0)15985 __ai int16x4_t vmovn_s32(int32x4_t __p0) {
15986   int16x4_t __ret;
15987   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15988   __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 1);
15989   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15990   return __ret;
15991 }
__noswap_vmovn_s32(int32x4_t __p0)15992 __ai int16x4_t __noswap_vmovn_s32(int32x4_t __p0) {
15993   int16x4_t __ret;
15994   __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
15995   return __ret;
15996 }
15997 #endif
15998 
15999 #ifdef __LITTLE_ENDIAN__
vmovn_s64(int64x2_t __p0)16000 __ai int32x2_t vmovn_s64(int64x2_t __p0) {
16001   int32x2_t __ret;
16002   __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
16003   return __ret;
16004 }
16005 #else
vmovn_s64(int64x2_t __p0)16006 __ai int32x2_t vmovn_s64(int64x2_t __p0) {
16007   int32x2_t __ret;
16008   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16009   __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 2);
16010   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16011   return __ret;
16012 }
__noswap_vmovn_s64(int64x2_t __p0)16013 __ai int32x2_t __noswap_vmovn_s64(int64x2_t __p0) {
16014   int32x2_t __ret;
16015   __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
16016   return __ret;
16017 }
16018 #endif
16019 
16020 #ifdef __LITTLE_ENDIAN__
vmovn_s16(int16x8_t __p0)16021 __ai int8x8_t vmovn_s16(int16x8_t __p0) {
16022   int8x8_t __ret;
16023   __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
16024   return __ret;
16025 }
16026 #else
vmovn_s16(int16x8_t __p0)16027 __ai int8x8_t vmovn_s16(int16x8_t __p0) {
16028   int8x8_t __ret;
16029   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16030   __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 0);
16031   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16032   return __ret;
16033 }
__noswap_vmovn_s16(int16x8_t __p0)16034 __ai int8x8_t __noswap_vmovn_s16(int16x8_t __p0) {
16035   int8x8_t __ret;
16036   __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
16037   return __ret;
16038 }
16039 #endif
16040 
16041 #ifdef __LITTLE_ENDIAN__
vmulq_u8(uint8x16_t __p0,uint8x16_t __p1)16042 __ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16043   uint8x16_t __ret;
16044   __ret = __p0 * __p1;
16045   return __ret;
16046 }
16047 #else
vmulq_u8(uint8x16_t __p0,uint8x16_t __p1)16048 __ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16049   uint8x16_t __ret;
16050   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16051   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16052   __ret = __rev0 * __rev1;
16053   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16054   return __ret;
16055 }
16056 #endif
16057 
16058 #ifdef __LITTLE_ENDIAN__
vmulq_u32(uint32x4_t __p0,uint32x4_t __p1)16059 __ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16060   uint32x4_t __ret;
16061   __ret = __p0 * __p1;
16062   return __ret;
16063 }
16064 #else
vmulq_u32(uint32x4_t __p0,uint32x4_t __p1)16065 __ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16066   uint32x4_t __ret;
16067   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16068   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16069   __ret = __rev0 * __rev1;
16070   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16071   return __ret;
16072 }
16073 #endif
16074 
16075 #ifdef __LITTLE_ENDIAN__
vmulq_u16(uint16x8_t __p0,uint16x8_t __p1)16076 __ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16077   uint16x8_t __ret;
16078   __ret = __p0 * __p1;
16079   return __ret;
16080 }
16081 #else
vmulq_u16(uint16x8_t __p0,uint16x8_t __p1)16082 __ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16083   uint16x8_t __ret;
16084   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16085   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16086   __ret = __rev0 * __rev1;
16087   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16088   return __ret;
16089 }
16090 #endif
16091 
16092 #ifdef __LITTLE_ENDIAN__
vmulq_s8(int8x16_t __p0,int8x16_t __p1)16093 __ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
16094   int8x16_t __ret;
16095   __ret = __p0 * __p1;
16096   return __ret;
16097 }
16098 #else
vmulq_s8(int8x16_t __p0,int8x16_t __p1)16099 __ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
16100   int8x16_t __ret;
16101   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16102   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16103   __ret = __rev0 * __rev1;
16104   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16105   return __ret;
16106 }
16107 #endif
16108 
16109 #ifdef __LITTLE_ENDIAN__
vmulq_f32(float32x4_t __p0,float32x4_t __p1)16110 __ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
16111   float32x4_t __ret;
16112   __ret = __p0 * __p1;
16113   return __ret;
16114 }
16115 #else
vmulq_f32(float32x4_t __p0,float32x4_t __p1)16116 __ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
16117   float32x4_t __ret;
16118   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16119   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16120   __ret = __rev0 * __rev1;
16121   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16122   return __ret;
16123 }
16124 #endif
16125 
16126 #ifdef __LITTLE_ENDIAN__
vmulq_s32(int32x4_t __p0,int32x4_t __p1)16127 __ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
16128   int32x4_t __ret;
16129   __ret = __p0 * __p1;
16130   return __ret;
16131 }
16132 #else
vmulq_s32(int32x4_t __p0,int32x4_t __p1)16133 __ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
16134   int32x4_t __ret;
16135   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16136   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16137   __ret = __rev0 * __rev1;
16138   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16139   return __ret;
16140 }
16141 #endif
16142 
16143 #ifdef __LITTLE_ENDIAN__
vmulq_s16(int16x8_t __p0,int16x8_t __p1)16144 __ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
16145   int16x8_t __ret;
16146   __ret = __p0 * __p1;
16147   return __ret;
16148 }
16149 #else
vmulq_s16(int16x8_t __p0,int16x8_t __p1)16150 __ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
16151   int16x8_t __ret;
16152   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16153   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16154   __ret = __rev0 * __rev1;
16155   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16156   return __ret;
16157 }
16158 #endif
16159 
16160 #ifdef __LITTLE_ENDIAN__
vmul_u8(uint8x8_t __p0,uint8x8_t __p1)16161 __ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
16162   uint8x8_t __ret;
16163   __ret = __p0 * __p1;
16164   return __ret;
16165 }
16166 #else
vmul_u8(uint8x8_t __p0,uint8x8_t __p1)16167 __ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
16168   uint8x8_t __ret;
16169   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16170   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16171   __ret = __rev0 * __rev1;
16172   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16173   return __ret;
16174 }
16175 #endif
16176 
16177 #ifdef __LITTLE_ENDIAN__
vmul_u32(uint32x2_t __p0,uint32x2_t __p1)16178 __ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
16179   uint32x2_t __ret;
16180   __ret = __p0 * __p1;
16181   return __ret;
16182 }
16183 #else
vmul_u32(uint32x2_t __p0,uint32x2_t __p1)16184 __ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
16185   uint32x2_t __ret;
16186   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16187   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16188   __ret = __rev0 * __rev1;
16189   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16190   return __ret;
16191 }
16192 #endif
16193 
16194 #ifdef __LITTLE_ENDIAN__
vmul_u16(uint16x4_t __p0,uint16x4_t __p1)16195 __ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
16196   uint16x4_t __ret;
16197   __ret = __p0 * __p1;
16198   return __ret;
16199 }
16200 #else
vmul_u16(uint16x4_t __p0,uint16x4_t __p1)16201 __ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
16202   uint16x4_t __ret;
16203   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16204   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16205   __ret = __rev0 * __rev1;
16206   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16207   return __ret;
16208 }
16209 #endif
16210 
16211 #ifdef __LITTLE_ENDIAN__
vmul_s8(int8x8_t __p0,int8x8_t __p1)16212 __ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
16213   int8x8_t __ret;
16214   __ret = __p0 * __p1;
16215   return __ret;
16216 }
16217 #else
vmul_s8(int8x8_t __p0,int8x8_t __p1)16218 __ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
16219   int8x8_t __ret;
16220   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16221   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16222   __ret = __rev0 * __rev1;
16223   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16224   return __ret;
16225 }
16226 #endif
16227 
16228 #ifdef __LITTLE_ENDIAN__
vmul_f32(float32x2_t __p0,float32x2_t __p1)16229 __ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
16230   float32x2_t __ret;
16231   __ret = __p0 * __p1;
16232   return __ret;
16233 }
16234 #else
vmul_f32(float32x2_t __p0,float32x2_t __p1)16235 __ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
16236   float32x2_t __ret;
16237   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16238   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16239   __ret = __rev0 * __rev1;
16240   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16241   return __ret;
16242 }
16243 #endif
16244 
16245 #ifdef __LITTLE_ENDIAN__
vmul_s32(int32x2_t __p0,int32x2_t __p1)16246 __ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
16247   int32x2_t __ret;
16248   __ret = __p0 * __p1;
16249   return __ret;
16250 }
16251 #else
vmul_s32(int32x2_t __p0,int32x2_t __p1)16252 __ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
16253   int32x2_t __ret;
16254   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16255   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16256   __ret = __rev0 * __rev1;
16257   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16258   return __ret;
16259 }
16260 #endif
16261 
16262 #ifdef __LITTLE_ENDIAN__
vmul_s16(int16x4_t __p0,int16x4_t __p1)16263 __ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
16264   int16x4_t __ret;
16265   __ret = __p0 * __p1;
16266   return __ret;
16267 }
16268 #else
vmul_s16(int16x4_t __p0,int16x4_t __p1)16269 __ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
16270   int16x4_t __ret;
16271   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16272   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16273   __ret = __rev0 * __rev1;
16274   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16275   return __ret;
16276 }
16277 #endif
16278 
16279 #ifdef __LITTLE_ENDIAN__
vmul_p8(poly8x8_t __p0,poly8x8_t __p1)16280 __ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
16281   poly8x8_t __ret;
16282   __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
16283   return __ret;
16284 }
16285 #else
vmul_p8(poly8x8_t __p0,poly8x8_t __p1)16286 __ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
16287   poly8x8_t __ret;
16288   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16289   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16290   __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
16291   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16292   return __ret;
16293 }
16294 #endif
16295 
16296 #ifdef __LITTLE_ENDIAN__
vmulq_p8(poly8x16_t __p0,poly8x16_t __p1)16297 __ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
16298   poly8x16_t __ret;
16299   __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
16300   return __ret;
16301 }
16302 #else
vmulq_p8(poly8x16_t __p0,poly8x16_t __p1)16303 __ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
16304   poly8x16_t __ret;
16305   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16306   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16307   __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
16308   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16309   return __ret;
16310 }
16311 #endif
16312 
16313 #ifdef __LITTLE_ENDIAN__
16314 #define vmulq_lane_u32(__p0_86, __p1_86, __p2_86) __extension__ ({ \
16315   uint32x4_t __ret_86; \
16316   uint32x4_t __s0_86 = __p0_86; \
16317   uint32x2_t __s1_86 = __p1_86; \
16318   __ret_86 = __s0_86 * splatq_lane_u32(__s1_86, __p2_86); \
16319   __ret_86; \
16320 })
16321 #else
16322 #define vmulq_lane_u32(__p0_87, __p1_87, __p2_87) __extension__ ({ \
16323   uint32x4_t __ret_87; \
16324   uint32x4_t __s0_87 = __p0_87; \
16325   uint32x2_t __s1_87 = __p1_87; \
16326   uint32x4_t __rev0_87;  __rev0_87 = __builtin_shufflevector(__s0_87, __s0_87, 3, 2, 1, 0); \
16327   uint32x2_t __rev1_87;  __rev1_87 = __builtin_shufflevector(__s1_87, __s1_87, 1, 0); \
16328   __ret_87 = __rev0_87 * __noswap_splatq_lane_u32(__rev1_87, __p2_87); \
16329   __ret_87 = __builtin_shufflevector(__ret_87, __ret_87, 3, 2, 1, 0); \
16330   __ret_87; \
16331 })
16332 #endif
16333 
16334 #ifdef __LITTLE_ENDIAN__
16335 #define vmulq_lane_u16(__p0_88, __p1_88, __p2_88) __extension__ ({ \
16336   uint16x8_t __ret_88; \
16337   uint16x8_t __s0_88 = __p0_88; \
16338   uint16x4_t __s1_88 = __p1_88; \
16339   __ret_88 = __s0_88 * splatq_lane_u16(__s1_88, __p2_88); \
16340   __ret_88; \
16341 })
16342 #else
16343 #define vmulq_lane_u16(__p0_89, __p1_89, __p2_89) __extension__ ({ \
16344   uint16x8_t __ret_89; \
16345   uint16x8_t __s0_89 = __p0_89; \
16346   uint16x4_t __s1_89 = __p1_89; \
16347   uint16x8_t __rev0_89;  __rev0_89 = __builtin_shufflevector(__s0_89, __s0_89, 7, 6, 5, 4, 3, 2, 1, 0); \
16348   uint16x4_t __rev1_89;  __rev1_89 = __builtin_shufflevector(__s1_89, __s1_89, 3, 2, 1, 0); \
16349   __ret_89 = __rev0_89 * __noswap_splatq_lane_u16(__rev1_89, __p2_89); \
16350   __ret_89 = __builtin_shufflevector(__ret_89, __ret_89, 7, 6, 5, 4, 3, 2, 1, 0); \
16351   __ret_89; \
16352 })
16353 #endif
16354 
16355 #ifdef __LITTLE_ENDIAN__
16356 #define vmulq_lane_f32(__p0_90, __p1_90, __p2_90) __extension__ ({ \
16357   float32x4_t __ret_90; \
16358   float32x4_t __s0_90 = __p0_90; \
16359   float32x2_t __s1_90 = __p1_90; \
16360   __ret_90 = __s0_90 * splatq_lane_f32(__s1_90, __p2_90); \
16361   __ret_90; \
16362 })
16363 #else
16364 #define vmulq_lane_f32(__p0_91, __p1_91, __p2_91) __extension__ ({ \
16365   float32x4_t __ret_91; \
16366   float32x4_t __s0_91 = __p0_91; \
16367   float32x2_t __s1_91 = __p1_91; \
16368   float32x4_t __rev0_91;  __rev0_91 = __builtin_shufflevector(__s0_91, __s0_91, 3, 2, 1, 0); \
16369   float32x2_t __rev1_91;  __rev1_91 = __builtin_shufflevector(__s1_91, __s1_91, 1, 0); \
16370   __ret_91 = __rev0_91 * __noswap_splatq_lane_f32(__rev1_91, __p2_91); \
16371   __ret_91 = __builtin_shufflevector(__ret_91, __ret_91, 3, 2, 1, 0); \
16372   __ret_91; \
16373 })
16374 #endif
16375 
16376 #ifdef __LITTLE_ENDIAN__
16377 #define vmulq_lane_s32(__p0_92, __p1_92, __p2_92) __extension__ ({ \
16378   int32x4_t __ret_92; \
16379   int32x4_t __s0_92 = __p0_92; \
16380   int32x2_t __s1_92 = __p1_92; \
16381   __ret_92 = __s0_92 * splatq_lane_s32(__s1_92, __p2_92); \
16382   __ret_92; \
16383 })
16384 #else
16385 #define vmulq_lane_s32(__p0_93, __p1_93, __p2_93) __extension__ ({ \
16386   int32x4_t __ret_93; \
16387   int32x4_t __s0_93 = __p0_93; \
16388   int32x2_t __s1_93 = __p1_93; \
16389   int32x4_t __rev0_93;  __rev0_93 = __builtin_shufflevector(__s0_93, __s0_93, 3, 2, 1, 0); \
16390   int32x2_t __rev1_93;  __rev1_93 = __builtin_shufflevector(__s1_93, __s1_93, 1, 0); \
16391   __ret_93 = __rev0_93 * __noswap_splatq_lane_s32(__rev1_93, __p2_93); \
16392   __ret_93 = __builtin_shufflevector(__ret_93, __ret_93, 3, 2, 1, 0); \
16393   __ret_93; \
16394 })
16395 #endif
16396 
16397 #ifdef __LITTLE_ENDIAN__
16398 #define vmulq_lane_s16(__p0_94, __p1_94, __p2_94) __extension__ ({ \
16399   int16x8_t __ret_94; \
16400   int16x8_t __s0_94 = __p0_94; \
16401   int16x4_t __s1_94 = __p1_94; \
16402   __ret_94 = __s0_94 * splatq_lane_s16(__s1_94, __p2_94); \
16403   __ret_94; \
16404 })
16405 #else
16406 #define vmulq_lane_s16(__p0_95, __p1_95, __p2_95) __extension__ ({ \
16407   int16x8_t __ret_95; \
16408   int16x8_t __s0_95 = __p0_95; \
16409   int16x4_t __s1_95 = __p1_95; \
16410   int16x8_t __rev0_95;  __rev0_95 = __builtin_shufflevector(__s0_95, __s0_95, 7, 6, 5, 4, 3, 2, 1, 0); \
16411   int16x4_t __rev1_95;  __rev1_95 = __builtin_shufflevector(__s1_95, __s1_95, 3, 2, 1, 0); \
16412   __ret_95 = __rev0_95 * __noswap_splatq_lane_s16(__rev1_95, __p2_95); \
16413   __ret_95 = __builtin_shufflevector(__ret_95, __ret_95, 7, 6, 5, 4, 3, 2, 1, 0); \
16414   __ret_95; \
16415 })
16416 #endif
16417 
16418 #ifdef __LITTLE_ENDIAN__
16419 #define vmul_lane_u32(__p0_96, __p1_96, __p2_96) __extension__ ({ \
16420   uint32x2_t __ret_96; \
16421   uint32x2_t __s0_96 = __p0_96; \
16422   uint32x2_t __s1_96 = __p1_96; \
16423   __ret_96 = __s0_96 * splat_lane_u32(__s1_96, __p2_96); \
16424   __ret_96; \
16425 })
16426 #else
16427 #define vmul_lane_u32(__p0_97, __p1_97, __p2_97) __extension__ ({ \
16428   uint32x2_t __ret_97; \
16429   uint32x2_t __s0_97 = __p0_97; \
16430   uint32x2_t __s1_97 = __p1_97; \
16431   uint32x2_t __rev0_97;  __rev0_97 = __builtin_shufflevector(__s0_97, __s0_97, 1, 0); \
16432   uint32x2_t __rev1_97;  __rev1_97 = __builtin_shufflevector(__s1_97, __s1_97, 1, 0); \
16433   __ret_97 = __rev0_97 * __noswap_splat_lane_u32(__rev1_97, __p2_97); \
16434   __ret_97 = __builtin_shufflevector(__ret_97, __ret_97, 1, 0); \
16435   __ret_97; \
16436 })
16437 #endif
16438 
16439 #ifdef __LITTLE_ENDIAN__
16440 #define vmul_lane_u16(__p0_98, __p1_98, __p2_98) __extension__ ({ \
16441   uint16x4_t __ret_98; \
16442   uint16x4_t __s0_98 = __p0_98; \
16443   uint16x4_t __s1_98 = __p1_98; \
16444   __ret_98 = __s0_98 * splat_lane_u16(__s1_98, __p2_98); \
16445   __ret_98; \
16446 })
16447 #else
16448 #define vmul_lane_u16(__p0_99, __p1_99, __p2_99) __extension__ ({ \
16449   uint16x4_t __ret_99; \
16450   uint16x4_t __s0_99 = __p0_99; \
16451   uint16x4_t __s1_99 = __p1_99; \
16452   uint16x4_t __rev0_99;  __rev0_99 = __builtin_shufflevector(__s0_99, __s0_99, 3, 2, 1, 0); \
16453   uint16x4_t __rev1_99;  __rev1_99 = __builtin_shufflevector(__s1_99, __s1_99, 3, 2, 1, 0); \
16454   __ret_99 = __rev0_99 * __noswap_splat_lane_u16(__rev1_99, __p2_99); \
16455   __ret_99 = __builtin_shufflevector(__ret_99, __ret_99, 3, 2, 1, 0); \
16456   __ret_99; \
16457 })
16458 #endif
16459 
16460 #ifdef __LITTLE_ENDIAN__
16461 #define vmul_lane_f32(__p0_100, __p1_100, __p2_100) __extension__ ({ \
16462   float32x2_t __ret_100; \
16463   float32x2_t __s0_100 = __p0_100; \
16464   float32x2_t __s1_100 = __p1_100; \
16465   __ret_100 = __s0_100 * splat_lane_f32(__s1_100, __p2_100); \
16466   __ret_100; \
16467 })
16468 #else
16469 #define vmul_lane_f32(__p0_101, __p1_101, __p2_101) __extension__ ({ \
16470   float32x2_t __ret_101; \
16471   float32x2_t __s0_101 = __p0_101; \
16472   float32x2_t __s1_101 = __p1_101; \
16473   float32x2_t __rev0_101;  __rev0_101 = __builtin_shufflevector(__s0_101, __s0_101, 1, 0); \
16474   float32x2_t __rev1_101;  __rev1_101 = __builtin_shufflevector(__s1_101, __s1_101, 1, 0); \
16475   __ret_101 = __rev0_101 * __noswap_splat_lane_f32(__rev1_101, __p2_101); \
16476   __ret_101 = __builtin_shufflevector(__ret_101, __ret_101, 1, 0); \
16477   __ret_101; \
16478 })
16479 #endif
16480 
16481 #ifdef __LITTLE_ENDIAN__
16482 #define vmul_lane_s32(__p0_102, __p1_102, __p2_102) __extension__ ({ \
16483   int32x2_t __ret_102; \
16484   int32x2_t __s0_102 = __p0_102; \
16485   int32x2_t __s1_102 = __p1_102; \
16486   __ret_102 = __s0_102 * splat_lane_s32(__s1_102, __p2_102); \
16487   __ret_102; \
16488 })
16489 #else
16490 #define vmul_lane_s32(__p0_103, __p1_103, __p2_103) __extension__ ({ \
16491   int32x2_t __ret_103; \
16492   int32x2_t __s0_103 = __p0_103; \
16493   int32x2_t __s1_103 = __p1_103; \
16494   int32x2_t __rev0_103;  __rev0_103 = __builtin_shufflevector(__s0_103, __s0_103, 1, 0); \
16495   int32x2_t __rev1_103;  __rev1_103 = __builtin_shufflevector(__s1_103, __s1_103, 1, 0); \
16496   __ret_103 = __rev0_103 * __noswap_splat_lane_s32(__rev1_103, __p2_103); \
16497   __ret_103 = __builtin_shufflevector(__ret_103, __ret_103, 1, 0); \
16498   __ret_103; \
16499 })
16500 #endif
16501 
16502 #ifdef __LITTLE_ENDIAN__
16503 #define vmul_lane_s16(__p0_104, __p1_104, __p2_104) __extension__ ({ \
16504   int16x4_t __ret_104; \
16505   int16x4_t __s0_104 = __p0_104; \
16506   int16x4_t __s1_104 = __p1_104; \
16507   __ret_104 = __s0_104 * splat_lane_s16(__s1_104, __p2_104); \
16508   __ret_104; \
16509 })
16510 #else
16511 #define vmul_lane_s16(__p0_105, __p1_105, __p2_105) __extension__ ({ \
16512   int16x4_t __ret_105; \
16513   int16x4_t __s0_105 = __p0_105; \
16514   int16x4_t __s1_105 = __p1_105; \
16515   int16x4_t __rev0_105;  __rev0_105 = __builtin_shufflevector(__s0_105, __s0_105, 3, 2, 1, 0); \
16516   int16x4_t __rev1_105;  __rev1_105 = __builtin_shufflevector(__s1_105, __s1_105, 3, 2, 1, 0); \
16517   __ret_105 = __rev0_105 * __noswap_splat_lane_s16(__rev1_105, __p2_105); \
16518   __ret_105 = __builtin_shufflevector(__ret_105, __ret_105, 3, 2, 1, 0); \
16519   __ret_105; \
16520 })
16521 #endif
16522 
16523 #ifdef __LITTLE_ENDIAN__
vmulq_n_u32(uint32x4_t __p0,uint32_t __p1)16524 __ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
16525   uint32x4_t __ret;
16526   __ret = __p0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
16527   return __ret;
16528 }
16529 #else
vmulq_n_u32(uint32x4_t __p0,uint32_t __p1)16530 __ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
16531   uint32x4_t __ret;
16532   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16533   __ret = __rev0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
16534   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16535   return __ret;
16536 }
16537 #endif
16538 
16539 #ifdef __LITTLE_ENDIAN__
vmulq_n_u16(uint16x8_t __p0,uint16_t __p1)16540 __ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
16541   uint16x8_t __ret;
16542   __ret = __p0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
16543   return __ret;
16544 }
16545 #else
vmulq_n_u16(uint16x8_t __p0,uint16_t __p1)16546 __ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
16547   uint16x8_t __ret;
16548   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16549   __ret = __rev0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
16550   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16551   return __ret;
16552 }
16553 #endif
16554 
16555 #ifdef __LITTLE_ENDIAN__
vmulq_n_f32(float32x4_t __p0,float32_t __p1)16556 __ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
16557   float32x4_t __ret;
16558   __ret = __p0 * (float32x4_t) {__p1, __p1, __p1, __p1};
16559   return __ret;
16560 }
16561 #else
vmulq_n_f32(float32x4_t __p0,float32_t __p1)16562 __ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
16563   float32x4_t __ret;
16564   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16565   __ret = __rev0 * (float32x4_t) {__p1, __p1, __p1, __p1};
16566   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16567   return __ret;
16568 }
16569 #endif
16570 
16571 #ifdef __LITTLE_ENDIAN__
vmulq_n_s32(int32x4_t __p0,int32_t __p1)16572 __ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
16573   int32x4_t __ret;
16574   __ret = __p0 * (int32x4_t) {__p1, __p1, __p1, __p1};
16575   return __ret;
16576 }
16577 #else
vmulq_n_s32(int32x4_t __p0,int32_t __p1)16578 __ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
16579   int32x4_t __ret;
16580   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16581   __ret = __rev0 * (int32x4_t) {__p1, __p1, __p1, __p1};
16582   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16583   return __ret;
16584 }
16585 #endif
16586 
16587 #ifdef __LITTLE_ENDIAN__
vmulq_n_s16(int16x8_t __p0,int16_t __p1)16588 __ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
16589   int16x8_t __ret;
16590   __ret = __p0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
16591   return __ret;
16592 }
16593 #else
vmulq_n_s16(int16x8_t __p0,int16_t __p1)16594 __ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
16595   int16x8_t __ret;
16596   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16597   __ret = __rev0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
16598   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16599   return __ret;
16600 }
16601 #endif
16602 
16603 #ifdef __LITTLE_ENDIAN__
vmul_n_u32(uint32x2_t __p0,uint32_t __p1)16604 __ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
16605   uint32x2_t __ret;
16606   __ret = __p0 * (uint32x2_t) {__p1, __p1};
16607   return __ret;
16608 }
16609 #else
vmul_n_u32(uint32x2_t __p0,uint32_t __p1)16610 __ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
16611   uint32x2_t __ret;
16612   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16613   __ret = __rev0 * (uint32x2_t) {__p1, __p1};
16614   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16615   return __ret;
16616 }
16617 #endif
16618 
16619 #ifdef __LITTLE_ENDIAN__
vmul_n_u16(uint16x4_t __p0,uint16_t __p1)16620 __ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
16621   uint16x4_t __ret;
16622   __ret = __p0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
16623   return __ret;
16624 }
16625 #else
vmul_n_u16(uint16x4_t __p0,uint16_t __p1)16626 __ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
16627   uint16x4_t __ret;
16628   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16629   __ret = __rev0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
16630   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16631   return __ret;
16632 }
16633 #endif
16634 
16635 #ifdef __LITTLE_ENDIAN__
vmul_n_f32(float32x2_t __p0,float32_t __p1)16636 __ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
16637   float32x2_t __ret;
16638   __ret = __p0 * (float32x2_t) {__p1, __p1};
16639   return __ret;
16640 }
16641 #else
vmul_n_f32(float32x2_t __p0,float32_t __p1)16642 __ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
16643   float32x2_t __ret;
16644   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16645   __ret = __rev0 * (float32x2_t) {__p1, __p1};
16646   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16647   return __ret;
16648 }
16649 #endif
16650 
16651 #ifdef __LITTLE_ENDIAN__
vmul_n_s32(int32x2_t __p0,int32_t __p1)16652 __ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
16653   int32x2_t __ret;
16654   __ret = __p0 * (int32x2_t) {__p1, __p1};
16655   return __ret;
16656 }
16657 #else
vmul_n_s32(int32x2_t __p0,int32_t __p1)16658 __ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
16659   int32x2_t __ret;
16660   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16661   __ret = __rev0 * (int32x2_t) {__p1, __p1};
16662   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16663   return __ret;
16664 }
16665 #endif
16666 
16667 #ifdef __LITTLE_ENDIAN__
vmul_n_s16(int16x4_t __p0,int16_t __p1)16668 __ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
16669   int16x4_t __ret;
16670   __ret = __p0 * (int16x4_t) {__p1, __p1, __p1, __p1};
16671   return __ret;
16672 }
16673 #else
vmul_n_s16(int16x4_t __p0,int16_t __p1)16674 __ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
16675   int16x4_t __ret;
16676   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16677   __ret = __rev0 * (int16x4_t) {__p1, __p1, __p1, __p1};
16678   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16679   return __ret;
16680 }
16681 #endif
16682 
16683 #ifdef __LITTLE_ENDIAN__
vmull_p8(poly8x8_t __p0,poly8x8_t __p1)16684 __ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
16685   poly16x8_t __ret;
16686   __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
16687   return __ret;
16688 }
16689 #else
vmull_p8(poly8x8_t __p0,poly8x8_t __p1)16690 __ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
16691   poly16x8_t __ret;
16692   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16693   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16694   __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 37);
16695   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16696   return __ret;
16697 }
__noswap_vmull_p8(poly8x8_t __p0,poly8x8_t __p1)16698 __ai poly16x8_t __noswap_vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
16699   poly16x8_t __ret;
16700   __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
16701   return __ret;
16702 }
16703 #endif
16704 
16705 #ifdef __LITTLE_ENDIAN__
vmull_u8(uint8x8_t __p0,uint8x8_t __p1)16706 __ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
16707   uint16x8_t __ret;
16708   __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
16709   return __ret;
16710 }
16711 #else
vmull_u8(uint8x8_t __p0,uint8x8_t __p1)16712 __ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
16713   uint16x8_t __ret;
16714   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16715   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16716   __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 49);
16717   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16718   return __ret;
16719 }
__noswap_vmull_u8(uint8x8_t __p0,uint8x8_t __p1)16720 __ai uint16x8_t __noswap_vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
16721   uint16x8_t __ret;
16722   __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
16723   return __ret;
16724 }
16725 #endif
16726 
16727 #ifdef __LITTLE_ENDIAN__
vmull_u32(uint32x2_t __p0,uint32x2_t __p1)16728 __ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
16729   uint64x2_t __ret;
16730   __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
16731   return __ret;
16732 }
16733 #else
vmull_u32(uint32x2_t __p0,uint32x2_t __p1)16734 __ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
16735   uint64x2_t __ret;
16736   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16737   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16738   __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 51);
16739   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16740   return __ret;
16741 }
__noswap_vmull_u32(uint32x2_t __p0,uint32x2_t __p1)16742 __ai uint64x2_t __noswap_vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
16743   uint64x2_t __ret;
16744   __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
16745   return __ret;
16746 }
16747 #endif
16748 
16749 #ifdef __LITTLE_ENDIAN__
vmull_u16(uint16x4_t __p0,uint16x4_t __p1)16750 __ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
16751   uint32x4_t __ret;
16752   __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
16753   return __ret;
16754 }
16755 #else
vmull_u16(uint16x4_t __p0,uint16x4_t __p1)16756 __ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
16757   uint32x4_t __ret;
16758   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16759   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16760   __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 50);
16761   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16762   return __ret;
16763 }
__noswap_vmull_u16(uint16x4_t __p0,uint16x4_t __p1)16764 __ai uint32x4_t __noswap_vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
16765   uint32x4_t __ret;
16766   __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
16767   return __ret;
16768 }
16769 #endif
16770 
16771 #ifdef __LITTLE_ENDIAN__
vmull_s8(int8x8_t __p0,int8x8_t __p1)16772 __ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
16773   int16x8_t __ret;
16774   __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
16775   return __ret;
16776 }
16777 #else
vmull_s8(int8x8_t __p0,int8x8_t __p1)16778 __ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
16779   int16x8_t __ret;
16780   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16781   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16782   __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 33);
16783   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16784   return __ret;
16785 }
__noswap_vmull_s8(int8x8_t __p0,int8x8_t __p1)16786 __ai int16x8_t __noswap_vmull_s8(int8x8_t __p0, int8x8_t __p1) {
16787   int16x8_t __ret;
16788   __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
16789   return __ret;
16790 }
16791 #endif
16792 
16793 #ifdef __LITTLE_ENDIAN__
vmull_s32(int32x2_t __p0,int32x2_t __p1)16794 __ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
16795   int64x2_t __ret;
16796   __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
16797   return __ret;
16798 }
16799 #else
vmull_s32(int32x2_t __p0,int32x2_t __p1)16800 __ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
16801   int64x2_t __ret;
16802   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16803   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16804   __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
16805   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16806   return __ret;
16807 }
__noswap_vmull_s32(int32x2_t __p0,int32x2_t __p1)16808 __ai int64x2_t __noswap_vmull_s32(int32x2_t __p0, int32x2_t __p1) {
16809   int64x2_t __ret;
16810   __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
16811   return __ret;
16812 }
16813 #endif
16814 
16815 #ifdef __LITTLE_ENDIAN__
vmull_s16(int16x4_t __p0,int16x4_t __p1)16816 __ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
16817   int32x4_t __ret;
16818   __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
16819   return __ret;
16820 }
16821 #else
vmull_s16(int16x4_t __p0,int16x4_t __p1)16822 __ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
16823   int32x4_t __ret;
16824   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16825   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16826   __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
16827   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16828   return __ret;
16829 }
__noswap_vmull_s16(int16x4_t __p0,int16x4_t __p1)16830 __ai int32x4_t __noswap_vmull_s16(int16x4_t __p0, int16x4_t __p1) {
16831   int32x4_t __ret;
16832   __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
16833   return __ret;
16834 }
16835 #endif
16836 
16837 #ifdef __LITTLE_ENDIAN__
16838 #define vmull_lane_u32(__p0_106, __p1_106, __p2_106) __extension__ ({ \
16839   uint64x2_t __ret_106; \
16840   uint32x2_t __s0_106 = __p0_106; \
16841   uint32x2_t __s1_106 = __p1_106; \
16842   __ret_106 = vmull_u32(__s0_106, splat_lane_u32(__s1_106, __p2_106)); \
16843   __ret_106; \
16844 })
16845 #else
16846 #define vmull_lane_u32(__p0_107, __p1_107, __p2_107) __extension__ ({ \
16847   uint64x2_t __ret_107; \
16848   uint32x2_t __s0_107 = __p0_107; \
16849   uint32x2_t __s1_107 = __p1_107; \
16850   uint32x2_t __rev0_107;  __rev0_107 = __builtin_shufflevector(__s0_107, __s0_107, 1, 0); \
16851   uint32x2_t __rev1_107;  __rev1_107 = __builtin_shufflevector(__s1_107, __s1_107, 1, 0); \
16852   __ret_107 = __noswap_vmull_u32(__rev0_107, __noswap_splat_lane_u32(__rev1_107, __p2_107)); \
16853   __ret_107 = __builtin_shufflevector(__ret_107, __ret_107, 1, 0); \
16854   __ret_107; \
16855 })
16856 #endif
16857 
16858 #ifdef __LITTLE_ENDIAN__
16859 #define vmull_lane_u16(__p0_108, __p1_108, __p2_108) __extension__ ({ \
16860   uint32x4_t __ret_108; \
16861   uint16x4_t __s0_108 = __p0_108; \
16862   uint16x4_t __s1_108 = __p1_108; \
16863   __ret_108 = vmull_u16(__s0_108, splat_lane_u16(__s1_108, __p2_108)); \
16864   __ret_108; \
16865 })
16866 #else
16867 #define vmull_lane_u16(__p0_109, __p1_109, __p2_109) __extension__ ({ \
16868   uint32x4_t __ret_109; \
16869   uint16x4_t __s0_109 = __p0_109; \
16870   uint16x4_t __s1_109 = __p1_109; \
16871   uint16x4_t __rev0_109;  __rev0_109 = __builtin_shufflevector(__s0_109, __s0_109, 3, 2, 1, 0); \
16872   uint16x4_t __rev1_109;  __rev1_109 = __builtin_shufflevector(__s1_109, __s1_109, 3, 2, 1, 0); \
16873   __ret_109 = __noswap_vmull_u16(__rev0_109, __noswap_splat_lane_u16(__rev1_109, __p2_109)); \
16874   __ret_109 = __builtin_shufflevector(__ret_109, __ret_109, 3, 2, 1, 0); \
16875   __ret_109; \
16876 })
16877 #endif
16878 
16879 #ifdef __LITTLE_ENDIAN__
16880 #define vmull_lane_s32(__p0_110, __p1_110, __p2_110) __extension__ ({ \
16881   int64x2_t __ret_110; \
16882   int32x2_t __s0_110 = __p0_110; \
16883   int32x2_t __s1_110 = __p1_110; \
16884   __ret_110 = vmull_s32(__s0_110, splat_lane_s32(__s1_110, __p2_110)); \
16885   __ret_110; \
16886 })
16887 #else
16888 #define vmull_lane_s32(__p0_111, __p1_111, __p2_111) __extension__ ({ \
16889   int64x2_t __ret_111; \
16890   int32x2_t __s0_111 = __p0_111; \
16891   int32x2_t __s1_111 = __p1_111; \
16892   int32x2_t __rev0_111;  __rev0_111 = __builtin_shufflevector(__s0_111, __s0_111, 1, 0); \
16893   int32x2_t __rev1_111;  __rev1_111 = __builtin_shufflevector(__s1_111, __s1_111, 1, 0); \
16894   __ret_111 = __noswap_vmull_s32(__rev0_111, __noswap_splat_lane_s32(__rev1_111, __p2_111)); \
16895   __ret_111 = __builtin_shufflevector(__ret_111, __ret_111, 1, 0); \
16896   __ret_111; \
16897 })
16898 #endif
16899 
16900 #ifdef __LITTLE_ENDIAN__
16901 #define vmull_lane_s16(__p0_112, __p1_112, __p2_112) __extension__ ({ \
16902   int32x4_t __ret_112; \
16903   int16x4_t __s0_112 = __p0_112; \
16904   int16x4_t __s1_112 = __p1_112; \
16905   __ret_112 = vmull_s16(__s0_112, splat_lane_s16(__s1_112, __p2_112)); \
16906   __ret_112; \
16907 })
16908 #else
16909 #define vmull_lane_s16(__p0_113, __p1_113, __p2_113) __extension__ ({ \
16910   int32x4_t __ret_113; \
16911   int16x4_t __s0_113 = __p0_113; \
16912   int16x4_t __s1_113 = __p1_113; \
16913   int16x4_t __rev0_113;  __rev0_113 = __builtin_shufflevector(__s0_113, __s0_113, 3, 2, 1, 0); \
16914   int16x4_t __rev1_113;  __rev1_113 = __builtin_shufflevector(__s1_113, __s1_113, 3, 2, 1, 0); \
16915   __ret_113 = __noswap_vmull_s16(__rev0_113, __noswap_splat_lane_s16(__rev1_113, __p2_113)); \
16916   __ret_113 = __builtin_shufflevector(__ret_113, __ret_113, 3, 2, 1, 0); \
16917   __ret_113; \
16918 })
16919 #endif
16920 
16921 #ifdef __LITTLE_ENDIAN__
vmull_n_u32(uint32x2_t __p0,uint32_t __p1)16922 __ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
16923   uint64x2_t __ret;
16924   __ret = vmull_u32(__p0, (uint32x2_t) {__p1, __p1});
16925   return __ret;
16926 }
16927 #else
vmull_n_u32(uint32x2_t __p0,uint32_t __p1)16928 __ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
16929   uint64x2_t __ret;
16930   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16931   __ret = __noswap_vmull_u32(__rev0, (uint32x2_t) {__p1, __p1});
16932   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16933   return __ret;
16934 }
__noswap_vmull_n_u32(uint32x2_t __p0,uint32_t __p1)16935 __ai uint64x2_t __noswap_vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
16936   uint64x2_t __ret;
16937   __ret = __noswap_vmull_u32(__p0, (uint32x2_t) {__p1, __p1});
16938   return __ret;
16939 }
16940 #endif
16941 
16942 #ifdef __LITTLE_ENDIAN__
vmull_n_u16(uint16x4_t __p0,uint16_t __p1)16943 __ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
16944   uint32x4_t __ret;
16945   __ret = vmull_u16(__p0, (uint16x4_t) {__p1, __p1, __p1, __p1});
16946   return __ret;
16947 }
16948 #else
vmull_n_u16(uint16x4_t __p0,uint16_t __p1)16949 __ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
16950   uint32x4_t __ret;
16951   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16952   __ret = __noswap_vmull_u16(__rev0, (uint16x4_t) {__p1, __p1, __p1, __p1});
16953   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16954   return __ret;
16955 }
__noswap_vmull_n_u16(uint16x4_t __p0,uint16_t __p1)16956 __ai uint32x4_t __noswap_vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
16957   uint32x4_t __ret;
16958   __ret = __noswap_vmull_u16(__p0, (uint16x4_t) {__p1, __p1, __p1, __p1});
16959   return __ret;
16960 }
16961 #endif
16962 
16963 #ifdef __LITTLE_ENDIAN__
vmull_n_s32(int32x2_t __p0,int32_t __p1)16964 __ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
16965   int64x2_t __ret;
16966   __ret = vmull_s32(__p0, (int32x2_t) {__p1, __p1});
16967   return __ret;
16968 }
16969 #else
vmull_n_s32(int32x2_t __p0,int32_t __p1)16970 __ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
16971   int64x2_t __ret;
16972   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16973   __ret = __noswap_vmull_s32(__rev0, (int32x2_t) {__p1, __p1});
16974   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16975   return __ret;
16976 }
__noswap_vmull_n_s32(int32x2_t __p0,int32_t __p1)16977 __ai int64x2_t __noswap_vmull_n_s32(int32x2_t __p0, int32_t __p1) {
16978   int64x2_t __ret;
16979   __ret = __noswap_vmull_s32(__p0, (int32x2_t) {__p1, __p1});
16980   return __ret;
16981 }
16982 #endif
16983 
16984 #ifdef __LITTLE_ENDIAN__
vmull_n_s16(int16x4_t __p0,int16_t __p1)16985 __ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
16986   int32x4_t __ret;
16987   __ret = vmull_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
16988   return __ret;
16989 }
16990 #else
vmull_n_s16(int16x4_t __p0,int16_t __p1)16991 __ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
16992   int32x4_t __ret;
16993   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16994   __ret = __noswap_vmull_s16(__rev0, (int16x4_t) {__p1, __p1, __p1, __p1});
16995   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16996   return __ret;
16997 }
__noswap_vmull_n_s16(int16x4_t __p0,int16_t __p1)16998 __ai int32x4_t __noswap_vmull_n_s16(int16x4_t __p0, int16_t __p1) {
16999   int32x4_t __ret;
17000   __ret = __noswap_vmull_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
17001   return __ret;
17002 }
17003 #endif
17004 
17005 #ifdef __LITTLE_ENDIAN__
vmvn_p8(poly8x8_t __p0)17006 __ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
17007   poly8x8_t __ret;
17008   __ret = ~__p0;
17009   return __ret;
17010 }
17011 #else
vmvn_p8(poly8x8_t __p0)17012 __ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
17013   poly8x8_t __ret;
17014   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17015   __ret = ~__rev0;
17016   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17017   return __ret;
17018 }
17019 #endif
17020 
17021 #ifdef __LITTLE_ENDIAN__
vmvnq_p8(poly8x16_t __p0)17022 __ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
17023   poly8x16_t __ret;
17024   __ret = ~__p0;
17025   return __ret;
17026 }
17027 #else
vmvnq_p8(poly8x16_t __p0)17028 __ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
17029   poly8x16_t __ret;
17030   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17031   __ret = ~__rev0;
17032   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17033   return __ret;
17034 }
17035 #endif
17036 
17037 #ifdef __LITTLE_ENDIAN__
vmvnq_u8(uint8x16_t __p0)17038 __ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
17039   uint8x16_t __ret;
17040   __ret = ~__p0;
17041   return __ret;
17042 }
17043 #else
vmvnq_u8(uint8x16_t __p0)17044 __ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
17045   uint8x16_t __ret;
17046   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17047   __ret = ~__rev0;
17048   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17049   return __ret;
17050 }
17051 #endif
17052 
17053 #ifdef __LITTLE_ENDIAN__
vmvnq_u32(uint32x4_t __p0)17054 __ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
17055   uint32x4_t __ret;
17056   __ret = ~__p0;
17057   return __ret;
17058 }
17059 #else
vmvnq_u32(uint32x4_t __p0)17060 __ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
17061   uint32x4_t __ret;
17062   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17063   __ret = ~__rev0;
17064   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17065   return __ret;
17066 }
17067 #endif
17068 
17069 #ifdef __LITTLE_ENDIAN__
vmvnq_u16(uint16x8_t __p0)17070 __ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
17071   uint16x8_t __ret;
17072   __ret = ~__p0;
17073   return __ret;
17074 }
17075 #else
vmvnq_u16(uint16x8_t __p0)17076 __ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
17077   uint16x8_t __ret;
17078   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17079   __ret = ~__rev0;
17080   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17081   return __ret;
17082 }
17083 #endif
17084 
17085 #ifdef __LITTLE_ENDIAN__
vmvnq_s8(int8x16_t __p0)17086 __ai int8x16_t vmvnq_s8(int8x16_t __p0) {
17087   int8x16_t __ret;
17088   __ret = ~__p0;
17089   return __ret;
17090 }
17091 #else
vmvnq_s8(int8x16_t __p0)17092 __ai int8x16_t vmvnq_s8(int8x16_t __p0) {
17093   int8x16_t __ret;
17094   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17095   __ret = ~__rev0;
17096   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17097   return __ret;
17098 }
17099 #endif
17100 
17101 #ifdef __LITTLE_ENDIAN__
vmvnq_s32(int32x4_t __p0)17102 __ai int32x4_t vmvnq_s32(int32x4_t __p0) {
17103   int32x4_t __ret;
17104   __ret = ~__p0;
17105   return __ret;
17106 }
17107 #else
vmvnq_s32(int32x4_t __p0)17108 __ai int32x4_t vmvnq_s32(int32x4_t __p0) {
17109   int32x4_t __ret;
17110   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17111   __ret = ~__rev0;
17112   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17113   return __ret;
17114 }
17115 #endif
17116 
17117 #ifdef __LITTLE_ENDIAN__
vmvnq_s16(int16x8_t __p0)17118 __ai int16x8_t vmvnq_s16(int16x8_t __p0) {
17119   int16x8_t __ret;
17120   __ret = ~__p0;
17121   return __ret;
17122 }
17123 #else
vmvnq_s16(int16x8_t __p0)17124 __ai int16x8_t vmvnq_s16(int16x8_t __p0) {
17125   int16x8_t __ret;
17126   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17127   __ret = ~__rev0;
17128   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17129   return __ret;
17130 }
17131 #endif
17132 
17133 #ifdef __LITTLE_ENDIAN__
vmvn_u8(uint8x8_t __p0)17134 __ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
17135   uint8x8_t __ret;
17136   __ret = ~__p0;
17137   return __ret;
17138 }
17139 #else
vmvn_u8(uint8x8_t __p0)17140 __ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
17141   uint8x8_t __ret;
17142   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17143   __ret = ~__rev0;
17144   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17145   return __ret;
17146 }
17147 #endif
17148 
17149 #ifdef __LITTLE_ENDIAN__
vmvn_u32(uint32x2_t __p0)17150 __ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
17151   uint32x2_t __ret;
17152   __ret = ~__p0;
17153   return __ret;
17154 }
17155 #else
vmvn_u32(uint32x2_t __p0)17156 __ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
17157   uint32x2_t __ret;
17158   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17159   __ret = ~__rev0;
17160   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17161   return __ret;
17162 }
17163 #endif
17164 
17165 #ifdef __LITTLE_ENDIAN__
vmvn_u16(uint16x4_t __p0)17166 __ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
17167   uint16x4_t __ret;
17168   __ret = ~__p0;
17169   return __ret;
17170 }
17171 #else
vmvn_u16(uint16x4_t __p0)17172 __ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
17173   uint16x4_t __ret;
17174   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17175   __ret = ~__rev0;
17176   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17177   return __ret;
17178 }
17179 #endif
17180 
17181 #ifdef __LITTLE_ENDIAN__
vmvn_s8(int8x8_t __p0)17182 __ai int8x8_t vmvn_s8(int8x8_t __p0) {
17183   int8x8_t __ret;
17184   __ret = ~__p0;
17185   return __ret;
17186 }
17187 #else
vmvn_s8(int8x8_t __p0)17188 __ai int8x8_t vmvn_s8(int8x8_t __p0) {
17189   int8x8_t __ret;
17190   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17191   __ret = ~__rev0;
17192   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17193   return __ret;
17194 }
17195 #endif
17196 
17197 #ifdef __LITTLE_ENDIAN__
vmvn_s32(int32x2_t __p0)17198 __ai int32x2_t vmvn_s32(int32x2_t __p0) {
17199   int32x2_t __ret;
17200   __ret = ~__p0;
17201   return __ret;
17202 }
17203 #else
vmvn_s32(int32x2_t __p0)17204 __ai int32x2_t vmvn_s32(int32x2_t __p0) {
17205   int32x2_t __ret;
17206   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17207   __ret = ~__rev0;
17208   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17209   return __ret;
17210 }
17211 #endif
17212 
17213 #ifdef __LITTLE_ENDIAN__
vmvn_s16(int16x4_t __p0)17214 __ai int16x4_t vmvn_s16(int16x4_t __p0) {
17215   int16x4_t __ret;
17216   __ret = ~__p0;
17217   return __ret;
17218 }
17219 #else
vmvn_s16(int16x4_t __p0)17220 __ai int16x4_t vmvn_s16(int16x4_t __p0) {
17221   int16x4_t __ret;
17222   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17223   __ret = ~__rev0;
17224   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17225   return __ret;
17226 }
17227 #endif
17228 
17229 #ifdef __LITTLE_ENDIAN__
vnegq_s8(int8x16_t __p0)17230 __ai int8x16_t vnegq_s8(int8x16_t __p0) {
17231   int8x16_t __ret;
17232   __ret = -__p0;
17233   return __ret;
17234 }
17235 #else
vnegq_s8(int8x16_t __p0)17236 __ai int8x16_t vnegq_s8(int8x16_t __p0) {
17237   int8x16_t __ret;
17238   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17239   __ret = -__rev0;
17240   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17241   return __ret;
17242 }
17243 #endif
17244 
17245 #ifdef __LITTLE_ENDIAN__
vnegq_f32(float32x4_t __p0)17246 __ai float32x4_t vnegq_f32(float32x4_t __p0) {
17247   float32x4_t __ret;
17248   __ret = -__p0;
17249   return __ret;
17250 }
17251 #else
vnegq_f32(float32x4_t __p0)17252 __ai float32x4_t vnegq_f32(float32x4_t __p0) {
17253   float32x4_t __ret;
17254   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17255   __ret = -__rev0;
17256   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17257   return __ret;
17258 }
17259 #endif
17260 
17261 #ifdef __LITTLE_ENDIAN__
vnegq_s32(int32x4_t __p0)17262 __ai int32x4_t vnegq_s32(int32x4_t __p0) {
17263   int32x4_t __ret;
17264   __ret = -__p0;
17265   return __ret;
17266 }
17267 #else
vnegq_s32(int32x4_t __p0)17268 __ai int32x4_t vnegq_s32(int32x4_t __p0) {
17269   int32x4_t __ret;
17270   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17271   __ret = -__rev0;
17272   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17273   return __ret;
17274 }
17275 #endif
17276 
17277 #ifdef __LITTLE_ENDIAN__
vnegq_s16(int16x8_t __p0)17278 __ai int16x8_t vnegq_s16(int16x8_t __p0) {
17279   int16x8_t __ret;
17280   __ret = -__p0;
17281   return __ret;
17282 }
17283 #else
vnegq_s16(int16x8_t __p0)17284 __ai int16x8_t vnegq_s16(int16x8_t __p0) {
17285   int16x8_t __ret;
17286   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17287   __ret = -__rev0;
17288   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17289   return __ret;
17290 }
17291 #endif
17292 
17293 #ifdef __LITTLE_ENDIAN__
vneg_s8(int8x8_t __p0)17294 __ai int8x8_t vneg_s8(int8x8_t __p0) {
17295   int8x8_t __ret;
17296   __ret = -__p0;
17297   return __ret;
17298 }
17299 #else
vneg_s8(int8x8_t __p0)17300 __ai int8x8_t vneg_s8(int8x8_t __p0) {
17301   int8x8_t __ret;
17302   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17303   __ret = -__rev0;
17304   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17305   return __ret;
17306 }
17307 #endif
17308 
17309 #ifdef __LITTLE_ENDIAN__
vneg_f32(float32x2_t __p0)17310 __ai float32x2_t vneg_f32(float32x2_t __p0) {
17311   float32x2_t __ret;
17312   __ret = -__p0;
17313   return __ret;
17314 }
17315 #else
vneg_f32(float32x2_t __p0)17316 __ai float32x2_t vneg_f32(float32x2_t __p0) {
17317   float32x2_t __ret;
17318   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17319   __ret = -__rev0;
17320   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17321   return __ret;
17322 }
17323 #endif
17324 
17325 #ifdef __LITTLE_ENDIAN__
vneg_s32(int32x2_t __p0)17326 __ai int32x2_t vneg_s32(int32x2_t __p0) {
17327   int32x2_t __ret;
17328   __ret = -__p0;
17329   return __ret;
17330 }
17331 #else
vneg_s32(int32x2_t __p0)17332 __ai int32x2_t vneg_s32(int32x2_t __p0) {
17333   int32x2_t __ret;
17334   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17335   __ret = -__rev0;
17336   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17337   return __ret;
17338 }
17339 #endif
17340 
17341 #ifdef __LITTLE_ENDIAN__
vneg_s16(int16x4_t __p0)17342 __ai int16x4_t vneg_s16(int16x4_t __p0) {
17343   int16x4_t __ret;
17344   __ret = -__p0;
17345   return __ret;
17346 }
17347 #else
vneg_s16(int16x4_t __p0)17348 __ai int16x4_t vneg_s16(int16x4_t __p0) {
17349   int16x4_t __ret;
17350   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17351   __ret = -__rev0;
17352   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17353   return __ret;
17354 }
17355 #endif
17356 
17357 #ifdef __LITTLE_ENDIAN__
vornq_u8(uint8x16_t __p0,uint8x16_t __p1)17358 __ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
17359   uint8x16_t __ret;
17360   __ret = __p0 | ~__p1;
17361   return __ret;
17362 }
17363 #else
vornq_u8(uint8x16_t __p0,uint8x16_t __p1)17364 __ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
17365   uint8x16_t __ret;
17366   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17367   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17368   __ret = __rev0 | ~__rev1;
17369   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17370   return __ret;
17371 }
17372 #endif
17373 
17374 #ifdef __LITTLE_ENDIAN__
vornq_u32(uint32x4_t __p0,uint32x4_t __p1)17375 __ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
17376   uint32x4_t __ret;
17377   __ret = __p0 | ~__p1;
17378   return __ret;
17379 }
17380 #else
vornq_u32(uint32x4_t __p0,uint32x4_t __p1)17381 __ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
17382   uint32x4_t __ret;
17383   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17384   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17385   __ret = __rev0 | ~__rev1;
17386   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17387   return __ret;
17388 }
17389 #endif
17390 
17391 #ifdef __LITTLE_ENDIAN__
vornq_u64(uint64x2_t __p0,uint64x2_t __p1)17392 __ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
17393   uint64x2_t __ret;
17394   __ret = __p0 | ~__p1;
17395   return __ret;
17396 }
17397 #else
vornq_u64(uint64x2_t __p0,uint64x2_t __p1)17398 __ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
17399   uint64x2_t __ret;
17400   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17401   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17402   __ret = __rev0 | ~__rev1;
17403   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17404   return __ret;
17405 }
17406 #endif
17407 
17408 #ifdef __LITTLE_ENDIAN__
vornq_u16(uint16x8_t __p0,uint16x8_t __p1)17409 __ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
17410   uint16x8_t __ret;
17411   __ret = __p0 | ~__p1;
17412   return __ret;
17413 }
17414 #else
vornq_u16(uint16x8_t __p0,uint16x8_t __p1)17415 __ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
17416   uint16x8_t __ret;
17417   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17418   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17419   __ret = __rev0 | ~__rev1;
17420   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17421   return __ret;
17422 }
17423 #endif
17424 
17425 #ifdef __LITTLE_ENDIAN__
vornq_s8(int8x16_t __p0,int8x16_t __p1)17426 __ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
17427   int8x16_t __ret;
17428   __ret = __p0 | ~__p1;
17429   return __ret;
17430 }
17431 #else
vornq_s8(int8x16_t __p0,int8x16_t __p1)17432 __ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
17433   int8x16_t __ret;
17434   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17435   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17436   __ret = __rev0 | ~__rev1;
17437   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17438   return __ret;
17439 }
17440 #endif
17441 
17442 #ifdef __LITTLE_ENDIAN__
vornq_s32(int32x4_t __p0,int32x4_t __p1)17443 __ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
17444   int32x4_t __ret;
17445   __ret = __p0 | ~__p1;
17446   return __ret;
17447 }
17448 #else
vornq_s32(int32x4_t __p0,int32x4_t __p1)17449 __ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
17450   int32x4_t __ret;
17451   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17452   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17453   __ret = __rev0 | ~__rev1;
17454   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17455   return __ret;
17456 }
17457 #endif
17458 
17459 #ifdef __LITTLE_ENDIAN__
vornq_s64(int64x2_t __p0,int64x2_t __p1)17460 __ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
17461   int64x2_t __ret;
17462   __ret = __p0 | ~__p1;
17463   return __ret;
17464 }
17465 #else
vornq_s64(int64x2_t __p0,int64x2_t __p1)17466 __ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
17467   int64x2_t __ret;
17468   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17469   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17470   __ret = __rev0 | ~__rev1;
17471   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17472   return __ret;
17473 }
17474 #endif
17475 
17476 #ifdef __LITTLE_ENDIAN__
vornq_s16(int16x8_t __p0,int16x8_t __p1)17477 __ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
17478   int16x8_t __ret;
17479   __ret = __p0 | ~__p1;
17480   return __ret;
17481 }
17482 #else
vornq_s16(int16x8_t __p0,int16x8_t __p1)17483 __ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
17484   int16x8_t __ret;
17485   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17486   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17487   __ret = __rev0 | ~__rev1;
17488   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17489   return __ret;
17490 }
17491 #endif
17492 
17493 #ifdef __LITTLE_ENDIAN__
vorn_u8(uint8x8_t __p0,uint8x8_t __p1)17494 __ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
17495   uint8x8_t __ret;
17496   __ret = __p0 | ~__p1;
17497   return __ret;
17498 }
17499 #else
vorn_u8(uint8x8_t __p0,uint8x8_t __p1)17500 __ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
17501   uint8x8_t __ret;
17502   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17503   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17504   __ret = __rev0 | ~__rev1;
17505   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17506   return __ret;
17507 }
17508 #endif
17509 
17510 #ifdef __LITTLE_ENDIAN__
vorn_u32(uint32x2_t __p0,uint32x2_t __p1)17511 __ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
17512   uint32x2_t __ret;
17513   __ret = __p0 | ~__p1;
17514   return __ret;
17515 }
17516 #else
vorn_u32(uint32x2_t __p0,uint32x2_t __p1)17517 __ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
17518   uint32x2_t __ret;
17519   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17520   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17521   __ret = __rev0 | ~__rev1;
17522   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17523   return __ret;
17524 }
17525 #endif
17526 
vorn_u64(uint64x1_t __p0,uint64x1_t __p1)17527 __ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
17528   uint64x1_t __ret;
17529   __ret = __p0 | ~__p1;
17530   return __ret;
17531 }
17532 #ifdef __LITTLE_ENDIAN__
vorn_u16(uint16x4_t __p0,uint16x4_t __p1)17533 __ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
17534   uint16x4_t __ret;
17535   __ret = __p0 | ~__p1;
17536   return __ret;
17537 }
17538 #else
vorn_u16(uint16x4_t __p0,uint16x4_t __p1)17539 __ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
17540   uint16x4_t __ret;
17541   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17542   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17543   __ret = __rev0 | ~__rev1;
17544   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17545   return __ret;
17546 }
17547 #endif
17548 
17549 #ifdef __LITTLE_ENDIAN__
vorn_s8(int8x8_t __p0,int8x8_t __p1)17550 __ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
17551   int8x8_t __ret;
17552   __ret = __p0 | ~__p1;
17553   return __ret;
17554 }
17555 #else
vorn_s8(int8x8_t __p0,int8x8_t __p1)17556 __ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
17557   int8x8_t __ret;
17558   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17559   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17560   __ret = __rev0 | ~__rev1;
17561   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17562   return __ret;
17563 }
17564 #endif
17565 
17566 #ifdef __LITTLE_ENDIAN__
vorn_s32(int32x2_t __p0,int32x2_t __p1)17567 __ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
17568   int32x2_t __ret;
17569   __ret = __p0 | ~__p1;
17570   return __ret;
17571 }
17572 #else
vorn_s32(int32x2_t __p0,int32x2_t __p1)17573 __ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
17574   int32x2_t __ret;
17575   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17576   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17577   __ret = __rev0 | ~__rev1;
17578   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17579   return __ret;
17580 }
17581 #endif
17582 
vorn_s64(int64x1_t __p0,int64x1_t __p1)17583 __ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
17584   int64x1_t __ret;
17585   __ret = __p0 | ~__p1;
17586   return __ret;
17587 }
17588 #ifdef __LITTLE_ENDIAN__
vorn_s16(int16x4_t __p0,int16x4_t __p1)17589 __ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
17590   int16x4_t __ret;
17591   __ret = __p0 | ~__p1;
17592   return __ret;
17593 }
17594 #else
vorn_s16(int16x4_t __p0,int16x4_t __p1)17595 __ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
17596   int16x4_t __ret;
17597   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17598   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17599   __ret = __rev0 | ~__rev1;
17600   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17601   return __ret;
17602 }
17603 #endif
17604 
17605 #ifdef __LITTLE_ENDIAN__
vorrq_u8(uint8x16_t __p0,uint8x16_t __p1)17606 __ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
17607   uint8x16_t __ret;
17608   __ret = __p0 | __p1;
17609   return __ret;
17610 }
17611 #else
vorrq_u8(uint8x16_t __p0,uint8x16_t __p1)17612 __ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
17613   uint8x16_t __ret;
17614   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17615   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17616   __ret = __rev0 | __rev1;
17617   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17618   return __ret;
17619 }
17620 #endif
17621 
17622 #ifdef __LITTLE_ENDIAN__
vorrq_u32(uint32x4_t __p0,uint32x4_t __p1)17623 __ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
17624   uint32x4_t __ret;
17625   __ret = __p0 | __p1;
17626   return __ret;
17627 }
17628 #else
vorrq_u32(uint32x4_t __p0,uint32x4_t __p1)17629 __ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
17630   uint32x4_t __ret;
17631   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17632   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17633   __ret = __rev0 | __rev1;
17634   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17635   return __ret;
17636 }
17637 #endif
17638 
17639 #ifdef __LITTLE_ENDIAN__
vorrq_u64(uint64x2_t __p0,uint64x2_t __p1)17640 __ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
17641   uint64x2_t __ret;
17642   __ret = __p0 | __p1;
17643   return __ret;
17644 }
17645 #else
vorrq_u64(uint64x2_t __p0,uint64x2_t __p1)17646 __ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
17647   uint64x2_t __ret;
17648   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17649   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17650   __ret = __rev0 | __rev1;
17651   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17652   return __ret;
17653 }
17654 #endif
17655 
17656 #ifdef __LITTLE_ENDIAN__
vorrq_u16(uint16x8_t __p0,uint16x8_t __p1)17657 __ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
17658   uint16x8_t __ret;
17659   __ret = __p0 | __p1;
17660   return __ret;
17661 }
17662 #else
vorrq_u16(uint16x8_t __p0,uint16x8_t __p1)17663 __ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
17664   uint16x8_t __ret;
17665   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17666   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17667   __ret = __rev0 | __rev1;
17668   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17669   return __ret;
17670 }
17671 #endif
17672 
17673 #ifdef __LITTLE_ENDIAN__
vorrq_s8(int8x16_t __p0,int8x16_t __p1)17674 __ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
17675   int8x16_t __ret;
17676   __ret = __p0 | __p1;
17677   return __ret;
17678 }
17679 #else
vorrq_s8(int8x16_t __p0,int8x16_t __p1)17680 __ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
17681   int8x16_t __ret;
17682   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17683   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17684   __ret = __rev0 | __rev1;
17685   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17686   return __ret;
17687 }
17688 #endif
17689 
17690 #ifdef __LITTLE_ENDIAN__
vorrq_s32(int32x4_t __p0,int32x4_t __p1)17691 __ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
17692   int32x4_t __ret;
17693   __ret = __p0 | __p1;
17694   return __ret;
17695 }
17696 #else
vorrq_s32(int32x4_t __p0,int32x4_t __p1)17697 __ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
17698   int32x4_t __ret;
17699   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17700   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17701   __ret = __rev0 | __rev1;
17702   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17703   return __ret;
17704 }
17705 #endif
17706 
17707 #ifdef __LITTLE_ENDIAN__
vorrq_s64(int64x2_t __p0,int64x2_t __p1)17708 __ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
17709   int64x2_t __ret;
17710   __ret = __p0 | __p1;
17711   return __ret;
17712 }
17713 #else
vorrq_s64(int64x2_t __p0,int64x2_t __p1)17714 __ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
17715   int64x2_t __ret;
17716   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17717   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17718   __ret = __rev0 | __rev1;
17719   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17720   return __ret;
17721 }
17722 #endif
17723 
17724 #ifdef __LITTLE_ENDIAN__
vorrq_s16(int16x8_t __p0,int16x8_t __p1)17725 __ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
17726   int16x8_t __ret;
17727   __ret = __p0 | __p1;
17728   return __ret;
17729 }
17730 #else
vorrq_s16(int16x8_t __p0,int16x8_t __p1)17731 __ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
17732   int16x8_t __ret;
17733   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17734   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17735   __ret = __rev0 | __rev1;
17736   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17737   return __ret;
17738 }
17739 #endif
17740 
17741 #ifdef __LITTLE_ENDIAN__
vorr_u8(uint8x8_t __p0,uint8x8_t __p1)17742 __ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
17743   uint8x8_t __ret;
17744   __ret = __p0 | __p1;
17745   return __ret;
17746 }
17747 #else
vorr_u8(uint8x8_t __p0,uint8x8_t __p1)17748 __ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
17749   uint8x8_t __ret;
17750   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17751   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17752   __ret = __rev0 | __rev1;
17753   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17754   return __ret;
17755 }
17756 #endif
17757 
17758 #ifdef __LITTLE_ENDIAN__
vorr_u32(uint32x2_t __p0,uint32x2_t __p1)17759 __ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
17760   uint32x2_t __ret;
17761   __ret = __p0 | __p1;
17762   return __ret;
17763 }
17764 #else
vorr_u32(uint32x2_t __p0,uint32x2_t __p1)17765 __ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
17766   uint32x2_t __ret;
17767   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17768   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17769   __ret = __rev0 | __rev1;
17770   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17771   return __ret;
17772 }
17773 #endif
17774 
vorr_u64(uint64x1_t __p0,uint64x1_t __p1)17775 __ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
17776   uint64x1_t __ret;
17777   __ret = __p0 | __p1;
17778   return __ret;
17779 }
17780 #ifdef __LITTLE_ENDIAN__
vorr_u16(uint16x4_t __p0,uint16x4_t __p1)17781 __ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
17782   uint16x4_t __ret;
17783   __ret = __p0 | __p1;
17784   return __ret;
17785 }
17786 #else
vorr_u16(uint16x4_t __p0,uint16x4_t __p1)17787 __ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
17788   uint16x4_t __ret;
17789   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17790   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17791   __ret = __rev0 | __rev1;
17792   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17793   return __ret;
17794 }
17795 #endif
17796 
17797 #ifdef __LITTLE_ENDIAN__
vorr_s8(int8x8_t __p0,int8x8_t __p1)17798 __ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
17799   int8x8_t __ret;
17800   __ret = __p0 | __p1;
17801   return __ret;
17802 }
17803 #else
vorr_s8(int8x8_t __p0,int8x8_t __p1)17804 __ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
17805   int8x8_t __ret;
17806   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17807   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17808   __ret = __rev0 | __rev1;
17809   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17810   return __ret;
17811 }
17812 #endif
17813 
17814 #ifdef __LITTLE_ENDIAN__
vorr_s32(int32x2_t __p0,int32x2_t __p1)17815 __ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
17816   int32x2_t __ret;
17817   __ret = __p0 | __p1;
17818   return __ret;
17819 }
17820 #else
vorr_s32(int32x2_t __p0,int32x2_t __p1)17821 __ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
17822   int32x2_t __ret;
17823   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17824   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17825   __ret = __rev0 | __rev1;
17826   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17827   return __ret;
17828 }
17829 #endif
17830 
vorr_s64(int64x1_t __p0,int64x1_t __p1)17831 __ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
17832   int64x1_t __ret;
17833   __ret = __p0 | __p1;
17834   return __ret;
17835 }
17836 #ifdef __LITTLE_ENDIAN__
vorr_s16(int16x4_t __p0,int16x4_t __p1)17837 __ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
17838   int16x4_t __ret;
17839   __ret = __p0 | __p1;
17840   return __ret;
17841 }
17842 #else
vorr_s16(int16x4_t __p0,int16x4_t __p1)17843 __ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
17844   int16x4_t __ret;
17845   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17846   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17847   __ret = __rev0 | __rev1;
17848   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17849   return __ret;
17850 }
17851 #endif
17852 
17853 #ifdef __LITTLE_ENDIAN__
vpadalq_u8(uint16x8_t __p0,uint8x16_t __p1)17854 __ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
17855   uint16x8_t __ret;
17856   __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
17857   return __ret;
17858 }
17859 #else
vpadalq_u8(uint16x8_t __p0,uint8x16_t __p1)17860 __ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
17861   uint16x8_t __ret;
17862   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17863   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17864   __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
17865   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17866   return __ret;
17867 }
17868 #endif
17869 
17870 #ifdef __LITTLE_ENDIAN__
vpadalq_u32(uint64x2_t __p0,uint32x4_t __p1)17871 __ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
17872   uint64x2_t __ret;
17873   __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
17874   return __ret;
17875 }
17876 #else
vpadalq_u32(uint64x2_t __p0,uint32x4_t __p1)17877 __ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
17878   uint64x2_t __ret;
17879   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17880   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17881   __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
17882   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17883   return __ret;
17884 }
17885 #endif
17886 
17887 #ifdef __LITTLE_ENDIAN__
vpadalq_u16(uint32x4_t __p0,uint16x8_t __p1)17888 __ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
17889   uint32x4_t __ret;
17890   __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
17891   return __ret;
17892 }
17893 #else
vpadalq_u16(uint32x4_t __p0,uint16x8_t __p1)17894 __ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
17895   uint32x4_t __ret;
17896   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17897   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17898   __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
17899   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17900   return __ret;
17901 }
17902 #endif
17903 
17904 #ifdef __LITTLE_ENDIAN__
vpadalq_s8(int16x8_t __p0,int8x16_t __p1)17905 __ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
17906   int16x8_t __ret;
17907   __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17908   return __ret;
17909 }
17910 #else
vpadalq_s8(int16x8_t __p0,int8x16_t __p1)17911 __ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
17912   int16x8_t __ret;
17913   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17914   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17915   __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
17916   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17917   return __ret;
17918 }
17919 #endif
17920 
17921 #ifdef __LITTLE_ENDIAN__
vpadalq_s32(int64x2_t __p0,int32x4_t __p1)17922 __ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
17923   int64x2_t __ret;
17924   __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
17925   return __ret;
17926 }
17927 #else
vpadalq_s32(int64x2_t __p0,int32x4_t __p1)17928 __ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
17929   int64x2_t __ret;
17930   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17931   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17932   __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
17933   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17934   return __ret;
17935 }
17936 #endif
17937 
17938 #ifdef __LITTLE_ENDIAN__
vpadalq_s16(int32x4_t __p0,int16x8_t __p1)17939 __ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
17940   int32x4_t __ret;
17941   __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17942   return __ret;
17943 }
17944 #else
vpadalq_s16(int32x4_t __p0,int16x8_t __p1)17945 __ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
17946   int32x4_t __ret;
17947   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17948   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17949   __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
17950   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17951   return __ret;
17952 }
17953 #endif
17954 
17955 #ifdef __LITTLE_ENDIAN__
vpadal_u8(uint16x4_t __p0,uint8x8_t __p1)17956 __ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
17957   uint16x4_t __ret;
17958   __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
17959   return __ret;
17960 }
17961 #else
vpadal_u8(uint16x4_t __p0,uint8x8_t __p1)17962 __ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
17963   uint16x4_t __ret;
17964   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17965   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17966   __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
17967   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17968   return __ret;
17969 }
17970 #endif
17971 
17972 #ifdef __LITTLE_ENDIAN__
vpadal_u32(uint64x1_t __p0,uint32x2_t __p1)17973 __ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
17974   uint64x1_t __ret;
17975   __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
17976   return __ret;
17977 }
17978 #else
vpadal_u32(uint64x1_t __p0,uint32x2_t __p1)17979 __ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
17980   uint64x1_t __ret;
17981   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17982   __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 19);
17983   return __ret;
17984 }
17985 #endif
17986 
17987 #ifdef __LITTLE_ENDIAN__
vpadal_u16(uint32x2_t __p0,uint16x4_t __p1)17988 __ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
17989   uint32x2_t __ret;
17990   __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
17991   return __ret;
17992 }
17993 #else
vpadal_u16(uint32x2_t __p0,uint16x4_t __p1)17994 __ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
17995   uint32x2_t __ret;
17996   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17997   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17998   __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
17999   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18000   return __ret;
18001 }
18002 #endif
18003 
18004 #ifdef __LITTLE_ENDIAN__
vpadal_s8(int16x4_t __p0,int8x8_t __p1)18005 __ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
18006   int16x4_t __ret;
18007   __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18008   return __ret;
18009 }
18010 #else
vpadal_s8(int16x4_t __p0,int8x8_t __p1)18011 __ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
18012   int16x4_t __ret;
18013   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18014   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18015   __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18016   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18017   return __ret;
18018 }
18019 #endif
18020 
18021 #ifdef __LITTLE_ENDIAN__
vpadal_s32(int64x1_t __p0,int32x2_t __p1)18022 __ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
18023   int64x1_t __ret;
18024   __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18025   return __ret;
18026 }
18027 #else
vpadal_s32(int64x1_t __p0,int32x2_t __p1)18028 __ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
18029   int64x1_t __ret;
18030   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18031   __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 3);
18032   return __ret;
18033 }
18034 #endif
18035 
18036 #ifdef __LITTLE_ENDIAN__
vpadal_s16(int32x2_t __p0,int16x4_t __p1)18037 __ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
18038   int32x2_t __ret;
18039   __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18040   return __ret;
18041 }
18042 #else
vpadal_s16(int32x2_t __p0,int16x4_t __p1)18043 __ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
18044   int32x2_t __ret;
18045   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18046   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18047   __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18048   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18049   return __ret;
18050 }
18051 #endif
18052 
18053 #ifdef __LITTLE_ENDIAN__
vpadd_u8(uint8x8_t __p0,uint8x8_t __p1)18054 __ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
18055   uint8x8_t __ret;
18056   __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18057   return __ret;
18058 }
18059 #else
vpadd_u8(uint8x8_t __p0,uint8x8_t __p1)18060 __ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
18061   uint8x8_t __ret;
18062   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18063   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18064   __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18065   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18066   return __ret;
18067 }
18068 #endif
18069 
18070 #ifdef __LITTLE_ENDIAN__
vpadd_u32(uint32x2_t __p0,uint32x2_t __p1)18071 __ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
18072   uint32x2_t __ret;
18073   __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18074   return __ret;
18075 }
18076 #else
vpadd_u32(uint32x2_t __p0,uint32x2_t __p1)18077 __ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
18078   uint32x2_t __ret;
18079   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18080   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18081   __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18082   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18083   return __ret;
18084 }
18085 #endif
18086 
18087 #ifdef __LITTLE_ENDIAN__
vpadd_u16(uint16x4_t __p0,uint16x4_t __p1)18088 __ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
18089   uint16x4_t __ret;
18090   __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18091   return __ret;
18092 }
18093 #else
vpadd_u16(uint16x4_t __p0,uint16x4_t __p1)18094 __ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
18095   uint16x4_t __ret;
18096   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18097   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18098   __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18099   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18100   return __ret;
18101 }
18102 #endif
18103 
18104 #ifdef __LITTLE_ENDIAN__
vpadd_s8(int8x8_t __p0,int8x8_t __p1)18105 __ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
18106   int8x8_t __ret;
18107   __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18108   return __ret;
18109 }
18110 #else
vpadd_s8(int8x8_t __p0,int8x8_t __p1)18111 __ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
18112   int8x8_t __ret;
18113   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18114   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18115   __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18116   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18117   return __ret;
18118 }
18119 #endif
18120 
18121 #ifdef __LITTLE_ENDIAN__
vpadd_f32(float32x2_t __p0,float32x2_t __p1)18122 __ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
18123   float32x2_t __ret;
18124   __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
18125   return __ret;
18126 }
18127 #else
vpadd_f32(float32x2_t __p0,float32x2_t __p1)18128 __ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
18129   float32x2_t __ret;
18130   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18131   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18132   __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
18133   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18134   return __ret;
18135 }
18136 #endif
18137 
18138 #ifdef __LITTLE_ENDIAN__
vpadd_s32(int32x2_t __p0,int32x2_t __p1)18139 __ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
18140   int32x2_t __ret;
18141   __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18142   return __ret;
18143 }
18144 #else
vpadd_s32(int32x2_t __p0,int32x2_t __p1)18145 __ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
18146   int32x2_t __ret;
18147   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18148   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18149   __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18150   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18151   return __ret;
18152 }
18153 #endif
18154 
18155 #ifdef __LITTLE_ENDIAN__
vpadd_s16(int16x4_t __p0,int16x4_t __p1)18156 __ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
18157   int16x4_t __ret;
18158   __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18159   return __ret;
18160 }
18161 #else
vpadd_s16(int16x4_t __p0,int16x4_t __p1)18162 __ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
18163   int16x4_t __ret;
18164   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18165   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18166   __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18167   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18168   return __ret;
18169 }
18170 #endif
18171 
18172 #ifdef __LITTLE_ENDIAN__
vpaddlq_u8(uint8x16_t __p0)18173 __ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
18174   uint16x8_t __ret;
18175   __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 49);
18176   return __ret;
18177 }
18178 #else
vpaddlq_u8(uint8x16_t __p0)18179 __ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
18180   uint16x8_t __ret;
18181   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18182   __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 49);
18183   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18184   return __ret;
18185 }
18186 #endif
18187 
18188 #ifdef __LITTLE_ENDIAN__
vpaddlq_u32(uint32x4_t __p0)18189 __ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
18190   uint64x2_t __ret;
18191   __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 51);
18192   return __ret;
18193 }
18194 #else
vpaddlq_u32(uint32x4_t __p0)18195 __ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
18196   uint64x2_t __ret;
18197   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18198   __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 51);
18199   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18200   return __ret;
18201 }
18202 #endif
18203 
18204 #ifdef __LITTLE_ENDIAN__
vpaddlq_u16(uint16x8_t __p0)18205 __ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
18206   uint32x4_t __ret;
18207   __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 50);
18208   return __ret;
18209 }
18210 #else
vpaddlq_u16(uint16x8_t __p0)18211 __ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
18212   uint32x4_t __ret;
18213   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18214   __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 50);
18215   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18216   return __ret;
18217 }
18218 #endif
18219 
18220 #ifdef __LITTLE_ENDIAN__
vpaddlq_s8(int8x16_t __p0)18221 __ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
18222   int16x8_t __ret;
18223   __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 33);
18224   return __ret;
18225 }
18226 #else
vpaddlq_s8(int8x16_t __p0)18227 __ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
18228   int16x8_t __ret;
18229   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18230   __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 33);
18231   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18232   return __ret;
18233 }
18234 #endif
18235 
18236 #ifdef __LITTLE_ENDIAN__
vpaddlq_s32(int32x4_t __p0)18237 __ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
18238   int64x2_t __ret;
18239   __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 35);
18240   return __ret;
18241 }
18242 #else
vpaddlq_s32(int32x4_t __p0)18243 __ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
18244   int64x2_t __ret;
18245   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18246   __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 35);
18247   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18248   return __ret;
18249 }
18250 #endif
18251 
18252 #ifdef __LITTLE_ENDIAN__
vpaddlq_s16(int16x8_t __p0)18253 __ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
18254   int32x4_t __ret;
18255   __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 34);
18256   return __ret;
18257 }
18258 #else
vpaddlq_s16(int16x8_t __p0)18259 __ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
18260   int32x4_t __ret;
18261   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18262   __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 34);
18263   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18264   return __ret;
18265 }
18266 #endif
18267 
18268 #ifdef __LITTLE_ENDIAN__
vpaddl_u8(uint8x8_t __p0)18269 __ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
18270   uint16x4_t __ret;
18271   __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 17);
18272   return __ret;
18273 }
18274 #else
vpaddl_u8(uint8x8_t __p0)18275 __ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
18276   uint16x4_t __ret;
18277   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18278   __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 17);
18279   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18280   return __ret;
18281 }
18282 #endif
18283 
18284 #ifdef __LITTLE_ENDIAN__
vpaddl_u32(uint32x2_t __p0)18285 __ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
18286   uint64x1_t __ret;
18287   __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 19);
18288   return __ret;
18289 }
18290 #else
vpaddl_u32(uint32x2_t __p0)18291 __ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
18292   uint64x1_t __ret;
18293   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18294   __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 19);
18295   return __ret;
18296 }
18297 #endif
18298 
18299 #ifdef __LITTLE_ENDIAN__
vpaddl_u16(uint16x4_t __p0)18300 __ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
18301   uint32x2_t __ret;
18302   __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 18);
18303   return __ret;
18304 }
18305 #else
vpaddl_u16(uint16x4_t __p0)18306 __ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
18307   uint32x2_t __ret;
18308   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18309   __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 18);
18310   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18311   return __ret;
18312 }
18313 #endif
18314 
18315 #ifdef __LITTLE_ENDIAN__
vpaddl_s8(int8x8_t __p0)18316 __ai int16x4_t vpaddl_s8(int8x8_t __p0) {
18317   int16x4_t __ret;
18318   __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 1);
18319   return __ret;
18320 }
18321 #else
vpaddl_s8(int8x8_t __p0)18322 __ai int16x4_t vpaddl_s8(int8x8_t __p0) {
18323   int16x4_t __ret;
18324   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18325   __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 1);
18326   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18327   return __ret;
18328 }
18329 #endif
18330 
18331 #ifdef __LITTLE_ENDIAN__
vpaddl_s32(int32x2_t __p0)18332 __ai int64x1_t vpaddl_s32(int32x2_t __p0) {
18333   int64x1_t __ret;
18334   __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 3);
18335   return __ret;
18336 }
18337 #else
vpaddl_s32(int32x2_t __p0)18338 __ai int64x1_t vpaddl_s32(int32x2_t __p0) {
18339   int64x1_t __ret;
18340   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18341   __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 3);
18342   return __ret;
18343 }
18344 #endif
18345 
18346 #ifdef __LITTLE_ENDIAN__
vpaddl_s16(int16x4_t __p0)18347 __ai int32x2_t vpaddl_s16(int16x4_t __p0) {
18348   int32x2_t __ret;
18349   __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 2);
18350   return __ret;
18351 }
18352 #else
vpaddl_s16(int16x4_t __p0)18353 __ai int32x2_t vpaddl_s16(int16x4_t __p0) {
18354   int32x2_t __ret;
18355   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18356   __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 2);
18357   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18358   return __ret;
18359 }
18360 #endif
18361 
18362 #ifdef __LITTLE_ENDIAN__
vpmax_u8(uint8x8_t __p0,uint8x8_t __p1)18363 __ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
18364   uint8x8_t __ret;
18365   __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18366   return __ret;
18367 }
18368 #else
vpmax_u8(uint8x8_t __p0,uint8x8_t __p1)18369 __ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
18370   uint8x8_t __ret;
18371   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18372   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18373   __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18374   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18375   return __ret;
18376 }
18377 #endif
18378 
18379 #ifdef __LITTLE_ENDIAN__
vpmax_u32(uint32x2_t __p0,uint32x2_t __p1)18380 __ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
18381   uint32x2_t __ret;
18382   __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18383   return __ret;
18384 }
18385 #else
vpmax_u32(uint32x2_t __p0,uint32x2_t __p1)18386 __ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
18387   uint32x2_t __ret;
18388   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18389   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18390   __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18391   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18392   return __ret;
18393 }
18394 #endif
18395 
18396 #ifdef __LITTLE_ENDIAN__
vpmax_u16(uint16x4_t __p0,uint16x4_t __p1)18397 __ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
18398   uint16x4_t __ret;
18399   __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18400   return __ret;
18401 }
18402 #else
vpmax_u16(uint16x4_t __p0,uint16x4_t __p1)18403 __ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
18404   uint16x4_t __ret;
18405   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18406   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18407   __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18408   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18409   return __ret;
18410 }
18411 #endif
18412 
18413 #ifdef __LITTLE_ENDIAN__
vpmax_s8(int8x8_t __p0,int8x8_t __p1)18414 __ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
18415   int8x8_t __ret;
18416   __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18417   return __ret;
18418 }
18419 #else
vpmax_s8(int8x8_t __p0,int8x8_t __p1)18420 __ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
18421   int8x8_t __ret;
18422   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18423   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18424   __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18425   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18426   return __ret;
18427 }
18428 #endif
18429 
18430 #ifdef __LITTLE_ENDIAN__
vpmax_f32(float32x2_t __p0,float32x2_t __p1)18431 __ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
18432   float32x2_t __ret;
18433   __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
18434   return __ret;
18435 }
18436 #else
vpmax_f32(float32x2_t __p0,float32x2_t __p1)18437 __ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
18438   float32x2_t __ret;
18439   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18440   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18441   __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
18442   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18443   return __ret;
18444 }
18445 #endif
18446 
18447 #ifdef __LITTLE_ENDIAN__
vpmax_s32(int32x2_t __p0,int32x2_t __p1)18448 __ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
18449   int32x2_t __ret;
18450   __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18451   return __ret;
18452 }
18453 #else
vpmax_s32(int32x2_t __p0,int32x2_t __p1)18454 __ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
18455   int32x2_t __ret;
18456   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18457   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18458   __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18459   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18460   return __ret;
18461 }
18462 #endif
18463 
18464 #ifdef __LITTLE_ENDIAN__
vpmax_s16(int16x4_t __p0,int16x4_t __p1)18465 __ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
18466   int16x4_t __ret;
18467   __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18468   return __ret;
18469 }
18470 #else
vpmax_s16(int16x4_t __p0,int16x4_t __p1)18471 __ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
18472   int16x4_t __ret;
18473   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18474   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18475   __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18476   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18477   return __ret;
18478 }
18479 #endif
18480 
18481 #ifdef __LITTLE_ENDIAN__
vpmin_u8(uint8x8_t __p0,uint8x8_t __p1)18482 __ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
18483   uint8x8_t __ret;
18484   __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18485   return __ret;
18486 }
18487 #else
vpmin_u8(uint8x8_t __p0,uint8x8_t __p1)18488 __ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
18489   uint8x8_t __ret;
18490   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18491   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18492   __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18493   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18494   return __ret;
18495 }
18496 #endif
18497 
18498 #ifdef __LITTLE_ENDIAN__
vpmin_u32(uint32x2_t __p0,uint32x2_t __p1)18499 __ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
18500   uint32x2_t __ret;
18501   __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18502   return __ret;
18503 }
18504 #else
vpmin_u32(uint32x2_t __p0,uint32x2_t __p1)18505 __ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
18506   uint32x2_t __ret;
18507   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18508   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18509   __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18510   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18511   return __ret;
18512 }
18513 #endif
18514 
18515 #ifdef __LITTLE_ENDIAN__
vpmin_u16(uint16x4_t __p0,uint16x4_t __p1)18516 __ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
18517   uint16x4_t __ret;
18518   __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18519   return __ret;
18520 }
18521 #else
vpmin_u16(uint16x4_t __p0,uint16x4_t __p1)18522 __ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
18523   uint16x4_t __ret;
18524   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18525   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18526   __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18527   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18528   return __ret;
18529 }
18530 #endif
18531 
18532 #ifdef __LITTLE_ENDIAN__
vpmin_s8(int8x8_t __p0,int8x8_t __p1)18533 __ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
18534   int8x8_t __ret;
18535   __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18536   return __ret;
18537 }
18538 #else
vpmin_s8(int8x8_t __p0,int8x8_t __p1)18539 __ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
18540   int8x8_t __ret;
18541   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18542   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18543   __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18544   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18545   return __ret;
18546 }
18547 #endif
18548 
18549 #ifdef __LITTLE_ENDIAN__
vpmin_f32(float32x2_t __p0,float32x2_t __p1)18550 __ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
18551   float32x2_t __ret;
18552   __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
18553   return __ret;
18554 }
18555 #else
vpmin_f32(float32x2_t __p0,float32x2_t __p1)18556 __ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
18557   float32x2_t __ret;
18558   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18559   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18560   __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
18561   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18562   return __ret;
18563 }
18564 #endif
18565 
18566 #ifdef __LITTLE_ENDIAN__
vpmin_s32(int32x2_t __p0,int32x2_t __p1)18567 __ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
18568   int32x2_t __ret;
18569   __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18570   return __ret;
18571 }
18572 #else
vpmin_s32(int32x2_t __p0,int32x2_t __p1)18573 __ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
18574   int32x2_t __ret;
18575   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18576   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18577   __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18578   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18579   return __ret;
18580 }
18581 #endif
18582 
18583 #ifdef __LITTLE_ENDIAN__
vpmin_s16(int16x4_t __p0,int16x4_t __p1)18584 __ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
18585   int16x4_t __ret;
18586   __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18587   return __ret;
18588 }
18589 #else
vpmin_s16(int16x4_t __p0,int16x4_t __p1)18590 __ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
18591   int16x4_t __ret;
18592   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18593   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18594   __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18595   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18596   return __ret;
18597 }
18598 #endif
18599 
18600 #ifdef __LITTLE_ENDIAN__
vqabsq_s8(int8x16_t __p0)18601 __ai int8x16_t vqabsq_s8(int8x16_t __p0) {
18602   int8x16_t __ret;
18603   __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 32);
18604   return __ret;
18605 }
18606 #else
vqabsq_s8(int8x16_t __p0)18607 __ai int8x16_t vqabsq_s8(int8x16_t __p0) {
18608   int8x16_t __ret;
18609   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18610   __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 32);
18611   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18612   return __ret;
18613 }
18614 #endif
18615 
18616 #ifdef __LITTLE_ENDIAN__
vqabsq_s32(int32x4_t __p0)18617 __ai int32x4_t vqabsq_s32(int32x4_t __p0) {
18618   int32x4_t __ret;
18619   __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 34);
18620   return __ret;
18621 }
18622 #else
vqabsq_s32(int32x4_t __p0)18623 __ai int32x4_t vqabsq_s32(int32x4_t __p0) {
18624   int32x4_t __ret;
18625   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18626   __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 34);
18627   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18628   return __ret;
18629 }
18630 #endif
18631 
18632 #ifdef __LITTLE_ENDIAN__
vqabsq_s16(int16x8_t __p0)18633 __ai int16x8_t vqabsq_s16(int16x8_t __p0) {
18634   int16x8_t __ret;
18635   __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 33);
18636   return __ret;
18637 }
18638 #else
vqabsq_s16(int16x8_t __p0)18639 __ai int16x8_t vqabsq_s16(int16x8_t __p0) {
18640   int16x8_t __ret;
18641   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18642   __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 33);
18643   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18644   return __ret;
18645 }
18646 #endif
18647 
18648 #ifdef __LITTLE_ENDIAN__
vqabs_s8(int8x8_t __p0)18649 __ai int8x8_t vqabs_s8(int8x8_t __p0) {
18650   int8x8_t __ret;
18651   __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 0);
18652   return __ret;
18653 }
18654 #else
vqabs_s8(int8x8_t __p0)18655 __ai int8x8_t vqabs_s8(int8x8_t __p0) {
18656   int8x8_t __ret;
18657   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18658   __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 0);
18659   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18660   return __ret;
18661 }
18662 #endif
18663 
18664 #ifdef __LITTLE_ENDIAN__
vqabs_s32(int32x2_t __p0)18665 __ai int32x2_t vqabs_s32(int32x2_t __p0) {
18666   int32x2_t __ret;
18667   __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 2);
18668   return __ret;
18669 }
18670 #else
vqabs_s32(int32x2_t __p0)18671 __ai int32x2_t vqabs_s32(int32x2_t __p0) {
18672   int32x2_t __ret;
18673   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18674   __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 2);
18675   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18676   return __ret;
18677 }
18678 #endif
18679 
18680 #ifdef __LITTLE_ENDIAN__
vqabs_s16(int16x4_t __p0)18681 __ai int16x4_t vqabs_s16(int16x4_t __p0) {
18682   int16x4_t __ret;
18683   __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 1);
18684   return __ret;
18685 }
18686 #else
vqabs_s16(int16x4_t __p0)18687 __ai int16x4_t vqabs_s16(int16x4_t __p0) {
18688   int16x4_t __ret;
18689   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18690   __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 1);
18691   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18692   return __ret;
18693 }
18694 #endif
18695 
18696 #ifdef __LITTLE_ENDIAN__
vqaddq_u8(uint8x16_t __p0,uint8x16_t __p1)18697 __ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
18698   uint8x16_t __ret;
18699   __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18700   return __ret;
18701 }
18702 #else
vqaddq_u8(uint8x16_t __p0,uint8x16_t __p1)18703 __ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
18704   uint8x16_t __ret;
18705   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18706   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18707   __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18708   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18709   return __ret;
18710 }
18711 #endif
18712 
18713 #ifdef __LITTLE_ENDIAN__
vqaddq_u32(uint32x4_t __p0,uint32x4_t __p1)18714 __ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
18715   uint32x4_t __ret;
18716   __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18717   return __ret;
18718 }
18719 #else
vqaddq_u32(uint32x4_t __p0,uint32x4_t __p1)18720 __ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
18721   uint32x4_t __ret;
18722   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18723   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18724   __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18725   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18726   return __ret;
18727 }
18728 #endif
18729 
18730 #ifdef __LITTLE_ENDIAN__
vqaddq_u64(uint64x2_t __p0,uint64x2_t __p1)18731 __ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
18732   uint64x2_t __ret;
18733   __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18734   return __ret;
18735 }
18736 #else
vqaddq_u64(uint64x2_t __p0,uint64x2_t __p1)18737 __ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
18738   uint64x2_t __ret;
18739   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18740   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18741   __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18742   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18743   return __ret;
18744 }
18745 #endif
18746 
18747 #ifdef __LITTLE_ENDIAN__
vqaddq_u16(uint16x8_t __p0,uint16x8_t __p1)18748 __ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
18749   uint16x8_t __ret;
18750   __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18751   return __ret;
18752 }
18753 #else
vqaddq_u16(uint16x8_t __p0,uint16x8_t __p1)18754 __ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
18755   uint16x8_t __ret;
18756   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18757   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18758   __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18759   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18760   return __ret;
18761 }
18762 #endif
18763 
18764 #ifdef __LITTLE_ENDIAN__
vqaddq_s8(int8x16_t __p0,int8x16_t __p1)18765 __ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
18766   int8x16_t __ret;
18767   __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18768   return __ret;
18769 }
18770 #else
vqaddq_s8(int8x16_t __p0,int8x16_t __p1)18771 __ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
18772   int8x16_t __ret;
18773   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18774   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18775   __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18776   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18777   return __ret;
18778 }
18779 #endif
18780 
18781 #ifdef __LITTLE_ENDIAN__
vqaddq_s32(int32x4_t __p0,int32x4_t __p1)18782 __ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
18783   int32x4_t __ret;
18784   __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18785   return __ret;
18786 }
18787 #else
vqaddq_s32(int32x4_t __p0,int32x4_t __p1)18788 __ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
18789   int32x4_t __ret;
18790   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18791   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18792   __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18793   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18794   return __ret;
18795 }
18796 #endif
18797 
18798 #ifdef __LITTLE_ENDIAN__
vqaddq_s64(int64x2_t __p0,int64x2_t __p1)18799 __ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
18800   int64x2_t __ret;
18801   __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18802   return __ret;
18803 }
18804 #else
vqaddq_s64(int64x2_t __p0,int64x2_t __p1)18805 __ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
18806   int64x2_t __ret;
18807   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18808   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18809   __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18810   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18811   return __ret;
18812 }
18813 #endif
18814 
18815 #ifdef __LITTLE_ENDIAN__
vqaddq_s16(int16x8_t __p0,int16x8_t __p1)18816 __ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
18817   int16x8_t __ret;
18818   __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18819   return __ret;
18820 }
18821 #else
vqaddq_s16(int16x8_t __p0,int16x8_t __p1)18822 __ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
18823   int16x8_t __ret;
18824   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18825   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18826   __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18827   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18828   return __ret;
18829 }
18830 #endif
18831 
18832 #ifdef __LITTLE_ENDIAN__
vqadd_u8(uint8x8_t __p0,uint8x8_t __p1)18833 __ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
18834   uint8x8_t __ret;
18835   __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18836   return __ret;
18837 }
18838 #else
vqadd_u8(uint8x8_t __p0,uint8x8_t __p1)18839 __ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
18840   uint8x8_t __ret;
18841   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18842   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18843   __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18844   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18845   return __ret;
18846 }
18847 #endif
18848 
18849 #ifdef __LITTLE_ENDIAN__
vqadd_u32(uint32x2_t __p0,uint32x2_t __p1)18850 __ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
18851   uint32x2_t __ret;
18852   __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18853   return __ret;
18854 }
18855 #else
vqadd_u32(uint32x2_t __p0,uint32x2_t __p1)18856 __ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
18857   uint32x2_t __ret;
18858   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18859   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18860   __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18861   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18862   return __ret;
18863 }
18864 #endif
18865 
vqadd_u64(uint64x1_t __p0,uint64x1_t __p1)18866 __ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
18867   uint64x1_t __ret;
18868   __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18869   return __ret;
18870 }
18871 #ifdef __LITTLE_ENDIAN__
vqadd_u16(uint16x4_t __p0,uint16x4_t __p1)18872 __ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
18873   uint16x4_t __ret;
18874   __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18875   return __ret;
18876 }
18877 #else
vqadd_u16(uint16x4_t __p0,uint16x4_t __p1)18878 __ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
18879   uint16x4_t __ret;
18880   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18881   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18882   __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18883   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18884   return __ret;
18885 }
18886 #endif
18887 
18888 #ifdef __LITTLE_ENDIAN__
vqadd_s8(int8x8_t __p0,int8x8_t __p1)18889 __ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
18890   int8x8_t __ret;
18891   __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18892   return __ret;
18893 }
18894 #else
vqadd_s8(int8x8_t __p0,int8x8_t __p1)18895 __ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
18896   int8x8_t __ret;
18897   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18898   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18899   __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18900   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18901   return __ret;
18902 }
18903 #endif
18904 
18905 #ifdef __LITTLE_ENDIAN__
vqadd_s32(int32x2_t __p0,int32x2_t __p1)18906 __ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
18907   int32x2_t __ret;
18908   __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18909   return __ret;
18910 }
18911 #else
vqadd_s32(int32x2_t __p0,int32x2_t __p1)18912 __ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
18913   int32x2_t __ret;
18914   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18915   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18916   __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18917   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18918   return __ret;
18919 }
18920 #endif
18921 
vqadd_s64(int64x1_t __p0,int64x1_t __p1)18922 __ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
18923   int64x1_t __ret;
18924   __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18925   return __ret;
18926 }
18927 #ifdef __LITTLE_ENDIAN__
vqadd_s16(int16x4_t __p0,int16x4_t __p1)18928 __ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
18929   int16x4_t __ret;
18930   __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18931   return __ret;
18932 }
18933 #else
vqadd_s16(int16x4_t __p0,int16x4_t __p1)18934 __ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
18935   int16x4_t __ret;
18936   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18937   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18938   __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18939   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18940   return __ret;
18941 }
18942 #endif
18943 
18944 #ifdef __LITTLE_ENDIAN__
vqdmlal_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)18945 __ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
18946   int64x2_t __ret;
18947   __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
18948   return __ret;
18949 }
18950 #else
vqdmlal_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)18951 __ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
18952   int64x2_t __ret;
18953   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18954   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18955   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
18956   __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
18957   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18958   return __ret;
18959 }
__noswap_vqdmlal_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)18960 __ai int64x2_t __noswap_vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
18961   int64x2_t __ret;
18962   __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
18963   return __ret;
18964 }
18965 #endif
18966 
18967 #ifdef __LITTLE_ENDIAN__
vqdmlal_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)18968 __ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
18969   int32x4_t __ret;
18970   __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
18971   return __ret;
18972 }
18973 #else
vqdmlal_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)18974 __ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
18975   int32x4_t __ret;
18976   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18977   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18978   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
18979   __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
18980   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18981   return __ret;
18982 }
__noswap_vqdmlal_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)18983 __ai int32x4_t __noswap_vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
18984   int32x4_t __ret;
18985   __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
18986   return __ret;
18987 }
18988 #endif
18989 
18990 #ifdef __LITTLE_ENDIAN__
18991 #define vqdmlal_lane_s32(__p0_114, __p1_114, __p2_114, __p3_114) __extension__ ({ \
18992   int64x2_t __ret_114; \
18993   int64x2_t __s0_114 = __p0_114; \
18994   int32x2_t __s1_114 = __p1_114; \
18995   int32x2_t __s2_114 = __p2_114; \
18996   __ret_114 = vqdmlal_s32(__s0_114, __s1_114, splat_lane_s32(__s2_114, __p3_114)); \
18997   __ret_114; \
18998 })
18999 #else
19000 #define vqdmlal_lane_s32(__p0_115, __p1_115, __p2_115, __p3_115) __extension__ ({ \
19001   int64x2_t __ret_115; \
19002   int64x2_t __s0_115 = __p0_115; \
19003   int32x2_t __s1_115 = __p1_115; \
19004   int32x2_t __s2_115 = __p2_115; \
19005   int64x2_t __rev0_115;  __rev0_115 = __builtin_shufflevector(__s0_115, __s0_115, 1, 0); \
19006   int32x2_t __rev1_115;  __rev1_115 = __builtin_shufflevector(__s1_115, __s1_115, 1, 0); \
19007   int32x2_t __rev2_115;  __rev2_115 = __builtin_shufflevector(__s2_115, __s2_115, 1, 0); \
19008   __ret_115 = __noswap_vqdmlal_s32(__rev0_115, __rev1_115, __noswap_splat_lane_s32(__rev2_115, __p3_115)); \
19009   __ret_115 = __builtin_shufflevector(__ret_115, __ret_115, 1, 0); \
19010   __ret_115; \
19011 })
19012 #endif
19013 
19014 #ifdef __LITTLE_ENDIAN__
19015 #define vqdmlal_lane_s16(__p0_116, __p1_116, __p2_116, __p3_116) __extension__ ({ \
19016   int32x4_t __ret_116; \
19017   int32x4_t __s0_116 = __p0_116; \
19018   int16x4_t __s1_116 = __p1_116; \
19019   int16x4_t __s2_116 = __p2_116; \
19020   __ret_116 = vqdmlal_s16(__s0_116, __s1_116, splat_lane_s16(__s2_116, __p3_116)); \
19021   __ret_116; \
19022 })
19023 #else
19024 #define vqdmlal_lane_s16(__p0_117, __p1_117, __p2_117, __p3_117) __extension__ ({ \
19025   int32x4_t __ret_117; \
19026   int32x4_t __s0_117 = __p0_117; \
19027   int16x4_t __s1_117 = __p1_117; \
19028   int16x4_t __s2_117 = __p2_117; \
19029   int32x4_t __rev0_117;  __rev0_117 = __builtin_shufflevector(__s0_117, __s0_117, 3, 2, 1, 0); \
19030   int16x4_t __rev1_117;  __rev1_117 = __builtin_shufflevector(__s1_117, __s1_117, 3, 2, 1, 0); \
19031   int16x4_t __rev2_117;  __rev2_117 = __builtin_shufflevector(__s2_117, __s2_117, 3, 2, 1, 0); \
19032   __ret_117 = __noswap_vqdmlal_s16(__rev0_117, __rev1_117, __noswap_splat_lane_s16(__rev2_117, __p3_117)); \
19033   __ret_117 = __builtin_shufflevector(__ret_117, __ret_117, 3, 2, 1, 0); \
19034   __ret_117; \
19035 })
19036 #endif
19037 
19038 #ifdef __LITTLE_ENDIAN__
vqdmlal_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)19039 __ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19040   int64x2_t __ret;
19041   __ret = vqdmlal_s32(__p0, __p1, (int32x2_t) {__p2, __p2});
19042   return __ret;
19043 }
19044 #else
vqdmlal_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)19045 __ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19046   int64x2_t __ret;
19047   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19048   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19049   __ret = __noswap_vqdmlal_s32(__rev0, __rev1, (int32x2_t) {__p2, __p2});
19050   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19051   return __ret;
19052 }
__noswap_vqdmlal_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)19053 __ai int64x2_t __noswap_vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19054   int64x2_t __ret;
19055   __ret = __noswap_vqdmlal_s32(__p0, __p1, (int32x2_t) {__p2, __p2});
19056   return __ret;
19057 }
19058 #endif
19059 
19060 #ifdef __LITTLE_ENDIAN__
vqdmlal_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)19061 __ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19062   int32x4_t __ret;
19063   __ret = vqdmlal_s16(__p0, __p1, (int16x4_t) {__p2, __p2, __p2, __p2});
19064   return __ret;
19065 }
19066 #else
vqdmlal_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)19067 __ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19068   int32x4_t __ret;
19069   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19070   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19071   __ret = __noswap_vqdmlal_s16(__rev0, __rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
19072   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19073   return __ret;
19074 }
__noswap_vqdmlal_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)19075 __ai int32x4_t __noswap_vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19076   int32x4_t __ret;
19077   __ret = __noswap_vqdmlal_s16(__p0, __p1, (int16x4_t) {__p2, __p2, __p2, __p2});
19078   return __ret;
19079 }
19080 #endif
19081 
19082 #ifdef __LITTLE_ENDIAN__
vqdmlsl_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)19083 __ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
19084   int64x2_t __ret;
19085   __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
19086   return __ret;
19087 }
19088 #else
vqdmlsl_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)19089 __ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
19090   int64x2_t __ret;
19091   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19092   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19093   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
19094   __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
19095   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19096   return __ret;
19097 }
__noswap_vqdmlsl_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)19098 __ai int64x2_t __noswap_vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
19099   int64x2_t __ret;
19100   __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
19101   return __ret;
19102 }
19103 #endif
19104 
19105 #ifdef __LITTLE_ENDIAN__
vqdmlsl_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)19106 __ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
19107   int32x4_t __ret;
19108   __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
19109   return __ret;
19110 }
19111 #else
vqdmlsl_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)19112 __ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
19113   int32x4_t __ret;
19114   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19115   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19116   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
19117   __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
19118   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19119   return __ret;
19120 }
__noswap_vqdmlsl_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)19121 __ai int32x4_t __noswap_vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
19122   int32x4_t __ret;
19123   __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
19124   return __ret;
19125 }
19126 #endif
19127 
19128 #ifdef __LITTLE_ENDIAN__
19129 #define vqdmlsl_lane_s32(__p0_118, __p1_118, __p2_118, __p3_118) __extension__ ({ \
19130   int64x2_t __ret_118; \
19131   int64x2_t __s0_118 = __p0_118; \
19132   int32x2_t __s1_118 = __p1_118; \
19133   int32x2_t __s2_118 = __p2_118; \
19134   __ret_118 = vqdmlsl_s32(__s0_118, __s1_118, splat_lane_s32(__s2_118, __p3_118)); \
19135   __ret_118; \
19136 })
19137 #else
19138 #define vqdmlsl_lane_s32(__p0_119, __p1_119, __p2_119, __p3_119) __extension__ ({ \
19139   int64x2_t __ret_119; \
19140   int64x2_t __s0_119 = __p0_119; \
19141   int32x2_t __s1_119 = __p1_119; \
19142   int32x2_t __s2_119 = __p2_119; \
19143   int64x2_t __rev0_119;  __rev0_119 = __builtin_shufflevector(__s0_119, __s0_119, 1, 0); \
19144   int32x2_t __rev1_119;  __rev1_119 = __builtin_shufflevector(__s1_119, __s1_119, 1, 0); \
19145   int32x2_t __rev2_119;  __rev2_119 = __builtin_shufflevector(__s2_119, __s2_119, 1, 0); \
19146   __ret_119 = __noswap_vqdmlsl_s32(__rev0_119, __rev1_119, __noswap_splat_lane_s32(__rev2_119, __p3_119)); \
19147   __ret_119 = __builtin_shufflevector(__ret_119, __ret_119, 1, 0); \
19148   __ret_119; \
19149 })
19150 #endif
19151 
19152 #ifdef __LITTLE_ENDIAN__
19153 #define vqdmlsl_lane_s16(__p0_120, __p1_120, __p2_120, __p3_120) __extension__ ({ \
19154   int32x4_t __ret_120; \
19155   int32x4_t __s0_120 = __p0_120; \
19156   int16x4_t __s1_120 = __p1_120; \
19157   int16x4_t __s2_120 = __p2_120; \
19158   __ret_120 = vqdmlsl_s16(__s0_120, __s1_120, splat_lane_s16(__s2_120, __p3_120)); \
19159   __ret_120; \
19160 })
19161 #else
19162 #define vqdmlsl_lane_s16(__p0_121, __p1_121, __p2_121, __p3_121) __extension__ ({ \
19163   int32x4_t __ret_121; \
19164   int32x4_t __s0_121 = __p0_121; \
19165   int16x4_t __s1_121 = __p1_121; \
19166   int16x4_t __s2_121 = __p2_121; \
19167   int32x4_t __rev0_121;  __rev0_121 = __builtin_shufflevector(__s0_121, __s0_121, 3, 2, 1, 0); \
19168   int16x4_t __rev1_121;  __rev1_121 = __builtin_shufflevector(__s1_121, __s1_121, 3, 2, 1, 0); \
19169   int16x4_t __rev2_121;  __rev2_121 = __builtin_shufflevector(__s2_121, __s2_121, 3, 2, 1, 0); \
19170   __ret_121 = __noswap_vqdmlsl_s16(__rev0_121, __rev1_121, __noswap_splat_lane_s16(__rev2_121, __p3_121)); \
19171   __ret_121 = __builtin_shufflevector(__ret_121, __ret_121, 3, 2, 1, 0); \
19172   __ret_121; \
19173 })
19174 #endif
19175 
19176 #ifdef __LITTLE_ENDIAN__
vqdmlsl_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)19177 __ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19178   int64x2_t __ret;
19179   __ret = vqdmlsl_s32(__p0, __p1, (int32x2_t) {__p2, __p2});
19180   return __ret;
19181 }
19182 #else
vqdmlsl_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)19183 __ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19184   int64x2_t __ret;
19185   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19186   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19187   __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, (int32x2_t) {__p2, __p2});
19188   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19189   return __ret;
19190 }
__noswap_vqdmlsl_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)19191 __ai int64x2_t __noswap_vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19192   int64x2_t __ret;
19193   __ret = __noswap_vqdmlsl_s32(__p0, __p1, (int32x2_t) {__p2, __p2});
19194   return __ret;
19195 }
19196 #endif
19197 
19198 #ifdef __LITTLE_ENDIAN__
vqdmlsl_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)19199 __ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19200   int32x4_t __ret;
19201   __ret = vqdmlsl_s16(__p0, __p1, (int16x4_t) {__p2, __p2, __p2, __p2});
19202   return __ret;
19203 }
19204 #else
vqdmlsl_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)19205 __ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19206   int32x4_t __ret;
19207   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19208   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19209   __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
19210   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19211   return __ret;
19212 }
__noswap_vqdmlsl_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)19213 __ai int32x4_t __noswap_vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19214   int32x4_t __ret;
19215   __ret = __noswap_vqdmlsl_s16(__p0, __p1, (int16x4_t) {__p2, __p2, __p2, __p2});
19216   return __ret;
19217 }
19218 #endif
19219 
19220 #ifdef __LITTLE_ENDIAN__
vqdmulhq_s32(int32x4_t __p0,int32x4_t __p1)19221 __ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19222   int32x4_t __ret;
19223   __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19224   return __ret;
19225 }
19226 #else
vqdmulhq_s32(int32x4_t __p0,int32x4_t __p1)19227 __ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19228   int32x4_t __ret;
19229   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19230   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19231   __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
19232   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19233   return __ret;
19234 }
__noswap_vqdmulhq_s32(int32x4_t __p0,int32x4_t __p1)19235 __ai int32x4_t __noswap_vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19236   int32x4_t __ret;
19237   __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19238   return __ret;
19239 }
19240 #endif
19241 
19242 #ifdef __LITTLE_ENDIAN__
vqdmulhq_s16(int16x8_t __p0,int16x8_t __p1)19243 __ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19244   int16x8_t __ret;
19245   __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19246   return __ret;
19247 }
19248 #else
vqdmulhq_s16(int16x8_t __p0,int16x8_t __p1)19249 __ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19250   int16x8_t __ret;
19251   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19252   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19253   __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
19254   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19255   return __ret;
19256 }
__noswap_vqdmulhq_s16(int16x8_t __p0,int16x8_t __p1)19257 __ai int16x8_t __noswap_vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19258   int16x8_t __ret;
19259   __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19260   return __ret;
19261 }
19262 #endif
19263 
19264 #ifdef __LITTLE_ENDIAN__
vqdmulh_s32(int32x2_t __p0,int32x2_t __p1)19265 __ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19266   int32x2_t __ret;
19267   __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19268   return __ret;
19269 }
19270 #else
vqdmulh_s32(int32x2_t __p0,int32x2_t __p1)19271 __ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19272   int32x2_t __ret;
19273   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19274   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19275   __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
19276   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19277   return __ret;
19278 }
__noswap_vqdmulh_s32(int32x2_t __p0,int32x2_t __p1)19279 __ai int32x2_t __noswap_vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19280   int32x2_t __ret;
19281   __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19282   return __ret;
19283 }
19284 #endif
19285 
19286 #ifdef __LITTLE_ENDIAN__
vqdmulh_s16(int16x4_t __p0,int16x4_t __p1)19287 __ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19288   int16x4_t __ret;
19289   __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19290   return __ret;
19291 }
19292 #else
vqdmulh_s16(int16x4_t __p0,int16x4_t __p1)19293 __ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19294   int16x4_t __ret;
19295   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19296   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19297   __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
19298   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19299   return __ret;
19300 }
__noswap_vqdmulh_s16(int16x4_t __p0,int16x4_t __p1)19301 __ai int16x4_t __noswap_vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19302   int16x4_t __ret;
19303   __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19304   return __ret;
19305 }
19306 #endif
19307 
19308 #ifdef __LITTLE_ENDIAN__
vqdmulhq_n_s32(int32x4_t __p0,int32_t __p1)19309 __ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
19310   int32x4_t __ret;
19311   __ret = vqdmulhq_s32(__p0, (int32x4_t) {__p1, __p1, __p1, __p1});
19312   return __ret;
19313 }
19314 #else
vqdmulhq_n_s32(int32x4_t __p0,int32_t __p1)19315 __ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
19316   int32x4_t __ret;
19317   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19318   __ret = __noswap_vqdmulhq_s32(__rev0, (int32x4_t) {__p1, __p1, __p1, __p1});
19319   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19320   return __ret;
19321 }
19322 #endif
19323 
19324 #ifdef __LITTLE_ENDIAN__
vqdmulhq_n_s16(int16x8_t __p0,int16_t __p1)19325 __ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
19326   int16x8_t __ret;
19327   __ret = vqdmulhq_s16(__p0, (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1});
19328   return __ret;
19329 }
19330 #else
vqdmulhq_n_s16(int16x8_t __p0,int16_t __p1)19331 __ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
19332   int16x8_t __ret;
19333   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19334   __ret = __noswap_vqdmulhq_s16(__rev0, (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1});
19335   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19336   return __ret;
19337 }
19338 #endif
19339 
19340 #ifdef __LITTLE_ENDIAN__
vqdmulh_n_s32(int32x2_t __p0,int32_t __p1)19341 __ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
19342   int32x2_t __ret;
19343   __ret = vqdmulh_s32(__p0, (int32x2_t) {__p1, __p1});
19344   return __ret;
19345 }
19346 #else
vqdmulh_n_s32(int32x2_t __p0,int32_t __p1)19347 __ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
19348   int32x2_t __ret;
19349   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19350   __ret = __noswap_vqdmulh_s32(__rev0, (int32x2_t) {__p1, __p1});
19351   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19352   return __ret;
19353 }
19354 #endif
19355 
19356 #ifdef __LITTLE_ENDIAN__
vqdmulh_n_s16(int16x4_t __p0,int16_t __p1)19357 __ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
19358   int16x4_t __ret;
19359   __ret = vqdmulh_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
19360   return __ret;
19361 }
19362 #else
vqdmulh_n_s16(int16x4_t __p0,int16_t __p1)19363 __ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
19364   int16x4_t __ret;
19365   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19366   __ret = __noswap_vqdmulh_s16(__rev0, (int16x4_t) {__p1, __p1, __p1, __p1});
19367   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19368   return __ret;
19369 }
19370 #endif
19371 
19372 #ifdef __LITTLE_ENDIAN__
vqdmull_s32(int32x2_t __p0,int32x2_t __p1)19373 __ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
19374   int64x2_t __ret;
19375   __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
19376   return __ret;
19377 }
19378 #else
vqdmull_s32(int32x2_t __p0,int32x2_t __p1)19379 __ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
19380   int64x2_t __ret;
19381   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19382   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19383   __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
19384   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19385   return __ret;
19386 }
__noswap_vqdmull_s32(int32x2_t __p0,int32x2_t __p1)19387 __ai int64x2_t __noswap_vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
19388   int64x2_t __ret;
19389   __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
19390   return __ret;
19391 }
19392 #endif
19393 
19394 #ifdef __LITTLE_ENDIAN__
vqdmull_s16(int16x4_t __p0,int16x4_t __p1)19395 __ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
19396   int32x4_t __ret;
19397   __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
19398   return __ret;
19399 }
19400 #else
vqdmull_s16(int16x4_t __p0,int16x4_t __p1)19401 __ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
19402   int32x4_t __ret;
19403   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19404   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19405   __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
19406   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19407   return __ret;
19408 }
__noswap_vqdmull_s16(int16x4_t __p0,int16x4_t __p1)19409 __ai int32x4_t __noswap_vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
19410   int32x4_t __ret;
19411   __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
19412   return __ret;
19413 }
19414 #endif
19415 
19416 #ifdef __LITTLE_ENDIAN__
19417 #define vqdmull_lane_s32(__p0_122, __p1_122, __p2_122) __extension__ ({ \
19418   int64x2_t __ret_122; \
19419   int32x2_t __s0_122 = __p0_122; \
19420   int32x2_t __s1_122 = __p1_122; \
19421   __ret_122 = vqdmull_s32(__s0_122, splat_lane_s32(__s1_122, __p2_122)); \
19422   __ret_122; \
19423 })
19424 #else
19425 #define vqdmull_lane_s32(__p0_123, __p1_123, __p2_123) __extension__ ({ \
19426   int64x2_t __ret_123; \
19427   int32x2_t __s0_123 = __p0_123; \
19428   int32x2_t __s1_123 = __p1_123; \
19429   int32x2_t __rev0_123;  __rev0_123 = __builtin_shufflevector(__s0_123, __s0_123, 1, 0); \
19430   int32x2_t __rev1_123;  __rev1_123 = __builtin_shufflevector(__s1_123, __s1_123, 1, 0); \
19431   __ret_123 = __noswap_vqdmull_s32(__rev0_123, __noswap_splat_lane_s32(__rev1_123, __p2_123)); \
19432   __ret_123 = __builtin_shufflevector(__ret_123, __ret_123, 1, 0); \
19433   __ret_123; \
19434 })
19435 #endif
19436 
19437 #ifdef __LITTLE_ENDIAN__
19438 #define vqdmull_lane_s16(__p0_124, __p1_124, __p2_124) __extension__ ({ \
19439   int32x4_t __ret_124; \
19440   int16x4_t __s0_124 = __p0_124; \
19441   int16x4_t __s1_124 = __p1_124; \
19442   __ret_124 = vqdmull_s16(__s0_124, splat_lane_s16(__s1_124, __p2_124)); \
19443   __ret_124; \
19444 })
19445 #else
19446 #define vqdmull_lane_s16(__p0_125, __p1_125, __p2_125) __extension__ ({ \
19447   int32x4_t __ret_125; \
19448   int16x4_t __s0_125 = __p0_125; \
19449   int16x4_t __s1_125 = __p1_125; \
19450   int16x4_t __rev0_125;  __rev0_125 = __builtin_shufflevector(__s0_125, __s0_125, 3, 2, 1, 0); \
19451   int16x4_t __rev1_125;  __rev1_125 = __builtin_shufflevector(__s1_125, __s1_125, 3, 2, 1, 0); \
19452   __ret_125 = __noswap_vqdmull_s16(__rev0_125, __noswap_splat_lane_s16(__rev1_125, __p2_125)); \
19453   __ret_125 = __builtin_shufflevector(__ret_125, __ret_125, 3, 2, 1, 0); \
19454   __ret_125; \
19455 })
19456 #endif
19457 
19458 #ifdef __LITTLE_ENDIAN__
vqdmull_n_s32(int32x2_t __p0,int32_t __p1)19459 __ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
19460   int64x2_t __ret;
19461   __ret = vqdmull_s32(__p0, (int32x2_t) {__p1, __p1});
19462   return __ret;
19463 }
19464 #else
vqdmull_n_s32(int32x2_t __p0,int32_t __p1)19465 __ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
19466   int64x2_t __ret;
19467   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19468   __ret = __noswap_vqdmull_s32(__rev0, (int32x2_t) {__p1, __p1});
19469   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19470   return __ret;
19471 }
__noswap_vqdmull_n_s32(int32x2_t __p0,int32_t __p1)19472 __ai int64x2_t __noswap_vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
19473   int64x2_t __ret;
19474   __ret = __noswap_vqdmull_s32(__p0, (int32x2_t) {__p1, __p1});
19475   return __ret;
19476 }
19477 #endif
19478 
19479 #ifdef __LITTLE_ENDIAN__
vqdmull_n_s16(int16x4_t __p0,int16_t __p1)19480 __ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
19481   int32x4_t __ret;
19482   __ret = vqdmull_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
19483   return __ret;
19484 }
19485 #else
vqdmull_n_s16(int16x4_t __p0,int16_t __p1)19486 __ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
19487   int32x4_t __ret;
19488   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19489   __ret = __noswap_vqdmull_s16(__rev0, (int16x4_t) {__p1, __p1, __p1, __p1});
19490   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19491   return __ret;
19492 }
__noswap_vqdmull_n_s16(int16x4_t __p0,int16_t __p1)19493 __ai int32x4_t __noswap_vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
19494   int32x4_t __ret;
19495   __ret = __noswap_vqdmull_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
19496   return __ret;
19497 }
19498 #endif
19499 
19500 #ifdef __LITTLE_ENDIAN__
vqmovn_u32(uint32x4_t __p0)19501 __ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
19502   uint16x4_t __ret;
19503   __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
19504   return __ret;
19505 }
19506 #else
vqmovn_u32(uint32x4_t __p0)19507 __ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
19508   uint16x4_t __ret;
19509   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19510   __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 17);
19511   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19512   return __ret;
19513 }
__noswap_vqmovn_u32(uint32x4_t __p0)19514 __ai uint16x4_t __noswap_vqmovn_u32(uint32x4_t __p0) {
19515   uint16x4_t __ret;
19516   __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
19517   return __ret;
19518 }
19519 #endif
19520 
19521 #ifdef __LITTLE_ENDIAN__
vqmovn_u64(uint64x2_t __p0)19522 __ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
19523   uint32x2_t __ret;
19524   __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
19525   return __ret;
19526 }
19527 #else
vqmovn_u64(uint64x2_t __p0)19528 __ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
19529   uint32x2_t __ret;
19530   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19531   __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 18);
19532   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19533   return __ret;
19534 }
__noswap_vqmovn_u64(uint64x2_t __p0)19535 __ai uint32x2_t __noswap_vqmovn_u64(uint64x2_t __p0) {
19536   uint32x2_t __ret;
19537   __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
19538   return __ret;
19539 }
19540 #endif
19541 
19542 #ifdef __LITTLE_ENDIAN__
vqmovn_u16(uint16x8_t __p0)19543 __ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
19544   uint8x8_t __ret;
19545   __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
19546   return __ret;
19547 }
19548 #else
vqmovn_u16(uint16x8_t __p0)19549 __ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
19550   uint8x8_t __ret;
19551   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19552   __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 16);
19553   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19554   return __ret;
19555 }
__noswap_vqmovn_u16(uint16x8_t __p0)19556 __ai uint8x8_t __noswap_vqmovn_u16(uint16x8_t __p0) {
19557   uint8x8_t __ret;
19558   __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
19559   return __ret;
19560 }
19561 #endif
19562 
19563 #ifdef __LITTLE_ENDIAN__
vqmovn_s32(int32x4_t __p0)19564 __ai int16x4_t vqmovn_s32(int32x4_t __p0) {
19565   int16x4_t __ret;
19566   __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
19567   return __ret;
19568 }
19569 #else
vqmovn_s32(int32x4_t __p0)19570 __ai int16x4_t vqmovn_s32(int32x4_t __p0) {
19571   int16x4_t __ret;
19572   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19573   __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 1);
19574   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19575   return __ret;
19576 }
__noswap_vqmovn_s32(int32x4_t __p0)19577 __ai int16x4_t __noswap_vqmovn_s32(int32x4_t __p0) {
19578   int16x4_t __ret;
19579   __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
19580   return __ret;
19581 }
19582 #endif
19583 
19584 #ifdef __LITTLE_ENDIAN__
vqmovn_s64(int64x2_t __p0)19585 __ai int32x2_t vqmovn_s64(int64x2_t __p0) {
19586   int32x2_t __ret;
19587   __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
19588   return __ret;
19589 }
19590 #else
vqmovn_s64(int64x2_t __p0)19591 __ai int32x2_t vqmovn_s64(int64x2_t __p0) {
19592   int32x2_t __ret;
19593   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19594   __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 2);
19595   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19596   return __ret;
19597 }
__noswap_vqmovn_s64(int64x2_t __p0)19598 __ai int32x2_t __noswap_vqmovn_s64(int64x2_t __p0) {
19599   int32x2_t __ret;
19600   __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
19601   return __ret;
19602 }
19603 #endif
19604 
19605 #ifdef __LITTLE_ENDIAN__
vqmovn_s16(int16x8_t __p0)19606 __ai int8x8_t vqmovn_s16(int16x8_t __p0) {
19607   int8x8_t __ret;
19608   __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
19609   return __ret;
19610 }
19611 #else
vqmovn_s16(int16x8_t __p0)19612 __ai int8x8_t vqmovn_s16(int16x8_t __p0) {
19613   int8x8_t __ret;
19614   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19615   __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 0);
19616   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19617   return __ret;
19618 }
__noswap_vqmovn_s16(int16x8_t __p0)19619 __ai int8x8_t __noswap_vqmovn_s16(int16x8_t __p0) {
19620   int8x8_t __ret;
19621   __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
19622   return __ret;
19623 }
19624 #endif
19625 
19626 #ifdef __LITTLE_ENDIAN__
vqmovun_s32(int32x4_t __p0)19627 __ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
19628   uint16x4_t __ret;
19629   __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
19630   return __ret;
19631 }
19632 #else
vqmovun_s32(int32x4_t __p0)19633 __ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
19634   uint16x4_t __ret;
19635   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19636   __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 17);
19637   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19638   return __ret;
19639 }
__noswap_vqmovun_s32(int32x4_t __p0)19640 __ai uint16x4_t __noswap_vqmovun_s32(int32x4_t __p0) {
19641   uint16x4_t __ret;
19642   __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
19643   return __ret;
19644 }
19645 #endif
19646 
19647 #ifdef __LITTLE_ENDIAN__
vqmovun_s64(int64x2_t __p0)19648 __ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
19649   uint32x2_t __ret;
19650   __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
19651   return __ret;
19652 }
19653 #else
vqmovun_s64(int64x2_t __p0)19654 __ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
19655   uint32x2_t __ret;
19656   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19657   __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 18);
19658   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19659   return __ret;
19660 }
__noswap_vqmovun_s64(int64x2_t __p0)19661 __ai uint32x2_t __noswap_vqmovun_s64(int64x2_t __p0) {
19662   uint32x2_t __ret;
19663   __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
19664   return __ret;
19665 }
19666 #endif
19667 
19668 #ifdef __LITTLE_ENDIAN__
vqmovun_s16(int16x8_t __p0)19669 __ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
19670   uint8x8_t __ret;
19671   __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
19672   return __ret;
19673 }
19674 #else
vqmovun_s16(int16x8_t __p0)19675 __ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
19676   uint8x8_t __ret;
19677   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19678   __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 16);
19679   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19680   return __ret;
19681 }
__noswap_vqmovun_s16(int16x8_t __p0)19682 __ai uint8x8_t __noswap_vqmovun_s16(int16x8_t __p0) {
19683   uint8x8_t __ret;
19684   __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
19685   return __ret;
19686 }
19687 #endif
19688 
19689 #ifdef __LITTLE_ENDIAN__
vqnegq_s8(int8x16_t __p0)19690 __ai int8x16_t vqnegq_s8(int8x16_t __p0) {
19691   int8x16_t __ret;
19692   __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 32);
19693   return __ret;
19694 }
19695 #else
vqnegq_s8(int8x16_t __p0)19696 __ai int8x16_t vqnegq_s8(int8x16_t __p0) {
19697   int8x16_t __ret;
19698   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19699   __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 32);
19700   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19701   return __ret;
19702 }
19703 #endif
19704 
19705 #ifdef __LITTLE_ENDIAN__
vqnegq_s32(int32x4_t __p0)19706 __ai int32x4_t vqnegq_s32(int32x4_t __p0) {
19707   int32x4_t __ret;
19708   __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 34);
19709   return __ret;
19710 }
19711 #else
vqnegq_s32(int32x4_t __p0)19712 __ai int32x4_t vqnegq_s32(int32x4_t __p0) {
19713   int32x4_t __ret;
19714   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19715   __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 34);
19716   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19717   return __ret;
19718 }
19719 #endif
19720 
19721 #ifdef __LITTLE_ENDIAN__
vqnegq_s16(int16x8_t __p0)19722 __ai int16x8_t vqnegq_s16(int16x8_t __p0) {
19723   int16x8_t __ret;
19724   __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 33);
19725   return __ret;
19726 }
19727 #else
vqnegq_s16(int16x8_t __p0)19728 __ai int16x8_t vqnegq_s16(int16x8_t __p0) {
19729   int16x8_t __ret;
19730   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19731   __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 33);
19732   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19733   return __ret;
19734 }
19735 #endif
19736 
19737 #ifdef __LITTLE_ENDIAN__
vqneg_s8(int8x8_t __p0)19738 __ai int8x8_t vqneg_s8(int8x8_t __p0) {
19739   int8x8_t __ret;
19740   __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 0);
19741   return __ret;
19742 }
19743 #else
vqneg_s8(int8x8_t __p0)19744 __ai int8x8_t vqneg_s8(int8x8_t __p0) {
19745   int8x8_t __ret;
19746   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19747   __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 0);
19748   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19749   return __ret;
19750 }
19751 #endif
19752 
19753 #ifdef __LITTLE_ENDIAN__
vqneg_s32(int32x2_t __p0)19754 __ai int32x2_t vqneg_s32(int32x2_t __p0) {
19755   int32x2_t __ret;
19756   __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 2);
19757   return __ret;
19758 }
19759 #else
vqneg_s32(int32x2_t __p0)19760 __ai int32x2_t vqneg_s32(int32x2_t __p0) {
19761   int32x2_t __ret;
19762   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19763   __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 2);
19764   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19765   return __ret;
19766 }
19767 #endif
19768 
19769 #ifdef __LITTLE_ENDIAN__
vqneg_s16(int16x4_t __p0)19770 __ai int16x4_t vqneg_s16(int16x4_t __p0) {
19771   int16x4_t __ret;
19772   __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 1);
19773   return __ret;
19774 }
19775 #else
vqneg_s16(int16x4_t __p0)19776 __ai int16x4_t vqneg_s16(int16x4_t __p0) {
19777   int16x4_t __ret;
19778   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19779   __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 1);
19780   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19781   return __ret;
19782 }
19783 #endif
19784 
19785 #ifdef __LITTLE_ENDIAN__
vqrdmulhq_s32(int32x4_t __p0,int32x4_t __p1)19786 __ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19787   int32x4_t __ret;
19788   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19789   return __ret;
19790 }
19791 #else
vqrdmulhq_s32(int32x4_t __p0,int32x4_t __p1)19792 __ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19793   int32x4_t __ret;
19794   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19795   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19796   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
19797   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19798   return __ret;
19799 }
__noswap_vqrdmulhq_s32(int32x4_t __p0,int32x4_t __p1)19800 __ai int32x4_t __noswap_vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19801   int32x4_t __ret;
19802   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19803   return __ret;
19804 }
19805 #endif
19806 
19807 #ifdef __LITTLE_ENDIAN__
vqrdmulhq_s16(int16x8_t __p0,int16x8_t __p1)19808 __ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19809   int16x8_t __ret;
19810   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19811   return __ret;
19812 }
19813 #else
vqrdmulhq_s16(int16x8_t __p0,int16x8_t __p1)19814 __ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19815   int16x8_t __ret;
19816   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19817   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19818   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
19819   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19820   return __ret;
19821 }
__noswap_vqrdmulhq_s16(int16x8_t __p0,int16x8_t __p1)19822 __ai int16x8_t __noswap_vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19823   int16x8_t __ret;
19824   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19825   return __ret;
19826 }
19827 #endif
19828 
19829 #ifdef __LITTLE_ENDIAN__
vqrdmulh_s32(int32x2_t __p0,int32x2_t __p1)19830 __ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19831   int32x2_t __ret;
19832   __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19833   return __ret;
19834 }
19835 #else
vqrdmulh_s32(int32x2_t __p0,int32x2_t __p1)19836 __ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19837   int32x2_t __ret;
19838   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19839   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19840   __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
19841   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19842   return __ret;
19843 }
__noswap_vqrdmulh_s32(int32x2_t __p0,int32x2_t __p1)19844 __ai int32x2_t __noswap_vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19845   int32x2_t __ret;
19846   __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19847   return __ret;
19848 }
19849 #endif
19850 
19851 #ifdef __LITTLE_ENDIAN__
vqrdmulh_s16(int16x4_t __p0,int16x4_t __p1)19852 __ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19853   int16x4_t __ret;
19854   __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19855   return __ret;
19856 }
19857 #else
vqrdmulh_s16(int16x4_t __p0,int16x4_t __p1)19858 __ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19859   int16x4_t __ret;
19860   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19861   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19862   __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
19863   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19864   return __ret;
19865 }
__noswap_vqrdmulh_s16(int16x4_t __p0,int16x4_t __p1)19866 __ai int16x4_t __noswap_vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19867   int16x4_t __ret;
19868   __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19869   return __ret;
19870 }
19871 #endif
19872 
19873 #ifdef __LITTLE_ENDIAN__
vqrdmulhq_n_s32(int32x4_t __p0,int32_t __p1)19874 __ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
19875   int32x4_t __ret;
19876   __ret = vqrdmulhq_s32(__p0, (int32x4_t) {__p1, __p1, __p1, __p1});
19877   return __ret;
19878 }
19879 #else
vqrdmulhq_n_s32(int32x4_t __p0,int32_t __p1)19880 __ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
19881   int32x4_t __ret;
19882   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19883   __ret = __noswap_vqrdmulhq_s32(__rev0, (int32x4_t) {__p1, __p1, __p1, __p1});
19884   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19885   return __ret;
19886 }
19887 #endif
19888 
19889 #ifdef __LITTLE_ENDIAN__
vqrdmulhq_n_s16(int16x8_t __p0,int16_t __p1)19890 __ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
19891   int16x8_t __ret;
19892   __ret = vqrdmulhq_s16(__p0, (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1});
19893   return __ret;
19894 }
19895 #else
vqrdmulhq_n_s16(int16x8_t __p0,int16_t __p1)19896 __ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
19897   int16x8_t __ret;
19898   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19899   __ret = __noswap_vqrdmulhq_s16(__rev0, (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1});
19900   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19901   return __ret;
19902 }
19903 #endif
19904 
19905 #ifdef __LITTLE_ENDIAN__
vqrdmulh_n_s32(int32x2_t __p0,int32_t __p1)19906 __ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
19907   int32x2_t __ret;
19908   __ret = vqrdmulh_s32(__p0, (int32x2_t) {__p1, __p1});
19909   return __ret;
19910 }
19911 #else
vqrdmulh_n_s32(int32x2_t __p0,int32_t __p1)19912 __ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
19913   int32x2_t __ret;
19914   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19915   __ret = __noswap_vqrdmulh_s32(__rev0, (int32x2_t) {__p1, __p1});
19916   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19917   return __ret;
19918 }
19919 #endif
19920 
19921 #ifdef __LITTLE_ENDIAN__
vqrdmulh_n_s16(int16x4_t __p0,int16_t __p1)19922 __ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
19923   int16x4_t __ret;
19924   __ret = vqrdmulh_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
19925   return __ret;
19926 }
19927 #else
vqrdmulh_n_s16(int16x4_t __p0,int16_t __p1)19928 __ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
19929   int16x4_t __ret;
19930   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19931   __ret = __noswap_vqrdmulh_s16(__rev0, (int16x4_t) {__p1, __p1, __p1, __p1});
19932   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19933   return __ret;
19934 }
19935 #endif
19936 
19937 #ifdef __LITTLE_ENDIAN__
vqrshlq_u8(uint8x16_t __p0,int8x16_t __p1)19938 __ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
19939   uint8x16_t __ret;
19940   __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
19941   return __ret;
19942 }
19943 #else
vqrshlq_u8(uint8x16_t __p0,int8x16_t __p1)19944 __ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
19945   uint8x16_t __ret;
19946   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19947   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19948   __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
19949   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19950   return __ret;
19951 }
19952 #endif
19953 
19954 #ifdef __LITTLE_ENDIAN__
vqrshlq_u32(uint32x4_t __p0,int32x4_t __p1)19955 __ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
19956   uint32x4_t __ret;
19957   __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
19958   return __ret;
19959 }
19960 #else
vqrshlq_u32(uint32x4_t __p0,int32x4_t __p1)19961 __ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
19962   uint32x4_t __ret;
19963   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19964   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19965   __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
19966   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19967   return __ret;
19968 }
19969 #endif
19970 
19971 #ifdef __LITTLE_ENDIAN__
vqrshlq_u64(uint64x2_t __p0,int64x2_t __p1)19972 __ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
19973   uint64x2_t __ret;
19974   __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
19975   return __ret;
19976 }
19977 #else
vqrshlq_u64(uint64x2_t __p0,int64x2_t __p1)19978 __ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
19979   uint64x2_t __ret;
19980   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19981   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19982   __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
19983   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19984   return __ret;
19985 }
19986 #endif
19987 
19988 #ifdef __LITTLE_ENDIAN__
vqrshlq_u16(uint16x8_t __p0,int16x8_t __p1)19989 __ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
19990   uint16x8_t __ret;
19991   __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
19992   return __ret;
19993 }
19994 #else
vqrshlq_u16(uint16x8_t __p0,int16x8_t __p1)19995 __ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
19996   uint16x8_t __ret;
19997   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19998   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19999   __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20000   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20001   return __ret;
20002 }
20003 #endif
20004 
20005 #ifdef __LITTLE_ENDIAN__
vqrshlq_s8(int8x16_t __p0,int8x16_t __p1)20006 __ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20007   int8x16_t __ret;
20008   __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20009   return __ret;
20010 }
20011 #else
vqrshlq_s8(int8x16_t __p0,int8x16_t __p1)20012 __ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20013   int8x16_t __ret;
20014   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20015   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20016   __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20017   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20018   return __ret;
20019 }
20020 #endif
20021 
20022 #ifdef __LITTLE_ENDIAN__
vqrshlq_s32(int32x4_t __p0,int32x4_t __p1)20023 __ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20024   int32x4_t __ret;
20025   __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20026   return __ret;
20027 }
20028 #else
vqrshlq_s32(int32x4_t __p0,int32x4_t __p1)20029 __ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20030   int32x4_t __ret;
20031   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20032   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20033   __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20034   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20035   return __ret;
20036 }
20037 #endif
20038 
20039 #ifdef __LITTLE_ENDIAN__
vqrshlq_s64(int64x2_t __p0,int64x2_t __p1)20040 __ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20041   int64x2_t __ret;
20042   __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
20043   return __ret;
20044 }
20045 #else
vqrshlq_s64(int64x2_t __p0,int64x2_t __p1)20046 __ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20047   int64x2_t __ret;
20048   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20049   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20050   __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
20051   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20052   return __ret;
20053 }
20054 #endif
20055 
20056 #ifdef __LITTLE_ENDIAN__
vqrshlq_s16(int16x8_t __p0,int16x8_t __p1)20057 __ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20058   int16x8_t __ret;
20059   __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
20060   return __ret;
20061 }
20062 #else
vqrshlq_s16(int16x8_t __p0,int16x8_t __p1)20063 __ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20064   int16x8_t __ret;
20065   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20066   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20067   __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
20068   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20069   return __ret;
20070 }
20071 #endif
20072 
20073 #ifdef __LITTLE_ENDIAN__
vqrshl_u8(uint8x8_t __p0,int8x8_t __p1)20074 __ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20075   uint8x8_t __ret;
20076   __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
20077   return __ret;
20078 }
20079 #else
vqrshl_u8(uint8x8_t __p0,int8x8_t __p1)20080 __ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20081   uint8x8_t __ret;
20082   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20083   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20084   __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
20085   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20086   return __ret;
20087 }
20088 #endif
20089 
20090 #ifdef __LITTLE_ENDIAN__
vqrshl_u32(uint32x2_t __p0,int32x2_t __p1)20091 __ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
20092   uint32x2_t __ret;
20093   __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
20094   return __ret;
20095 }
20096 #else
vqrshl_u32(uint32x2_t __p0,int32x2_t __p1)20097 __ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
20098   uint32x2_t __ret;
20099   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20100   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20101   __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
20102   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20103   return __ret;
20104 }
20105 #endif
20106 
vqrshl_u64(uint64x1_t __p0,int64x1_t __p1)20107 __ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
20108   uint64x1_t __ret;
20109   __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
20110   return __ret;
20111 }
20112 #ifdef __LITTLE_ENDIAN__
vqrshl_u16(uint16x4_t __p0,int16x4_t __p1)20113 __ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
20114   uint16x4_t __ret;
20115   __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
20116   return __ret;
20117 }
20118 #else
vqrshl_u16(uint16x4_t __p0,int16x4_t __p1)20119 __ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
20120   uint16x4_t __ret;
20121   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20122   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20123   __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
20124   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20125   return __ret;
20126 }
20127 #endif
20128 
20129 #ifdef __LITTLE_ENDIAN__
vqrshl_s8(int8x8_t __p0,int8x8_t __p1)20130 __ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
20131   int8x8_t __ret;
20132   __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
20133   return __ret;
20134 }
20135 #else
vqrshl_s8(int8x8_t __p0,int8x8_t __p1)20136 __ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
20137   int8x8_t __ret;
20138   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20139   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20140   __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
20141   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20142   return __ret;
20143 }
20144 #endif
20145 
20146 #ifdef __LITTLE_ENDIAN__
vqrshl_s32(int32x2_t __p0,int32x2_t __p1)20147 __ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
20148   int32x2_t __ret;
20149   __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
20150   return __ret;
20151 }
20152 #else
vqrshl_s32(int32x2_t __p0,int32x2_t __p1)20153 __ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
20154   int32x2_t __ret;
20155   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20156   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20157   __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
20158   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20159   return __ret;
20160 }
20161 #endif
20162 
vqrshl_s64(int64x1_t __p0,int64x1_t __p1)20163 __ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
20164   int64x1_t __ret;
20165   __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
20166   return __ret;
20167 }
20168 #ifdef __LITTLE_ENDIAN__
vqrshl_s16(int16x4_t __p0,int16x4_t __p1)20169 __ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
20170   int16x4_t __ret;
20171   __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
20172   return __ret;
20173 }
20174 #else
vqrshl_s16(int16x4_t __p0,int16x4_t __p1)20175 __ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
20176   int16x4_t __ret;
20177   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20178   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20179   __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
20180   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20181   return __ret;
20182 }
20183 #endif
20184 
20185 #ifdef __LITTLE_ENDIAN__
20186 #define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
20187   uint16x4_t __ret; \
20188   uint32x4_t __s0 = __p0; \
20189   __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
20190   __ret; \
20191 })
20192 #else
20193 #define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
20194   uint16x4_t __ret; \
20195   uint32x4_t __s0 = __p0; \
20196   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20197   __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
20198   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20199   __ret; \
20200 })
20201 #define __noswap_vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
20202   uint16x4_t __ret; \
20203   uint32x4_t __s0 = __p0; \
20204   __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
20205   __ret; \
20206 })
20207 #endif
20208 
20209 #ifdef __LITTLE_ENDIAN__
20210 #define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
20211   uint32x2_t __ret; \
20212   uint64x2_t __s0 = __p0; \
20213   __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
20214   __ret; \
20215 })
20216 #else
20217 #define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
20218   uint32x2_t __ret; \
20219   uint64x2_t __s0 = __p0; \
20220   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20221   __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
20222   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20223   __ret; \
20224 })
20225 #define __noswap_vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
20226   uint32x2_t __ret; \
20227   uint64x2_t __s0 = __p0; \
20228   __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
20229   __ret; \
20230 })
20231 #endif
20232 
20233 #ifdef __LITTLE_ENDIAN__
20234 #define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
20235   uint8x8_t __ret; \
20236   uint16x8_t __s0 = __p0; \
20237   __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
20238   __ret; \
20239 })
20240 #else
20241 #define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
20242   uint8x8_t __ret; \
20243   uint16x8_t __s0 = __p0; \
20244   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20245   __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
20246   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20247   __ret; \
20248 })
20249 #define __noswap_vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
20250   uint8x8_t __ret; \
20251   uint16x8_t __s0 = __p0; \
20252   __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
20253   __ret; \
20254 })
20255 #endif
20256 
20257 #ifdef __LITTLE_ENDIAN__
20258 #define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
20259   int16x4_t __ret; \
20260   int32x4_t __s0 = __p0; \
20261   __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
20262   __ret; \
20263 })
20264 #else
20265 #define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
20266   int16x4_t __ret; \
20267   int32x4_t __s0 = __p0; \
20268   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20269   __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
20270   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20271   __ret; \
20272 })
20273 #define __noswap_vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
20274   int16x4_t __ret; \
20275   int32x4_t __s0 = __p0; \
20276   __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
20277   __ret; \
20278 })
20279 #endif
20280 
20281 #ifdef __LITTLE_ENDIAN__
20282 #define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
20283   int32x2_t __ret; \
20284   int64x2_t __s0 = __p0; \
20285   __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
20286   __ret; \
20287 })
20288 #else
20289 #define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
20290   int32x2_t __ret; \
20291   int64x2_t __s0 = __p0; \
20292   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20293   __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
20294   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20295   __ret; \
20296 })
20297 #define __noswap_vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
20298   int32x2_t __ret; \
20299   int64x2_t __s0 = __p0; \
20300   __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
20301   __ret; \
20302 })
20303 #endif
20304 
20305 #ifdef __LITTLE_ENDIAN__
20306 #define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
20307   int8x8_t __ret; \
20308   int16x8_t __s0 = __p0; \
20309   __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
20310   __ret; \
20311 })
20312 #else
20313 #define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
20314   int8x8_t __ret; \
20315   int16x8_t __s0 = __p0; \
20316   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20317   __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
20318   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20319   __ret; \
20320 })
20321 #define __noswap_vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
20322   int8x8_t __ret; \
20323   int16x8_t __s0 = __p0; \
20324   __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
20325   __ret; \
20326 })
20327 #endif
20328 
20329 #ifdef __LITTLE_ENDIAN__
20330 #define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
20331   uint16x4_t __ret; \
20332   int32x4_t __s0 = __p0; \
20333   __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
20334   __ret; \
20335 })
20336 #else
20337 #define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
20338   uint16x4_t __ret; \
20339   int32x4_t __s0 = __p0; \
20340   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20341   __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 17); \
20342   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20343   __ret; \
20344 })
20345 #define __noswap_vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
20346   uint16x4_t __ret; \
20347   int32x4_t __s0 = __p0; \
20348   __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
20349   __ret; \
20350 })
20351 #endif
20352 
20353 #ifdef __LITTLE_ENDIAN__
20354 #define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
20355   uint32x2_t __ret; \
20356   int64x2_t __s0 = __p0; \
20357   __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
20358   __ret; \
20359 })
20360 #else
20361 #define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
20362   uint32x2_t __ret; \
20363   int64x2_t __s0 = __p0; \
20364   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20365   __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 18); \
20366   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20367   __ret; \
20368 })
20369 #define __noswap_vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
20370   uint32x2_t __ret; \
20371   int64x2_t __s0 = __p0; \
20372   __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
20373   __ret; \
20374 })
20375 #endif
20376 
20377 #ifdef __LITTLE_ENDIAN__
20378 #define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
20379   uint8x8_t __ret; \
20380   int16x8_t __s0 = __p0; \
20381   __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
20382   __ret; \
20383 })
20384 #else
20385 #define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
20386   uint8x8_t __ret; \
20387   int16x8_t __s0 = __p0; \
20388   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20389   __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 16); \
20390   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20391   __ret; \
20392 })
20393 #define __noswap_vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
20394   uint8x8_t __ret; \
20395   int16x8_t __s0 = __p0; \
20396   __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
20397   __ret; \
20398 })
20399 #endif
20400 
20401 #ifdef __LITTLE_ENDIAN__
vqshlq_u8(uint8x16_t __p0,int8x16_t __p1)20402 __ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20403   uint8x16_t __ret;
20404   __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20405   return __ret;
20406 }
20407 #else
vqshlq_u8(uint8x16_t __p0,int8x16_t __p1)20408 __ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20409   uint8x16_t __ret;
20410   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20411   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20412   __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20413   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20414   return __ret;
20415 }
20416 #endif
20417 
20418 #ifdef __LITTLE_ENDIAN__
vqshlq_u32(uint32x4_t __p0,int32x4_t __p1)20419 __ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20420   uint32x4_t __ret;
20421   __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20422   return __ret;
20423 }
20424 #else
vqshlq_u32(uint32x4_t __p0,int32x4_t __p1)20425 __ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20426   uint32x4_t __ret;
20427   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20428   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20429   __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20430   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20431   return __ret;
20432 }
20433 #endif
20434 
20435 #ifdef __LITTLE_ENDIAN__
vqshlq_u64(uint64x2_t __p0,int64x2_t __p1)20436 __ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20437   uint64x2_t __ret;
20438   __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
20439   return __ret;
20440 }
20441 #else
vqshlq_u64(uint64x2_t __p0,int64x2_t __p1)20442 __ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20443   uint64x2_t __ret;
20444   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20445   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20446   __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
20447   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20448   return __ret;
20449 }
20450 #endif
20451 
20452 #ifdef __LITTLE_ENDIAN__
vqshlq_u16(uint16x8_t __p0,int16x8_t __p1)20453 __ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20454   uint16x8_t __ret;
20455   __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20456   return __ret;
20457 }
20458 #else
vqshlq_u16(uint16x8_t __p0,int16x8_t __p1)20459 __ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20460   uint16x8_t __ret;
20461   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20462   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20463   __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20464   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20465   return __ret;
20466 }
20467 #endif
20468 
20469 #ifdef __LITTLE_ENDIAN__
vqshlq_s8(int8x16_t __p0,int8x16_t __p1)20470 __ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20471   int8x16_t __ret;
20472   __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20473   return __ret;
20474 }
20475 #else
vqshlq_s8(int8x16_t __p0,int8x16_t __p1)20476 __ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20477   int8x16_t __ret;
20478   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20479   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20480   __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20481   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20482   return __ret;
20483 }
20484 #endif
20485 
20486 #ifdef __LITTLE_ENDIAN__
vqshlq_s32(int32x4_t __p0,int32x4_t __p1)20487 __ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20488   int32x4_t __ret;
20489   __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20490   return __ret;
20491 }
20492 #else
vqshlq_s32(int32x4_t __p0,int32x4_t __p1)20493 __ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20494   int32x4_t __ret;
20495   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20496   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20497   __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20498   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20499   return __ret;
20500 }
20501 #endif
20502 
20503 #ifdef __LITTLE_ENDIAN__
vqshlq_s64(int64x2_t __p0,int64x2_t __p1)20504 __ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20505   int64x2_t __ret;
20506   __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
20507   return __ret;
20508 }
20509 #else
vqshlq_s64(int64x2_t __p0,int64x2_t __p1)20510 __ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20511   int64x2_t __ret;
20512   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20513   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20514   __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
20515   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20516   return __ret;
20517 }
20518 #endif
20519 
20520 #ifdef __LITTLE_ENDIAN__
vqshlq_s16(int16x8_t __p0,int16x8_t __p1)20521 __ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20522   int16x8_t __ret;
20523   __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
20524   return __ret;
20525 }
20526 #else
vqshlq_s16(int16x8_t __p0,int16x8_t __p1)20527 __ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20528   int16x8_t __ret;
20529   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20530   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20531   __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
20532   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20533   return __ret;
20534 }
20535 #endif
20536 
20537 #ifdef __LITTLE_ENDIAN__
vqshl_u8(uint8x8_t __p0,int8x8_t __p1)20538 __ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20539   uint8x8_t __ret;
20540   __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
20541   return __ret;
20542 }
20543 #else
vqshl_u8(uint8x8_t __p0,int8x8_t __p1)20544 __ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20545   uint8x8_t __ret;
20546   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20547   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20548   __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
20549   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20550   return __ret;
20551 }
20552 #endif
20553 
20554 #ifdef __LITTLE_ENDIAN__
vqshl_u32(uint32x2_t __p0,int32x2_t __p1)20555 __ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
20556   uint32x2_t __ret;
20557   __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
20558   return __ret;
20559 }
20560 #else
vqshl_u32(uint32x2_t __p0,int32x2_t __p1)20561 __ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
20562   uint32x2_t __ret;
20563   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20564   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20565   __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
20566   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20567   return __ret;
20568 }
20569 #endif
20570 
vqshl_u64(uint64x1_t __p0,int64x1_t __p1)20571 __ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
20572   uint64x1_t __ret;
20573   __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
20574   return __ret;
20575 }
20576 #ifdef __LITTLE_ENDIAN__
vqshl_u16(uint16x4_t __p0,int16x4_t __p1)20577 __ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
20578   uint16x4_t __ret;
20579   __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
20580   return __ret;
20581 }
20582 #else
vqshl_u16(uint16x4_t __p0,int16x4_t __p1)20583 __ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
20584   uint16x4_t __ret;
20585   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20586   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20587   __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
20588   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20589   return __ret;
20590 }
20591 #endif
20592 
20593 #ifdef __LITTLE_ENDIAN__
vqshl_s8(int8x8_t __p0,int8x8_t __p1)20594 __ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
20595   int8x8_t __ret;
20596   __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
20597   return __ret;
20598 }
20599 #else
vqshl_s8(int8x8_t __p0,int8x8_t __p1)20600 __ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
20601   int8x8_t __ret;
20602   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20603   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20604   __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
20605   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20606   return __ret;
20607 }
20608 #endif
20609 
20610 #ifdef __LITTLE_ENDIAN__
vqshl_s32(int32x2_t __p0,int32x2_t __p1)20611 __ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
20612   int32x2_t __ret;
20613   __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
20614   return __ret;
20615 }
20616 #else
vqshl_s32(int32x2_t __p0,int32x2_t __p1)20617 __ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
20618   int32x2_t __ret;
20619   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20620   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20621   __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
20622   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20623   return __ret;
20624 }
20625 #endif
20626 
vqshl_s64(int64x1_t __p0,int64x1_t __p1)20627 __ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
20628   int64x1_t __ret;
20629   __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
20630   return __ret;
20631 }
20632 #ifdef __LITTLE_ENDIAN__
vqshl_s16(int16x4_t __p0,int16x4_t __p1)20633 __ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
20634   int16x4_t __ret;
20635   __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
20636   return __ret;
20637 }
20638 #else
vqshl_s16(int16x4_t __p0,int16x4_t __p1)20639 __ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
20640   int16x4_t __ret;
20641   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20642   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20643   __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
20644   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20645   return __ret;
20646 }
20647 #endif
20648 
20649 #ifdef __LITTLE_ENDIAN__
20650 #define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
20651   uint8x16_t __ret; \
20652   uint8x16_t __s0 = __p0; \
20653   __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 48); \
20654   __ret; \
20655 })
20656 #else
20657 #define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
20658   uint8x16_t __ret; \
20659   uint8x16_t __s0 = __p0; \
20660   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20661   __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 48); \
20662   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20663   __ret; \
20664 })
20665 #endif
20666 
20667 #ifdef __LITTLE_ENDIAN__
20668 #define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
20669   uint32x4_t __ret; \
20670   uint32x4_t __s0 = __p0; \
20671   __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 50); \
20672   __ret; \
20673 })
20674 #else
20675 #define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
20676   uint32x4_t __ret; \
20677   uint32x4_t __s0 = __p0; \
20678   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20679   __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 50); \
20680   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20681   __ret; \
20682 })
20683 #endif
20684 
20685 #ifdef __LITTLE_ENDIAN__
20686 #define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
20687   uint64x2_t __ret; \
20688   uint64x2_t __s0 = __p0; \
20689   __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 51); \
20690   __ret; \
20691 })
20692 #else
20693 #define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
20694   uint64x2_t __ret; \
20695   uint64x2_t __s0 = __p0; \
20696   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20697   __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 51); \
20698   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20699   __ret; \
20700 })
20701 #endif
20702 
20703 #ifdef __LITTLE_ENDIAN__
20704 #define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
20705   uint16x8_t __ret; \
20706   uint16x8_t __s0 = __p0; \
20707   __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 49); \
20708   __ret; \
20709 })
20710 #else
20711 #define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
20712   uint16x8_t __ret; \
20713   uint16x8_t __s0 = __p0; \
20714   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20715   __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 49); \
20716   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20717   __ret; \
20718 })
20719 #endif
20720 
20721 #ifdef __LITTLE_ENDIAN__
20722 #define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
20723   int8x16_t __ret; \
20724   int8x16_t __s0 = __p0; \
20725   __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 32); \
20726   __ret; \
20727 })
20728 #else
20729 #define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
20730   int8x16_t __ret; \
20731   int8x16_t __s0 = __p0; \
20732   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20733   __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 32); \
20734   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20735   __ret; \
20736 })
20737 #endif
20738 
20739 #ifdef __LITTLE_ENDIAN__
20740 #define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
20741   int32x4_t __ret; \
20742   int32x4_t __s0 = __p0; \
20743   __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 34); \
20744   __ret; \
20745 })
20746 #else
20747 #define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
20748   int32x4_t __ret; \
20749   int32x4_t __s0 = __p0; \
20750   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20751   __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 34); \
20752   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20753   __ret; \
20754 })
20755 #endif
20756 
20757 #ifdef __LITTLE_ENDIAN__
20758 #define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
20759   int64x2_t __ret; \
20760   int64x2_t __s0 = __p0; \
20761   __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 35); \
20762   __ret; \
20763 })
20764 #else
20765 #define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
20766   int64x2_t __ret; \
20767   int64x2_t __s0 = __p0; \
20768   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20769   __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 35); \
20770   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20771   __ret; \
20772 })
20773 #endif
20774 
20775 #ifdef __LITTLE_ENDIAN__
20776 #define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
20777   int16x8_t __ret; \
20778   int16x8_t __s0 = __p0; \
20779   __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 33); \
20780   __ret; \
20781 })
20782 #else
20783 #define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
20784   int16x8_t __ret; \
20785   int16x8_t __s0 = __p0; \
20786   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20787   __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 33); \
20788   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20789   __ret; \
20790 })
20791 #endif
20792 
20793 #ifdef __LITTLE_ENDIAN__
20794 #define vqshl_n_u8(__p0, __p1) __extension__ ({ \
20795   uint8x8_t __ret; \
20796   uint8x8_t __s0 = __p0; \
20797   __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 16); \
20798   __ret; \
20799 })
20800 #else
20801 #define vqshl_n_u8(__p0, __p1) __extension__ ({ \
20802   uint8x8_t __ret; \
20803   uint8x8_t __s0 = __p0; \
20804   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20805   __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 16); \
20806   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20807   __ret; \
20808 })
20809 #endif
20810 
20811 #ifdef __LITTLE_ENDIAN__
20812 #define vqshl_n_u32(__p0, __p1) __extension__ ({ \
20813   uint32x2_t __ret; \
20814   uint32x2_t __s0 = __p0; \
20815   __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 18); \
20816   __ret; \
20817 })
20818 #else
20819 #define vqshl_n_u32(__p0, __p1) __extension__ ({ \
20820   uint32x2_t __ret; \
20821   uint32x2_t __s0 = __p0; \
20822   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20823   __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 18); \
20824   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20825   __ret; \
20826 })
20827 #endif
20828 
20829 #define vqshl_n_u64(__p0, __p1) __extension__ ({ \
20830   uint64x1_t __ret; \
20831   uint64x1_t __s0 = __p0; \
20832   __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
20833   __ret; \
20834 })
20835 #ifdef __LITTLE_ENDIAN__
20836 #define vqshl_n_u16(__p0, __p1) __extension__ ({ \
20837   uint16x4_t __ret; \
20838   uint16x4_t __s0 = __p0; \
20839   __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 17); \
20840   __ret; \
20841 })
20842 #else
20843 #define vqshl_n_u16(__p0, __p1) __extension__ ({ \
20844   uint16x4_t __ret; \
20845   uint16x4_t __s0 = __p0; \
20846   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20847   __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 17); \
20848   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20849   __ret; \
20850 })
20851 #endif
20852 
20853 #ifdef __LITTLE_ENDIAN__
20854 #define vqshl_n_s8(__p0, __p1) __extension__ ({ \
20855   int8x8_t __ret; \
20856   int8x8_t __s0 = __p0; \
20857   __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 0); \
20858   __ret; \
20859 })
20860 #else
20861 #define vqshl_n_s8(__p0, __p1) __extension__ ({ \
20862   int8x8_t __ret; \
20863   int8x8_t __s0 = __p0; \
20864   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20865   __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 0); \
20866   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20867   __ret; \
20868 })
20869 #endif
20870 
20871 #ifdef __LITTLE_ENDIAN__
20872 #define vqshl_n_s32(__p0, __p1) __extension__ ({ \
20873   int32x2_t __ret; \
20874   int32x2_t __s0 = __p0; \
20875   __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 2); \
20876   __ret; \
20877 })
20878 #else
20879 #define vqshl_n_s32(__p0, __p1) __extension__ ({ \
20880   int32x2_t __ret; \
20881   int32x2_t __s0 = __p0; \
20882   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20883   __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 2); \
20884   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20885   __ret; \
20886 })
20887 #endif
20888 
20889 #define vqshl_n_s64(__p0, __p1) __extension__ ({ \
20890   int64x1_t __ret; \
20891   int64x1_t __s0 = __p0; \
20892   __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
20893   __ret; \
20894 })
20895 #ifdef __LITTLE_ENDIAN__
20896 #define vqshl_n_s16(__p0, __p1) __extension__ ({ \
20897   int16x4_t __ret; \
20898   int16x4_t __s0 = __p0; \
20899   __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 1); \
20900   __ret; \
20901 })
20902 #else
20903 #define vqshl_n_s16(__p0, __p1) __extension__ ({ \
20904   int16x4_t __ret; \
20905   int16x4_t __s0 = __p0; \
20906   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20907   __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 1); \
20908   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20909   __ret; \
20910 })
20911 #endif
20912 
20913 #ifdef __LITTLE_ENDIAN__
20914 #define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
20915   uint8x16_t __ret; \
20916   int8x16_t __s0 = __p0; \
20917   __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 48); \
20918   __ret; \
20919 })
20920 #else
20921 #define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
20922   uint8x16_t __ret; \
20923   int8x16_t __s0 = __p0; \
20924   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20925   __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 48); \
20926   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20927   __ret; \
20928 })
20929 #endif
20930 
20931 #ifdef __LITTLE_ENDIAN__
20932 #define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
20933   uint32x4_t __ret; \
20934   int32x4_t __s0 = __p0; \
20935   __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 50); \
20936   __ret; \
20937 })
20938 #else
20939 #define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
20940   uint32x4_t __ret; \
20941   int32x4_t __s0 = __p0; \
20942   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20943   __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 50); \
20944   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20945   __ret; \
20946 })
20947 #endif
20948 
20949 #ifdef __LITTLE_ENDIAN__
20950 #define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
20951   uint64x2_t __ret; \
20952   int64x2_t __s0 = __p0; \
20953   __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 51); \
20954   __ret; \
20955 })
20956 #else
20957 #define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
20958   uint64x2_t __ret; \
20959   int64x2_t __s0 = __p0; \
20960   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20961   __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 51); \
20962   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20963   __ret; \
20964 })
20965 #endif
20966 
20967 #ifdef __LITTLE_ENDIAN__
20968 #define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
20969   uint16x8_t __ret; \
20970   int16x8_t __s0 = __p0; \
20971   __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 49); \
20972   __ret; \
20973 })
20974 #else
20975 #define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
20976   uint16x8_t __ret; \
20977   int16x8_t __s0 = __p0; \
20978   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20979   __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 49); \
20980   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20981   __ret; \
20982 })
20983 #endif
20984 
20985 #ifdef __LITTLE_ENDIAN__
20986 #define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
20987   uint8x8_t __ret; \
20988   int8x8_t __s0 = __p0; \
20989   __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 16); \
20990   __ret; \
20991 })
20992 #else
20993 #define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
20994   uint8x8_t __ret; \
20995   int8x8_t __s0 = __p0; \
20996   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20997   __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 16); \
20998   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20999   __ret; \
21000 })
21001 #endif
21002 
21003 #ifdef __LITTLE_ENDIAN__
21004 #define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
21005   uint32x2_t __ret; \
21006   int32x2_t __s0 = __p0; \
21007   __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 18); \
21008   __ret; \
21009 })
21010 #else
21011 #define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
21012   uint32x2_t __ret; \
21013   int32x2_t __s0 = __p0; \
21014   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21015   __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 18); \
21016   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21017   __ret; \
21018 })
21019 #endif
21020 
21021 #define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
21022   uint64x1_t __ret; \
21023   int64x1_t __s0 = __p0; \
21024   __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
21025   __ret; \
21026 })
21027 #ifdef __LITTLE_ENDIAN__
21028 #define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
21029   uint16x4_t __ret; \
21030   int16x4_t __s0 = __p0; \
21031   __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 17); \
21032   __ret; \
21033 })
21034 #else
21035 #define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
21036   uint16x4_t __ret; \
21037   int16x4_t __s0 = __p0; \
21038   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21039   __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 17); \
21040   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21041   __ret; \
21042 })
21043 #endif
21044 
21045 #ifdef __LITTLE_ENDIAN__
21046 #define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
21047   uint16x4_t __ret; \
21048   uint32x4_t __s0 = __p0; \
21049   __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
21050   __ret; \
21051 })
21052 #else
21053 #define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
21054   uint16x4_t __ret; \
21055   uint32x4_t __s0 = __p0; \
21056   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21057   __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 17); \
21058   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21059   __ret; \
21060 })
21061 #define __noswap_vqshrn_n_u32(__p0, __p1) __extension__ ({ \
21062   uint16x4_t __ret; \
21063   uint32x4_t __s0 = __p0; \
21064   __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
21065   __ret; \
21066 })
21067 #endif
21068 
21069 #ifdef __LITTLE_ENDIAN__
21070 #define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
21071   uint32x2_t __ret; \
21072   uint64x2_t __s0 = __p0; \
21073   __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
21074   __ret; \
21075 })
21076 #else
21077 #define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
21078   uint32x2_t __ret; \
21079   uint64x2_t __s0 = __p0; \
21080   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21081   __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 18); \
21082   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21083   __ret; \
21084 })
21085 #define __noswap_vqshrn_n_u64(__p0, __p1) __extension__ ({ \
21086   uint32x2_t __ret; \
21087   uint64x2_t __s0 = __p0; \
21088   __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
21089   __ret; \
21090 })
21091 #endif
21092 
21093 #ifdef __LITTLE_ENDIAN__
21094 #define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
21095   uint8x8_t __ret; \
21096   uint16x8_t __s0 = __p0; \
21097   __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
21098   __ret; \
21099 })
21100 #else
21101 #define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
21102   uint8x8_t __ret; \
21103   uint16x8_t __s0 = __p0; \
21104   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21105   __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 16); \
21106   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21107   __ret; \
21108 })
21109 #define __noswap_vqshrn_n_u16(__p0, __p1) __extension__ ({ \
21110   uint8x8_t __ret; \
21111   uint16x8_t __s0 = __p0; \
21112   __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
21113   __ret; \
21114 })
21115 #endif
21116 
21117 #ifdef __LITTLE_ENDIAN__
21118 #define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
21119   int16x4_t __ret; \
21120   int32x4_t __s0 = __p0; \
21121   __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
21122   __ret; \
21123 })
21124 #else
21125 #define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
21126   int16x4_t __ret; \
21127   int32x4_t __s0 = __p0; \
21128   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21129   __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 1); \
21130   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21131   __ret; \
21132 })
21133 #define __noswap_vqshrn_n_s32(__p0, __p1) __extension__ ({ \
21134   int16x4_t __ret; \
21135   int32x4_t __s0 = __p0; \
21136   __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
21137   __ret; \
21138 })
21139 #endif
21140 
21141 #ifdef __LITTLE_ENDIAN__
21142 #define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
21143   int32x2_t __ret; \
21144   int64x2_t __s0 = __p0; \
21145   __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
21146   __ret; \
21147 })
21148 #else
21149 #define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
21150   int32x2_t __ret; \
21151   int64x2_t __s0 = __p0; \
21152   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21153   __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 2); \
21154   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21155   __ret; \
21156 })
21157 #define __noswap_vqshrn_n_s64(__p0, __p1) __extension__ ({ \
21158   int32x2_t __ret; \
21159   int64x2_t __s0 = __p0; \
21160   __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
21161   __ret; \
21162 })
21163 #endif
21164 
21165 #ifdef __LITTLE_ENDIAN__
21166 #define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
21167   int8x8_t __ret; \
21168   int16x8_t __s0 = __p0; \
21169   __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
21170   __ret; \
21171 })
21172 #else
21173 #define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
21174   int8x8_t __ret; \
21175   int16x8_t __s0 = __p0; \
21176   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21177   __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 0); \
21178   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21179   __ret; \
21180 })
21181 #define __noswap_vqshrn_n_s16(__p0, __p1) __extension__ ({ \
21182   int8x8_t __ret; \
21183   int16x8_t __s0 = __p0; \
21184   __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
21185   __ret; \
21186 })
21187 #endif
21188 
21189 #ifdef __LITTLE_ENDIAN__
21190 #define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
21191   uint16x4_t __ret; \
21192   int32x4_t __s0 = __p0; \
21193   __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
21194   __ret; \
21195 })
21196 #else
21197 #define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
21198   uint16x4_t __ret; \
21199   int32x4_t __s0 = __p0; \
21200   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21201   __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 17); \
21202   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21203   __ret; \
21204 })
21205 #define __noswap_vqshrun_n_s32(__p0, __p1) __extension__ ({ \
21206   uint16x4_t __ret; \
21207   int32x4_t __s0 = __p0; \
21208   __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
21209   __ret; \
21210 })
21211 #endif
21212 
21213 #ifdef __LITTLE_ENDIAN__
21214 #define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
21215   uint32x2_t __ret; \
21216   int64x2_t __s0 = __p0; \
21217   __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
21218   __ret; \
21219 })
21220 #else
21221 #define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
21222   uint32x2_t __ret; \
21223   int64x2_t __s0 = __p0; \
21224   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21225   __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 18); \
21226   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21227   __ret; \
21228 })
21229 #define __noswap_vqshrun_n_s64(__p0, __p1) __extension__ ({ \
21230   uint32x2_t __ret; \
21231   int64x2_t __s0 = __p0; \
21232   __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
21233   __ret; \
21234 })
21235 #endif
21236 
21237 #ifdef __LITTLE_ENDIAN__
21238 #define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
21239   uint8x8_t __ret; \
21240   int16x8_t __s0 = __p0; \
21241   __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
21242   __ret; \
21243 })
21244 #else
21245 #define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
21246   uint8x8_t __ret; \
21247   int16x8_t __s0 = __p0; \
21248   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21249   __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 16); \
21250   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21251   __ret; \
21252 })
21253 #define __noswap_vqshrun_n_s16(__p0, __p1) __extension__ ({ \
21254   uint8x8_t __ret; \
21255   int16x8_t __s0 = __p0; \
21256   __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
21257   __ret; \
21258 })
21259 #endif
21260 
21261 #ifdef __LITTLE_ENDIAN__
vqsubq_u8(uint8x16_t __p0,uint8x16_t __p1)21262 __ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
21263   uint8x16_t __ret;
21264   __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
21265   return __ret;
21266 }
21267 #else
vqsubq_u8(uint8x16_t __p0,uint8x16_t __p1)21268 __ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
21269   uint8x16_t __ret;
21270   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21271   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21272   __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
21273   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21274   return __ret;
21275 }
21276 #endif
21277 
21278 #ifdef __LITTLE_ENDIAN__
vqsubq_u32(uint32x4_t __p0,uint32x4_t __p1)21279 __ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
21280   uint32x4_t __ret;
21281   __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
21282   return __ret;
21283 }
21284 #else
vqsubq_u32(uint32x4_t __p0,uint32x4_t __p1)21285 __ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
21286   uint32x4_t __ret;
21287   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21288   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21289   __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
21290   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21291   return __ret;
21292 }
21293 #endif
21294 
21295 #ifdef __LITTLE_ENDIAN__
vqsubq_u64(uint64x2_t __p0,uint64x2_t __p1)21296 __ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
21297   uint64x2_t __ret;
21298   __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
21299   return __ret;
21300 }
21301 #else
vqsubq_u64(uint64x2_t __p0,uint64x2_t __p1)21302 __ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
21303   uint64x2_t __ret;
21304   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21305   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21306   __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
21307   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21308   return __ret;
21309 }
21310 #endif
21311 
21312 #ifdef __LITTLE_ENDIAN__
vqsubq_u16(uint16x8_t __p0,uint16x8_t __p1)21313 __ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
21314   uint16x8_t __ret;
21315   __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
21316   return __ret;
21317 }
21318 #else
vqsubq_u16(uint16x8_t __p0,uint16x8_t __p1)21319 __ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
21320   uint16x8_t __ret;
21321   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21322   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21323   __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
21324   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21325   return __ret;
21326 }
21327 #endif
21328 
21329 #ifdef __LITTLE_ENDIAN__
vqsubq_s8(int8x16_t __p0,int8x16_t __p1)21330 __ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
21331   int8x16_t __ret;
21332   __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
21333   return __ret;
21334 }
21335 #else
vqsubq_s8(int8x16_t __p0,int8x16_t __p1)21336 __ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
21337   int8x16_t __ret;
21338   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21339   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21340   __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
21341   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21342   return __ret;
21343 }
21344 #endif
21345 
21346 #ifdef __LITTLE_ENDIAN__
vqsubq_s32(int32x4_t __p0,int32x4_t __p1)21347 __ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
21348   int32x4_t __ret;
21349   __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
21350   return __ret;
21351 }
21352 #else
vqsubq_s32(int32x4_t __p0,int32x4_t __p1)21353 __ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
21354   int32x4_t __ret;
21355   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21356   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21357   __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
21358   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21359   return __ret;
21360 }
21361 #endif
21362 
21363 #ifdef __LITTLE_ENDIAN__
vqsubq_s64(int64x2_t __p0,int64x2_t __p1)21364 __ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
21365   int64x2_t __ret;
21366   __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
21367   return __ret;
21368 }
21369 #else
vqsubq_s64(int64x2_t __p0,int64x2_t __p1)21370 __ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
21371   int64x2_t __ret;
21372   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21373   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21374   __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
21375   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21376   return __ret;
21377 }
21378 #endif
21379 
21380 #ifdef __LITTLE_ENDIAN__
vqsubq_s16(int16x8_t __p0,int16x8_t __p1)21381 __ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
21382   int16x8_t __ret;
21383   __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
21384   return __ret;
21385 }
21386 #else
vqsubq_s16(int16x8_t __p0,int16x8_t __p1)21387 __ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
21388   int16x8_t __ret;
21389   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21390   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21391   __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
21392   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21393   return __ret;
21394 }
21395 #endif
21396 
21397 #ifdef __LITTLE_ENDIAN__
vqsub_u8(uint8x8_t __p0,uint8x8_t __p1)21398 __ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
21399   uint8x8_t __ret;
21400   __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
21401   return __ret;
21402 }
21403 #else
vqsub_u8(uint8x8_t __p0,uint8x8_t __p1)21404 __ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
21405   uint8x8_t __ret;
21406   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21407   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21408   __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
21409   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21410   return __ret;
21411 }
21412 #endif
21413 
21414 #ifdef __LITTLE_ENDIAN__
vqsub_u32(uint32x2_t __p0,uint32x2_t __p1)21415 __ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
21416   uint32x2_t __ret;
21417   __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
21418   return __ret;
21419 }
21420 #else
vqsub_u32(uint32x2_t __p0,uint32x2_t __p1)21421 __ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
21422   uint32x2_t __ret;
21423   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21424   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21425   __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
21426   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21427   return __ret;
21428 }
21429 #endif
21430 
vqsub_u64(uint64x1_t __p0,uint64x1_t __p1)21431 __ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
21432   uint64x1_t __ret;
21433   __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21434   return __ret;
21435 }
21436 #ifdef __LITTLE_ENDIAN__
vqsub_u16(uint16x4_t __p0,uint16x4_t __p1)21437 __ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
21438   uint16x4_t __ret;
21439   __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
21440   return __ret;
21441 }
21442 #else
vqsub_u16(uint16x4_t __p0,uint16x4_t __p1)21443 __ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
21444   uint16x4_t __ret;
21445   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21446   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21447   __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
21448   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21449   return __ret;
21450 }
21451 #endif
21452 
21453 #ifdef __LITTLE_ENDIAN__
vqsub_s8(int8x8_t __p0,int8x8_t __p1)21454 __ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
21455   int8x8_t __ret;
21456   __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
21457   return __ret;
21458 }
21459 #else
vqsub_s8(int8x8_t __p0,int8x8_t __p1)21460 __ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
21461   int8x8_t __ret;
21462   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21463   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21464   __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
21465   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21466   return __ret;
21467 }
21468 #endif
21469 
21470 #ifdef __LITTLE_ENDIAN__
vqsub_s32(int32x2_t __p0,int32x2_t __p1)21471 __ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
21472   int32x2_t __ret;
21473   __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
21474   return __ret;
21475 }
21476 #else
vqsub_s32(int32x2_t __p0,int32x2_t __p1)21477 __ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
21478   int32x2_t __ret;
21479   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21480   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21481   __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
21482   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21483   return __ret;
21484 }
21485 #endif
21486 
vqsub_s64(int64x1_t __p0,int64x1_t __p1)21487 __ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
21488   int64x1_t __ret;
21489   __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21490   return __ret;
21491 }
21492 #ifdef __LITTLE_ENDIAN__
vqsub_s16(int16x4_t __p0,int16x4_t __p1)21493 __ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
21494   int16x4_t __ret;
21495   __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
21496   return __ret;
21497 }
21498 #else
vqsub_s16(int16x4_t __p0,int16x4_t __p1)21499 __ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
21500   int16x4_t __ret;
21501   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21502   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21503   __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
21504   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21505   return __ret;
21506 }
21507 #endif
21508 
21509 #ifdef __LITTLE_ENDIAN__
vraddhn_u32(uint32x4_t __p0,uint32x4_t __p1)21510 __ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
21511   uint16x4_t __ret;
21512   __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
21513   return __ret;
21514 }
21515 #else
vraddhn_u32(uint32x4_t __p0,uint32x4_t __p1)21516 __ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
21517   uint16x4_t __ret;
21518   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21519   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21520   __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
21521   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21522   return __ret;
21523 }
__noswap_vraddhn_u32(uint32x4_t __p0,uint32x4_t __p1)21524 __ai uint16x4_t __noswap_vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
21525   uint16x4_t __ret;
21526   __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
21527   return __ret;
21528 }
21529 #endif
21530 
21531 #ifdef __LITTLE_ENDIAN__
vraddhn_u64(uint64x2_t __p0,uint64x2_t __p1)21532 __ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
21533   uint32x2_t __ret;
21534   __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
21535   return __ret;
21536 }
21537 #else
vraddhn_u64(uint64x2_t __p0,uint64x2_t __p1)21538 __ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
21539   uint32x2_t __ret;
21540   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21541   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21542   __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
21543   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21544   return __ret;
21545 }
__noswap_vraddhn_u64(uint64x2_t __p0,uint64x2_t __p1)21546 __ai uint32x2_t __noswap_vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
21547   uint32x2_t __ret;
21548   __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
21549   return __ret;
21550 }
21551 #endif
21552 
21553 #ifdef __LITTLE_ENDIAN__
vraddhn_u16(uint16x8_t __p0,uint16x8_t __p1)21554 __ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
21555   uint8x8_t __ret;
21556   __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
21557   return __ret;
21558 }
21559 #else
vraddhn_u16(uint16x8_t __p0,uint16x8_t __p1)21560 __ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
21561   uint8x8_t __ret;
21562   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21563   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21564   __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
21565   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21566   return __ret;
21567 }
__noswap_vraddhn_u16(uint16x8_t __p0,uint16x8_t __p1)21568 __ai uint8x8_t __noswap_vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
21569   uint8x8_t __ret;
21570   __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
21571   return __ret;
21572 }
21573 #endif
21574 
21575 #ifdef __LITTLE_ENDIAN__
vraddhn_s32(int32x4_t __p0,int32x4_t __p1)21576 __ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
21577   int16x4_t __ret;
21578   __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
21579   return __ret;
21580 }
21581 #else
vraddhn_s32(int32x4_t __p0,int32x4_t __p1)21582 __ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
21583   int16x4_t __ret;
21584   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21585   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21586   __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
21587   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21588   return __ret;
21589 }
__noswap_vraddhn_s32(int32x4_t __p0,int32x4_t __p1)21590 __ai int16x4_t __noswap_vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
21591   int16x4_t __ret;
21592   __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
21593   return __ret;
21594 }
21595 #endif
21596 
21597 #ifdef __LITTLE_ENDIAN__
vraddhn_s64(int64x2_t __p0,int64x2_t __p1)21598 __ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
21599   int32x2_t __ret;
21600   __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
21601   return __ret;
21602 }
21603 #else
vraddhn_s64(int64x2_t __p0,int64x2_t __p1)21604 __ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
21605   int32x2_t __ret;
21606   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21607   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21608   __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
21609   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21610   return __ret;
21611 }
__noswap_vraddhn_s64(int64x2_t __p0,int64x2_t __p1)21612 __ai int32x2_t __noswap_vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
21613   int32x2_t __ret;
21614   __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
21615   return __ret;
21616 }
21617 #endif
21618 
21619 #ifdef __LITTLE_ENDIAN__
vraddhn_s16(int16x8_t __p0,int16x8_t __p1)21620 __ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
21621   int8x8_t __ret;
21622   __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
21623   return __ret;
21624 }
21625 #else
vraddhn_s16(int16x8_t __p0,int16x8_t __p1)21626 __ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
21627   int8x8_t __ret;
21628   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21629   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21630   __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
21631   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21632   return __ret;
21633 }
__noswap_vraddhn_s16(int16x8_t __p0,int16x8_t __p1)21634 __ai int8x8_t __noswap_vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
21635   int8x8_t __ret;
21636   __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
21637   return __ret;
21638 }
21639 #endif
21640 
21641 #ifdef __LITTLE_ENDIAN__
vrecpeq_u32(uint32x4_t __p0)21642 __ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
21643   uint32x4_t __ret;
21644   __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 50);
21645   return __ret;
21646 }
21647 #else
vrecpeq_u32(uint32x4_t __p0)21648 __ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
21649   uint32x4_t __ret;
21650   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21651   __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 50);
21652   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21653   return __ret;
21654 }
21655 #endif
21656 
21657 #ifdef __LITTLE_ENDIAN__
vrecpeq_f32(float32x4_t __p0)21658 __ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
21659   float32x4_t __ret;
21660   __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 41);
21661   return __ret;
21662 }
21663 #else
vrecpeq_f32(float32x4_t __p0)21664 __ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
21665   float32x4_t __ret;
21666   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21667   __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 41);
21668   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21669   return __ret;
21670 }
21671 #endif
21672 
21673 #ifdef __LITTLE_ENDIAN__
vrecpe_u32(uint32x2_t __p0)21674 __ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
21675   uint32x2_t __ret;
21676   __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 18);
21677   return __ret;
21678 }
21679 #else
vrecpe_u32(uint32x2_t __p0)21680 __ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
21681   uint32x2_t __ret;
21682   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21683   __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 18);
21684   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21685   return __ret;
21686 }
21687 #endif
21688 
21689 #ifdef __LITTLE_ENDIAN__
vrecpe_f32(float32x2_t __p0)21690 __ai float32x2_t vrecpe_f32(float32x2_t __p0) {
21691   float32x2_t __ret;
21692   __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 9);
21693   return __ret;
21694 }
21695 #else
vrecpe_f32(float32x2_t __p0)21696 __ai float32x2_t vrecpe_f32(float32x2_t __p0) {
21697   float32x2_t __ret;
21698   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21699   __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 9);
21700   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21701   return __ret;
21702 }
21703 #endif
21704 
21705 #ifdef __LITTLE_ENDIAN__
vrecpsq_f32(float32x4_t __p0,float32x4_t __p1)21706 __ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
21707   float32x4_t __ret;
21708   __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
21709   return __ret;
21710 }
21711 #else
vrecpsq_f32(float32x4_t __p0,float32x4_t __p1)21712 __ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
21713   float32x4_t __ret;
21714   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21715   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21716   __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
21717   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21718   return __ret;
21719 }
21720 #endif
21721 
21722 #ifdef __LITTLE_ENDIAN__
vrecps_f32(float32x2_t __p0,float32x2_t __p1)21723 __ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
21724   float32x2_t __ret;
21725   __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
21726   return __ret;
21727 }
21728 #else
vrecps_f32(float32x2_t __p0,float32x2_t __p1)21729 __ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
21730   float32x2_t __ret;
21731   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21732   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21733   __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
21734   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21735   return __ret;
21736 }
21737 #endif
21738 
21739 #ifdef __LITTLE_ENDIAN__
vrev16_p8(poly8x8_t __p0)21740 __ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
21741   poly8x8_t __ret;
21742   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
21743   return __ret;
21744 }
21745 #else
vrev16_p8(poly8x8_t __p0)21746 __ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
21747   poly8x8_t __ret;
21748   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21749   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
21750   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21751   return __ret;
21752 }
21753 #endif
21754 
21755 #ifdef __LITTLE_ENDIAN__
vrev16q_p8(poly8x16_t __p0)21756 __ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
21757   poly8x16_t __ret;
21758   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21759   return __ret;
21760 }
21761 #else
vrev16q_p8(poly8x16_t __p0)21762 __ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
21763   poly8x16_t __ret;
21764   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21765   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21766   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21767   return __ret;
21768 }
21769 #endif
21770 
21771 #ifdef __LITTLE_ENDIAN__
vrev16q_u8(uint8x16_t __p0)21772 __ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
21773   uint8x16_t __ret;
21774   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21775   return __ret;
21776 }
21777 #else
vrev16q_u8(uint8x16_t __p0)21778 __ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
21779   uint8x16_t __ret;
21780   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21781   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21782   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21783   return __ret;
21784 }
21785 #endif
21786 
21787 #ifdef __LITTLE_ENDIAN__
vrev16q_s8(int8x16_t __p0)21788 __ai int8x16_t vrev16q_s8(int8x16_t __p0) {
21789   int8x16_t __ret;
21790   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21791   return __ret;
21792 }
21793 #else
vrev16q_s8(int8x16_t __p0)21794 __ai int8x16_t vrev16q_s8(int8x16_t __p0) {
21795   int8x16_t __ret;
21796   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21797   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21798   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21799   return __ret;
21800 }
21801 #endif
21802 
21803 #ifdef __LITTLE_ENDIAN__
vrev16_u8(uint8x8_t __p0)21804 __ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
21805   uint8x8_t __ret;
21806   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
21807   return __ret;
21808 }
21809 #else
vrev16_u8(uint8x8_t __p0)21810 __ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
21811   uint8x8_t __ret;
21812   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21813   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
21814   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21815   return __ret;
21816 }
21817 #endif
21818 
21819 #ifdef __LITTLE_ENDIAN__
vrev16_s8(int8x8_t __p0)21820 __ai int8x8_t vrev16_s8(int8x8_t __p0) {
21821   int8x8_t __ret;
21822   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
21823   return __ret;
21824 }
21825 #else
vrev16_s8(int8x8_t __p0)21826 __ai int8x8_t vrev16_s8(int8x8_t __p0) {
21827   int8x8_t __ret;
21828   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21829   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
21830   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21831   return __ret;
21832 }
21833 #endif
21834 
21835 #ifdef __LITTLE_ENDIAN__
vrev32_p8(poly8x8_t __p0)21836 __ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
21837   poly8x8_t __ret;
21838   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
21839   return __ret;
21840 }
21841 #else
vrev32_p8(poly8x8_t __p0)21842 __ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
21843   poly8x8_t __ret;
21844   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21845   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
21846   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21847   return __ret;
21848 }
21849 #endif
21850 
21851 #ifdef __LITTLE_ENDIAN__
vrev32_p16(poly16x4_t __p0)21852 __ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
21853   poly16x4_t __ret;
21854   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
21855   return __ret;
21856 }
21857 #else
vrev32_p16(poly16x4_t __p0)21858 __ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
21859   poly16x4_t __ret;
21860   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21861   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
21862   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21863   return __ret;
21864 }
21865 #endif
21866 
21867 #ifdef __LITTLE_ENDIAN__
vrev32q_p8(poly8x16_t __p0)21868 __ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
21869   poly8x16_t __ret;
21870   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
21871   return __ret;
21872 }
21873 #else
vrev32q_p8(poly8x16_t __p0)21874 __ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
21875   poly8x16_t __ret;
21876   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21877   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
21878   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21879   return __ret;
21880 }
21881 #endif
21882 
21883 #ifdef __LITTLE_ENDIAN__
vrev32q_p16(poly16x8_t __p0)21884 __ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
21885   poly16x8_t __ret;
21886   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
21887   return __ret;
21888 }
21889 #else
vrev32q_p16(poly16x8_t __p0)21890 __ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
21891   poly16x8_t __ret;
21892   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21893   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
21894   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21895   return __ret;
21896 }
21897 #endif
21898 
21899 #ifdef __LITTLE_ENDIAN__
vrev32q_u8(uint8x16_t __p0)21900 __ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
21901   uint8x16_t __ret;
21902   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
21903   return __ret;
21904 }
21905 #else
vrev32q_u8(uint8x16_t __p0)21906 __ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
21907   uint8x16_t __ret;
21908   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21909   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
21910   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21911   return __ret;
21912 }
21913 #endif
21914 
21915 #ifdef __LITTLE_ENDIAN__
vrev32q_u16(uint16x8_t __p0)21916 __ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
21917   uint16x8_t __ret;
21918   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
21919   return __ret;
21920 }
21921 #else
vrev32q_u16(uint16x8_t __p0)21922 __ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
21923   uint16x8_t __ret;
21924   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21925   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
21926   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21927   return __ret;
21928 }
21929 #endif
21930 
21931 #ifdef __LITTLE_ENDIAN__
vrev32q_s8(int8x16_t __p0)21932 __ai int8x16_t vrev32q_s8(int8x16_t __p0) {
21933   int8x16_t __ret;
21934   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
21935   return __ret;
21936 }
21937 #else
vrev32q_s8(int8x16_t __p0)21938 __ai int8x16_t vrev32q_s8(int8x16_t __p0) {
21939   int8x16_t __ret;
21940   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21941   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
21942   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21943   return __ret;
21944 }
21945 #endif
21946 
21947 #ifdef __LITTLE_ENDIAN__
vrev32q_s16(int16x8_t __p0)21948 __ai int16x8_t vrev32q_s16(int16x8_t __p0) {
21949   int16x8_t __ret;
21950   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
21951   return __ret;
21952 }
21953 #else
vrev32q_s16(int16x8_t __p0)21954 __ai int16x8_t vrev32q_s16(int16x8_t __p0) {
21955   int16x8_t __ret;
21956   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21957   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
21958   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21959   return __ret;
21960 }
21961 #endif
21962 
21963 #ifdef __LITTLE_ENDIAN__
vrev32_u8(uint8x8_t __p0)21964 __ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
21965   uint8x8_t __ret;
21966   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
21967   return __ret;
21968 }
21969 #else
vrev32_u8(uint8x8_t __p0)21970 __ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
21971   uint8x8_t __ret;
21972   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21973   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
21974   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21975   return __ret;
21976 }
21977 #endif
21978 
21979 #ifdef __LITTLE_ENDIAN__
vrev32_u16(uint16x4_t __p0)21980 __ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
21981   uint16x4_t __ret;
21982   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
21983   return __ret;
21984 }
21985 #else
vrev32_u16(uint16x4_t __p0)21986 __ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
21987   uint16x4_t __ret;
21988   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21989   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
21990   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21991   return __ret;
21992 }
21993 #endif
21994 
21995 #ifdef __LITTLE_ENDIAN__
vrev32_s8(int8x8_t __p0)21996 __ai int8x8_t vrev32_s8(int8x8_t __p0) {
21997   int8x8_t __ret;
21998   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
21999   return __ret;
22000 }
22001 #else
vrev32_s8(int8x8_t __p0)22002 __ai int8x8_t vrev32_s8(int8x8_t __p0) {
22003   int8x8_t __ret;
22004   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22005   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
22006   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22007   return __ret;
22008 }
22009 #endif
22010 
22011 #ifdef __LITTLE_ENDIAN__
vrev32_s16(int16x4_t __p0)22012 __ai int16x4_t vrev32_s16(int16x4_t __p0) {
22013   int16x4_t __ret;
22014   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22015   return __ret;
22016 }
22017 #else
vrev32_s16(int16x4_t __p0)22018 __ai int16x4_t vrev32_s16(int16x4_t __p0) {
22019   int16x4_t __ret;
22020   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22021   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22022   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22023   return __ret;
22024 }
22025 #endif
22026 
22027 #ifdef __LITTLE_ENDIAN__
vrev64_p8(poly8x8_t __p0)22028 __ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
22029   poly8x8_t __ret;
22030   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22031   return __ret;
22032 }
22033 #else
vrev64_p8(poly8x8_t __p0)22034 __ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
22035   poly8x8_t __ret;
22036   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22037   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
22038   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22039   return __ret;
22040 }
22041 #endif
22042 
22043 #ifdef __LITTLE_ENDIAN__
vrev64_p16(poly16x4_t __p0)22044 __ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
22045   poly16x4_t __ret;
22046   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22047   return __ret;
22048 }
22049 #else
vrev64_p16(poly16x4_t __p0)22050 __ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
22051   poly16x4_t __ret;
22052   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22053   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
22054   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22055   return __ret;
22056 }
22057 #endif
22058 
22059 #ifdef __LITTLE_ENDIAN__
vrev64q_p8(poly8x16_t __p0)22060 __ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
22061   poly8x16_t __ret;
22062   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22063   return __ret;
22064 }
22065 #else
vrev64q_p8(poly8x16_t __p0)22066 __ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
22067   poly8x16_t __ret;
22068   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22069   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22070   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22071   return __ret;
22072 }
22073 #endif
22074 
22075 #ifdef __LITTLE_ENDIAN__
vrev64q_p16(poly16x8_t __p0)22076 __ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
22077   poly16x8_t __ret;
22078   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
22079   return __ret;
22080 }
22081 #else
vrev64q_p16(poly16x8_t __p0)22082 __ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
22083   poly16x8_t __ret;
22084   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22085   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
22086   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22087   return __ret;
22088 }
22089 #endif
22090 
22091 #ifdef __LITTLE_ENDIAN__
vrev64q_u8(uint8x16_t __p0)22092 __ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
22093   uint8x16_t __ret;
22094   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22095   return __ret;
22096 }
22097 #else
vrev64q_u8(uint8x16_t __p0)22098 __ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
22099   uint8x16_t __ret;
22100   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22101   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22102   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22103   return __ret;
22104 }
22105 #endif
22106 
22107 #ifdef __LITTLE_ENDIAN__
vrev64q_u32(uint32x4_t __p0)22108 __ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
22109   uint32x4_t __ret;
22110   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22111   return __ret;
22112 }
22113 #else
vrev64q_u32(uint32x4_t __p0)22114 __ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
22115   uint32x4_t __ret;
22116   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22117   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22118   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22119   return __ret;
22120 }
22121 #endif
22122 
22123 #ifdef __LITTLE_ENDIAN__
vrev64q_u16(uint16x8_t __p0)22124 __ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
22125   uint16x8_t __ret;
22126   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
22127   return __ret;
22128 }
22129 #else
vrev64q_u16(uint16x8_t __p0)22130 __ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
22131   uint16x8_t __ret;
22132   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22133   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
22134   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22135   return __ret;
22136 }
22137 #endif
22138 
22139 #ifdef __LITTLE_ENDIAN__
vrev64q_s8(int8x16_t __p0)22140 __ai int8x16_t vrev64q_s8(int8x16_t __p0) {
22141   int8x16_t __ret;
22142   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22143   return __ret;
22144 }
22145 #else
vrev64q_s8(int8x16_t __p0)22146 __ai int8x16_t vrev64q_s8(int8x16_t __p0) {
22147   int8x16_t __ret;
22148   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22149   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22150   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22151   return __ret;
22152 }
22153 #endif
22154 
22155 #ifdef __LITTLE_ENDIAN__
vrev64q_f32(float32x4_t __p0)22156 __ai float32x4_t vrev64q_f32(float32x4_t __p0) {
22157   float32x4_t __ret;
22158   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22159   return __ret;
22160 }
22161 #else
vrev64q_f32(float32x4_t __p0)22162 __ai float32x4_t vrev64q_f32(float32x4_t __p0) {
22163   float32x4_t __ret;
22164   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22165   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22166   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22167   return __ret;
22168 }
22169 #endif
22170 
22171 #ifdef __LITTLE_ENDIAN__
vrev64q_s32(int32x4_t __p0)22172 __ai int32x4_t vrev64q_s32(int32x4_t __p0) {
22173   int32x4_t __ret;
22174   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22175   return __ret;
22176 }
22177 #else
vrev64q_s32(int32x4_t __p0)22178 __ai int32x4_t vrev64q_s32(int32x4_t __p0) {
22179   int32x4_t __ret;
22180   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22181   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22182   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22183   return __ret;
22184 }
22185 #endif
22186 
22187 #ifdef __LITTLE_ENDIAN__
vrev64q_s16(int16x8_t __p0)22188 __ai int16x8_t vrev64q_s16(int16x8_t __p0) {
22189   int16x8_t __ret;
22190   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
22191   return __ret;
22192 }
22193 #else
vrev64q_s16(int16x8_t __p0)22194 __ai int16x8_t vrev64q_s16(int16x8_t __p0) {
22195   int16x8_t __ret;
22196   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22197   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
22198   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22199   return __ret;
22200 }
22201 #endif
22202 
22203 #ifdef __LITTLE_ENDIAN__
vrev64_u8(uint8x8_t __p0)22204 __ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
22205   uint8x8_t __ret;
22206   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22207   return __ret;
22208 }
22209 #else
vrev64_u8(uint8x8_t __p0)22210 __ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
22211   uint8x8_t __ret;
22212   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22213   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
22214   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22215   return __ret;
22216 }
22217 #endif
22218 
22219 #ifdef __LITTLE_ENDIAN__
vrev64_u32(uint32x2_t __p0)22220 __ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
22221   uint32x2_t __ret;
22222   __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
22223   return __ret;
22224 }
22225 #else
vrev64_u32(uint32x2_t __p0)22226 __ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
22227   uint32x2_t __ret;
22228   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22229   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
22230   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22231   return __ret;
22232 }
22233 #endif
22234 
22235 #ifdef __LITTLE_ENDIAN__
vrev64_u16(uint16x4_t __p0)22236 __ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
22237   uint16x4_t __ret;
22238   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22239   return __ret;
22240 }
22241 #else
vrev64_u16(uint16x4_t __p0)22242 __ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
22243   uint16x4_t __ret;
22244   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22245   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
22246   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22247   return __ret;
22248 }
22249 #endif
22250 
22251 #ifdef __LITTLE_ENDIAN__
vrev64_s8(int8x8_t __p0)22252 __ai int8x8_t vrev64_s8(int8x8_t __p0) {
22253   int8x8_t __ret;
22254   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22255   return __ret;
22256 }
22257 #else
vrev64_s8(int8x8_t __p0)22258 __ai int8x8_t vrev64_s8(int8x8_t __p0) {
22259   int8x8_t __ret;
22260   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22261   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
22262   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22263   return __ret;
22264 }
22265 #endif
22266 
22267 #ifdef __LITTLE_ENDIAN__
vrev64_f32(float32x2_t __p0)22268 __ai float32x2_t vrev64_f32(float32x2_t __p0) {
22269   float32x2_t __ret;
22270   __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
22271   return __ret;
22272 }
22273 #else
vrev64_f32(float32x2_t __p0)22274 __ai float32x2_t vrev64_f32(float32x2_t __p0) {
22275   float32x2_t __ret;
22276   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22277   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
22278   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22279   return __ret;
22280 }
22281 #endif
22282 
22283 #ifdef __LITTLE_ENDIAN__
vrev64_s32(int32x2_t __p0)22284 __ai int32x2_t vrev64_s32(int32x2_t __p0) {
22285   int32x2_t __ret;
22286   __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
22287   return __ret;
22288 }
22289 #else
vrev64_s32(int32x2_t __p0)22290 __ai int32x2_t vrev64_s32(int32x2_t __p0) {
22291   int32x2_t __ret;
22292   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22293   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
22294   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22295   return __ret;
22296 }
22297 #endif
22298 
22299 #ifdef __LITTLE_ENDIAN__
vrev64_s16(int16x4_t __p0)22300 __ai int16x4_t vrev64_s16(int16x4_t __p0) {
22301   int16x4_t __ret;
22302   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22303   return __ret;
22304 }
22305 #else
vrev64_s16(int16x4_t __p0)22306 __ai int16x4_t vrev64_s16(int16x4_t __p0) {
22307   int16x4_t __ret;
22308   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22309   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
22310   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22311   return __ret;
22312 }
22313 #endif
22314 
22315 #ifdef __LITTLE_ENDIAN__
vrhaddq_u8(uint8x16_t __p0,uint8x16_t __p1)22316 __ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
22317   uint8x16_t __ret;
22318   __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
22319   return __ret;
22320 }
22321 #else
vrhaddq_u8(uint8x16_t __p0,uint8x16_t __p1)22322 __ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
22323   uint8x16_t __ret;
22324   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22325   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22326   __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
22327   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22328   return __ret;
22329 }
22330 #endif
22331 
22332 #ifdef __LITTLE_ENDIAN__
vrhaddq_u32(uint32x4_t __p0,uint32x4_t __p1)22333 __ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
22334   uint32x4_t __ret;
22335   __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
22336   return __ret;
22337 }
22338 #else
vrhaddq_u32(uint32x4_t __p0,uint32x4_t __p1)22339 __ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
22340   uint32x4_t __ret;
22341   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22342   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22343   __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
22344   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22345   return __ret;
22346 }
22347 #endif
22348 
22349 #ifdef __LITTLE_ENDIAN__
vrhaddq_u16(uint16x8_t __p0,uint16x8_t __p1)22350 __ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
22351   uint16x8_t __ret;
22352   __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
22353   return __ret;
22354 }
22355 #else
vrhaddq_u16(uint16x8_t __p0,uint16x8_t __p1)22356 __ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
22357   uint16x8_t __ret;
22358   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22359   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22360   __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
22361   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22362   return __ret;
22363 }
22364 #endif
22365 
22366 #ifdef __LITTLE_ENDIAN__
vrhaddq_s8(int8x16_t __p0,int8x16_t __p1)22367 __ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
22368   int8x16_t __ret;
22369   __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
22370   return __ret;
22371 }
22372 #else
vrhaddq_s8(int8x16_t __p0,int8x16_t __p1)22373 __ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
22374   int8x16_t __ret;
22375   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22376   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22377   __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
22378   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22379   return __ret;
22380 }
22381 #endif
22382 
22383 #ifdef __LITTLE_ENDIAN__
vrhaddq_s32(int32x4_t __p0,int32x4_t __p1)22384 __ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
22385   int32x4_t __ret;
22386   __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
22387   return __ret;
22388 }
22389 #else
vrhaddq_s32(int32x4_t __p0,int32x4_t __p1)22390 __ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
22391   int32x4_t __ret;
22392   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22393   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22394   __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
22395   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22396   return __ret;
22397 }
22398 #endif
22399 
22400 #ifdef __LITTLE_ENDIAN__
vrhaddq_s16(int16x8_t __p0,int16x8_t __p1)22401 __ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
22402   int16x8_t __ret;
22403   __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
22404   return __ret;
22405 }
22406 #else
vrhaddq_s16(int16x8_t __p0,int16x8_t __p1)22407 __ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
22408   int16x8_t __ret;
22409   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22410   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22411   __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
22412   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22413   return __ret;
22414 }
22415 #endif
22416 
22417 #ifdef __LITTLE_ENDIAN__
vrhadd_u8(uint8x8_t __p0,uint8x8_t __p1)22418 __ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
22419   uint8x8_t __ret;
22420   __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
22421   return __ret;
22422 }
22423 #else
vrhadd_u8(uint8x8_t __p0,uint8x8_t __p1)22424 __ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
22425   uint8x8_t __ret;
22426   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22427   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22428   __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
22429   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22430   return __ret;
22431 }
22432 #endif
22433 
22434 #ifdef __LITTLE_ENDIAN__
vrhadd_u32(uint32x2_t __p0,uint32x2_t __p1)22435 __ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
22436   uint32x2_t __ret;
22437   __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
22438   return __ret;
22439 }
22440 #else
vrhadd_u32(uint32x2_t __p0,uint32x2_t __p1)22441 __ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
22442   uint32x2_t __ret;
22443   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22444   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22445   __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
22446   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22447   return __ret;
22448 }
22449 #endif
22450 
22451 #ifdef __LITTLE_ENDIAN__
vrhadd_u16(uint16x4_t __p0,uint16x4_t __p1)22452 __ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
22453   uint16x4_t __ret;
22454   __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
22455   return __ret;
22456 }
22457 #else
vrhadd_u16(uint16x4_t __p0,uint16x4_t __p1)22458 __ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
22459   uint16x4_t __ret;
22460   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22461   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22462   __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
22463   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22464   return __ret;
22465 }
22466 #endif
22467 
22468 #ifdef __LITTLE_ENDIAN__
vrhadd_s8(int8x8_t __p0,int8x8_t __p1)22469 __ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
22470   int8x8_t __ret;
22471   __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
22472   return __ret;
22473 }
22474 #else
vrhadd_s8(int8x8_t __p0,int8x8_t __p1)22475 __ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
22476   int8x8_t __ret;
22477   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22478   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22479   __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
22480   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22481   return __ret;
22482 }
22483 #endif
22484 
22485 #ifdef __LITTLE_ENDIAN__
vrhadd_s32(int32x2_t __p0,int32x2_t __p1)22486 __ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
22487   int32x2_t __ret;
22488   __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
22489   return __ret;
22490 }
22491 #else
vrhadd_s32(int32x2_t __p0,int32x2_t __p1)22492 __ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
22493   int32x2_t __ret;
22494   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22495   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22496   __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
22497   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22498   return __ret;
22499 }
22500 #endif
22501 
22502 #ifdef __LITTLE_ENDIAN__
vrhadd_s16(int16x4_t __p0,int16x4_t __p1)22503 __ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
22504   int16x4_t __ret;
22505   __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
22506   return __ret;
22507 }
22508 #else
vrhadd_s16(int16x4_t __p0,int16x4_t __p1)22509 __ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
22510   int16x4_t __ret;
22511   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22512   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22513   __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
22514   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22515   return __ret;
22516 }
22517 #endif
22518 
22519 #ifdef __LITTLE_ENDIAN__
vrshlq_u8(uint8x16_t __p0,int8x16_t __p1)22520 __ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22521   uint8x16_t __ret;
22522   __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
22523   return __ret;
22524 }
22525 #else
vrshlq_u8(uint8x16_t __p0,int8x16_t __p1)22526 __ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22527   uint8x16_t __ret;
22528   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22529   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22530   __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
22531   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22532   return __ret;
22533 }
22534 #endif
22535 
22536 #ifdef __LITTLE_ENDIAN__
vrshlq_u32(uint32x4_t __p0,int32x4_t __p1)22537 __ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22538   uint32x4_t __ret;
22539   __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
22540   return __ret;
22541 }
22542 #else
vrshlq_u32(uint32x4_t __p0,int32x4_t __p1)22543 __ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22544   uint32x4_t __ret;
22545   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22546   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22547   __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
22548   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22549   return __ret;
22550 }
22551 #endif
22552 
22553 #ifdef __LITTLE_ENDIAN__
vrshlq_u64(uint64x2_t __p0,int64x2_t __p1)22554 __ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22555   uint64x2_t __ret;
22556   __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
22557   return __ret;
22558 }
22559 #else
vrshlq_u64(uint64x2_t __p0,int64x2_t __p1)22560 __ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22561   uint64x2_t __ret;
22562   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22563   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22564   __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
22565   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22566   return __ret;
22567 }
22568 #endif
22569 
22570 #ifdef __LITTLE_ENDIAN__
vrshlq_u16(uint16x8_t __p0,int16x8_t __p1)22571 __ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22572   uint16x8_t __ret;
22573   __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
22574   return __ret;
22575 }
22576 #else
vrshlq_u16(uint16x8_t __p0,int16x8_t __p1)22577 __ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22578   uint16x8_t __ret;
22579   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22580   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22581   __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
22582   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22583   return __ret;
22584 }
22585 #endif
22586 
22587 #ifdef __LITTLE_ENDIAN__
vrshlq_s8(int8x16_t __p0,int8x16_t __p1)22588 __ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22589   int8x16_t __ret;
22590   __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
22591   return __ret;
22592 }
22593 #else
vrshlq_s8(int8x16_t __p0,int8x16_t __p1)22594 __ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22595   int8x16_t __ret;
22596   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22597   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22598   __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
22599   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22600   return __ret;
22601 }
22602 #endif
22603 
22604 #ifdef __LITTLE_ENDIAN__
vrshlq_s32(int32x4_t __p0,int32x4_t __p1)22605 __ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22606   int32x4_t __ret;
22607   __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
22608   return __ret;
22609 }
22610 #else
vrshlq_s32(int32x4_t __p0,int32x4_t __p1)22611 __ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22612   int32x4_t __ret;
22613   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22614   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22615   __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
22616   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22617   return __ret;
22618 }
22619 #endif
22620 
22621 #ifdef __LITTLE_ENDIAN__
vrshlq_s64(int64x2_t __p0,int64x2_t __p1)22622 __ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22623   int64x2_t __ret;
22624   __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
22625   return __ret;
22626 }
22627 #else
vrshlq_s64(int64x2_t __p0,int64x2_t __p1)22628 __ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22629   int64x2_t __ret;
22630   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22631   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22632   __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
22633   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22634   return __ret;
22635 }
22636 #endif
22637 
22638 #ifdef __LITTLE_ENDIAN__
vrshlq_s16(int16x8_t __p0,int16x8_t __p1)22639 __ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22640   int16x8_t __ret;
22641   __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
22642   return __ret;
22643 }
22644 #else
vrshlq_s16(int16x8_t __p0,int16x8_t __p1)22645 __ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22646   int16x8_t __ret;
22647   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22648   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22649   __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
22650   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22651   return __ret;
22652 }
22653 #endif
22654 
22655 #ifdef __LITTLE_ENDIAN__
vrshl_u8(uint8x8_t __p0,int8x8_t __p1)22656 __ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22657   uint8x8_t __ret;
22658   __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
22659   return __ret;
22660 }
22661 #else
vrshl_u8(uint8x8_t __p0,int8x8_t __p1)22662 __ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22663   uint8x8_t __ret;
22664   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22665   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22666   __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
22667   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22668   return __ret;
22669 }
22670 #endif
22671 
22672 #ifdef __LITTLE_ENDIAN__
vrshl_u32(uint32x2_t __p0,int32x2_t __p1)22673 __ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22674   uint32x2_t __ret;
22675   __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
22676   return __ret;
22677 }
22678 #else
vrshl_u32(uint32x2_t __p0,int32x2_t __p1)22679 __ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22680   uint32x2_t __ret;
22681   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22682   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22683   __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
22684   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22685   return __ret;
22686 }
22687 #endif
22688 
vrshl_u64(uint64x1_t __p0,int64x1_t __p1)22689 __ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22690   uint64x1_t __ret;
22691   __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22692   return __ret;
22693 }
22694 #ifdef __LITTLE_ENDIAN__
vrshl_u16(uint16x4_t __p0,int16x4_t __p1)22695 __ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22696   uint16x4_t __ret;
22697   __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
22698   return __ret;
22699 }
22700 #else
vrshl_u16(uint16x4_t __p0,int16x4_t __p1)22701 __ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22702   uint16x4_t __ret;
22703   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22704   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22705   __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
22706   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22707   return __ret;
22708 }
22709 #endif
22710 
22711 #ifdef __LITTLE_ENDIAN__
vrshl_s8(int8x8_t __p0,int8x8_t __p1)22712 __ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
22713   int8x8_t __ret;
22714   __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
22715   return __ret;
22716 }
22717 #else
vrshl_s8(int8x8_t __p0,int8x8_t __p1)22718 __ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
22719   int8x8_t __ret;
22720   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22721   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22722   __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
22723   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22724   return __ret;
22725 }
22726 #endif
22727 
22728 #ifdef __LITTLE_ENDIAN__
vrshl_s32(int32x2_t __p0,int32x2_t __p1)22729 __ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
22730   int32x2_t __ret;
22731   __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
22732   return __ret;
22733 }
22734 #else
vrshl_s32(int32x2_t __p0,int32x2_t __p1)22735 __ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
22736   int32x2_t __ret;
22737   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22738   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22739   __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
22740   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22741   return __ret;
22742 }
22743 #endif
22744 
vrshl_s64(int64x1_t __p0,int64x1_t __p1)22745 __ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
22746   int64x1_t __ret;
22747   __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
22748   return __ret;
22749 }
22750 #ifdef __LITTLE_ENDIAN__
vrshl_s16(int16x4_t __p0,int16x4_t __p1)22751 __ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
22752   int16x4_t __ret;
22753   __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
22754   return __ret;
22755 }
22756 #else
vrshl_s16(int16x4_t __p0,int16x4_t __p1)22757 __ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
22758   int16x4_t __ret;
22759   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22760   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22761   __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
22762   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22763   return __ret;
22764 }
22765 #endif
22766 
22767 #ifdef __LITTLE_ENDIAN__
22768 #define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
22769   uint8x16_t __ret; \
22770   uint8x16_t __s0 = __p0; \
22771   __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 48); \
22772   __ret; \
22773 })
22774 #else
22775 #define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
22776   uint8x16_t __ret; \
22777   uint8x16_t __s0 = __p0; \
22778   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22779   __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 48); \
22780   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22781   __ret; \
22782 })
22783 #endif
22784 
22785 #ifdef __LITTLE_ENDIAN__
22786 #define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
22787   uint32x4_t __ret; \
22788   uint32x4_t __s0 = __p0; \
22789   __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 50); \
22790   __ret; \
22791 })
22792 #else
22793 #define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
22794   uint32x4_t __ret; \
22795   uint32x4_t __s0 = __p0; \
22796   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
22797   __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 50); \
22798   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22799   __ret; \
22800 })
22801 #endif
22802 
22803 #ifdef __LITTLE_ENDIAN__
22804 #define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
22805   uint64x2_t __ret; \
22806   uint64x2_t __s0 = __p0; \
22807   __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 51); \
22808   __ret; \
22809 })
22810 #else
22811 #define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
22812   uint64x2_t __ret; \
22813   uint64x2_t __s0 = __p0; \
22814   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
22815   __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 51); \
22816   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22817   __ret; \
22818 })
22819 #endif
22820 
22821 #ifdef __LITTLE_ENDIAN__
22822 #define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
22823   uint16x8_t __ret; \
22824   uint16x8_t __s0 = __p0; \
22825   __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 49); \
22826   __ret; \
22827 })
22828 #else
22829 #define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
22830   uint16x8_t __ret; \
22831   uint16x8_t __s0 = __p0; \
22832   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
22833   __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 49); \
22834   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22835   __ret; \
22836 })
22837 #endif
22838 
22839 #ifdef __LITTLE_ENDIAN__
22840 #define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
22841   int8x16_t __ret; \
22842   int8x16_t __s0 = __p0; \
22843   __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 32); \
22844   __ret; \
22845 })
22846 #else
22847 #define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
22848   int8x16_t __ret; \
22849   int8x16_t __s0 = __p0; \
22850   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22851   __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 32); \
22852   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22853   __ret; \
22854 })
22855 #endif
22856 
22857 #ifdef __LITTLE_ENDIAN__
22858 #define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
22859   int32x4_t __ret; \
22860   int32x4_t __s0 = __p0; \
22861   __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 34); \
22862   __ret; \
22863 })
22864 #else
22865 #define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
22866   int32x4_t __ret; \
22867   int32x4_t __s0 = __p0; \
22868   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
22869   __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 34); \
22870   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22871   __ret; \
22872 })
22873 #endif
22874 
22875 #ifdef __LITTLE_ENDIAN__
22876 #define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
22877   int64x2_t __ret; \
22878   int64x2_t __s0 = __p0; \
22879   __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 35); \
22880   __ret; \
22881 })
22882 #else
22883 #define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
22884   int64x2_t __ret; \
22885   int64x2_t __s0 = __p0; \
22886   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
22887   __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 35); \
22888   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22889   __ret; \
22890 })
22891 #endif
22892 
22893 #ifdef __LITTLE_ENDIAN__
22894 #define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
22895   int16x8_t __ret; \
22896   int16x8_t __s0 = __p0; \
22897   __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 33); \
22898   __ret; \
22899 })
22900 #else
22901 #define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
22902   int16x8_t __ret; \
22903   int16x8_t __s0 = __p0; \
22904   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
22905   __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 33); \
22906   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22907   __ret; \
22908 })
22909 #endif
22910 
22911 #ifdef __LITTLE_ENDIAN__
22912 #define vrshr_n_u8(__p0, __p1) __extension__ ({ \
22913   uint8x8_t __ret; \
22914   uint8x8_t __s0 = __p0; \
22915   __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 16); \
22916   __ret; \
22917 })
22918 #else
22919 #define vrshr_n_u8(__p0, __p1) __extension__ ({ \
22920   uint8x8_t __ret; \
22921   uint8x8_t __s0 = __p0; \
22922   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
22923   __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 16); \
22924   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22925   __ret; \
22926 })
22927 #endif
22928 
22929 #ifdef __LITTLE_ENDIAN__
22930 #define vrshr_n_u32(__p0, __p1) __extension__ ({ \
22931   uint32x2_t __ret; \
22932   uint32x2_t __s0 = __p0; \
22933   __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 18); \
22934   __ret; \
22935 })
22936 #else
22937 #define vrshr_n_u32(__p0, __p1) __extension__ ({ \
22938   uint32x2_t __ret; \
22939   uint32x2_t __s0 = __p0; \
22940   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
22941   __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 18); \
22942   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22943   __ret; \
22944 })
22945 #endif
22946 
22947 #define vrshr_n_u64(__p0, __p1) __extension__ ({ \
22948   uint64x1_t __ret; \
22949   uint64x1_t __s0 = __p0; \
22950   __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
22951   __ret; \
22952 })
22953 #ifdef __LITTLE_ENDIAN__
22954 #define vrshr_n_u16(__p0, __p1) __extension__ ({ \
22955   uint16x4_t __ret; \
22956   uint16x4_t __s0 = __p0; \
22957   __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 17); \
22958   __ret; \
22959 })
22960 #else
22961 #define vrshr_n_u16(__p0, __p1) __extension__ ({ \
22962   uint16x4_t __ret; \
22963   uint16x4_t __s0 = __p0; \
22964   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
22965   __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 17); \
22966   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22967   __ret; \
22968 })
22969 #endif
22970 
22971 #ifdef __LITTLE_ENDIAN__
22972 #define vrshr_n_s8(__p0, __p1) __extension__ ({ \
22973   int8x8_t __ret; \
22974   int8x8_t __s0 = __p0; \
22975   __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 0); \
22976   __ret; \
22977 })
22978 #else
22979 #define vrshr_n_s8(__p0, __p1) __extension__ ({ \
22980   int8x8_t __ret; \
22981   int8x8_t __s0 = __p0; \
22982   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
22983   __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 0); \
22984   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22985   __ret; \
22986 })
22987 #endif
22988 
22989 #ifdef __LITTLE_ENDIAN__
22990 #define vrshr_n_s32(__p0, __p1) __extension__ ({ \
22991   int32x2_t __ret; \
22992   int32x2_t __s0 = __p0; \
22993   __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 2); \
22994   __ret; \
22995 })
22996 #else
22997 #define vrshr_n_s32(__p0, __p1) __extension__ ({ \
22998   int32x2_t __ret; \
22999   int32x2_t __s0 = __p0; \
23000   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23001   __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 2); \
23002   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23003   __ret; \
23004 })
23005 #endif
23006 
23007 #define vrshr_n_s64(__p0, __p1) __extension__ ({ \
23008   int64x1_t __ret; \
23009   int64x1_t __s0 = __p0; \
23010   __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
23011   __ret; \
23012 })
23013 #ifdef __LITTLE_ENDIAN__
23014 #define vrshr_n_s16(__p0, __p1) __extension__ ({ \
23015   int16x4_t __ret; \
23016   int16x4_t __s0 = __p0; \
23017   __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 1); \
23018   __ret; \
23019 })
23020 #else
23021 #define vrshr_n_s16(__p0, __p1) __extension__ ({ \
23022   int16x4_t __ret; \
23023   int16x4_t __s0 = __p0; \
23024   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23025   __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 1); \
23026   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23027   __ret; \
23028 })
23029 #endif
23030 
23031 #ifdef __LITTLE_ENDIAN__
23032 #define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
23033   uint16x4_t __ret; \
23034   uint32x4_t __s0 = __p0; \
23035   __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
23036   __ret; \
23037 })
23038 #else
23039 #define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
23040   uint16x4_t __ret; \
23041   uint32x4_t __s0 = __p0; \
23042   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23043   __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
23044   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23045   __ret; \
23046 })
23047 #define __noswap_vrshrn_n_u32(__p0, __p1) __extension__ ({ \
23048   uint16x4_t __ret; \
23049   uint32x4_t __s0 = __p0; \
23050   __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
23051   __ret; \
23052 })
23053 #endif
23054 
23055 #ifdef __LITTLE_ENDIAN__
23056 #define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
23057   uint32x2_t __ret; \
23058   uint64x2_t __s0 = __p0; \
23059   __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
23060   __ret; \
23061 })
23062 #else
23063 #define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
23064   uint32x2_t __ret; \
23065   uint64x2_t __s0 = __p0; \
23066   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23067   __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
23068   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23069   __ret; \
23070 })
23071 #define __noswap_vrshrn_n_u64(__p0, __p1) __extension__ ({ \
23072   uint32x2_t __ret; \
23073   uint64x2_t __s0 = __p0; \
23074   __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
23075   __ret; \
23076 })
23077 #endif
23078 
23079 #ifdef __LITTLE_ENDIAN__
23080 #define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
23081   uint8x8_t __ret; \
23082   uint16x8_t __s0 = __p0; \
23083   __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
23084   __ret; \
23085 })
23086 #else
23087 #define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
23088   uint8x8_t __ret; \
23089   uint16x8_t __s0 = __p0; \
23090   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23091   __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
23092   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23093   __ret; \
23094 })
23095 #define __noswap_vrshrn_n_u16(__p0, __p1) __extension__ ({ \
23096   uint8x8_t __ret; \
23097   uint16x8_t __s0 = __p0; \
23098   __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
23099   __ret; \
23100 })
23101 #endif
23102 
23103 #ifdef __LITTLE_ENDIAN__
23104 #define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
23105   int16x4_t __ret; \
23106   int32x4_t __s0 = __p0; \
23107   __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
23108   __ret; \
23109 })
23110 #else
23111 #define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
23112   int16x4_t __ret; \
23113   int32x4_t __s0 = __p0; \
23114   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23115   __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
23116   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23117   __ret; \
23118 })
23119 #define __noswap_vrshrn_n_s32(__p0, __p1) __extension__ ({ \
23120   int16x4_t __ret; \
23121   int32x4_t __s0 = __p0; \
23122   __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
23123   __ret; \
23124 })
23125 #endif
23126 
23127 #ifdef __LITTLE_ENDIAN__
23128 #define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
23129   int32x2_t __ret; \
23130   int64x2_t __s0 = __p0; \
23131   __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
23132   __ret; \
23133 })
23134 #else
23135 #define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
23136   int32x2_t __ret; \
23137   int64x2_t __s0 = __p0; \
23138   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23139   __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
23140   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23141   __ret; \
23142 })
23143 #define __noswap_vrshrn_n_s64(__p0, __p1) __extension__ ({ \
23144   int32x2_t __ret; \
23145   int64x2_t __s0 = __p0; \
23146   __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
23147   __ret; \
23148 })
23149 #endif
23150 
23151 #ifdef __LITTLE_ENDIAN__
23152 #define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
23153   int8x8_t __ret; \
23154   int16x8_t __s0 = __p0; \
23155   __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
23156   __ret; \
23157 })
23158 #else
23159 #define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
23160   int8x8_t __ret; \
23161   int16x8_t __s0 = __p0; \
23162   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23163   __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
23164   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23165   __ret; \
23166 })
23167 #define __noswap_vrshrn_n_s16(__p0, __p1) __extension__ ({ \
23168   int8x8_t __ret; \
23169   int16x8_t __s0 = __p0; \
23170   __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
23171   __ret; \
23172 })
23173 #endif
23174 
23175 #ifdef __LITTLE_ENDIAN__
vrsqrteq_u32(uint32x4_t __p0)23176 __ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
23177   uint32x4_t __ret;
23178   __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 50);
23179   return __ret;
23180 }
23181 #else
vrsqrteq_u32(uint32x4_t __p0)23182 __ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
23183   uint32x4_t __ret;
23184   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23185   __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 50);
23186   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23187   return __ret;
23188 }
23189 #endif
23190 
23191 #ifdef __LITTLE_ENDIAN__
vrsqrteq_f32(float32x4_t __p0)23192 __ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
23193   float32x4_t __ret;
23194   __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 41);
23195   return __ret;
23196 }
23197 #else
vrsqrteq_f32(float32x4_t __p0)23198 __ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
23199   float32x4_t __ret;
23200   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23201   __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 41);
23202   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23203   return __ret;
23204 }
23205 #endif
23206 
23207 #ifdef __LITTLE_ENDIAN__
vrsqrte_u32(uint32x2_t __p0)23208 __ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
23209   uint32x2_t __ret;
23210   __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 18);
23211   return __ret;
23212 }
23213 #else
vrsqrte_u32(uint32x2_t __p0)23214 __ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
23215   uint32x2_t __ret;
23216   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23217   __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 18);
23218   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23219   return __ret;
23220 }
23221 #endif
23222 
23223 #ifdef __LITTLE_ENDIAN__
vrsqrte_f32(float32x2_t __p0)23224 __ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
23225   float32x2_t __ret;
23226   __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 9);
23227   return __ret;
23228 }
23229 #else
vrsqrte_f32(float32x2_t __p0)23230 __ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
23231   float32x2_t __ret;
23232   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23233   __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 9);
23234   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23235   return __ret;
23236 }
23237 #endif
23238 
23239 #ifdef __LITTLE_ENDIAN__
vrsqrtsq_f32(float32x4_t __p0,float32x4_t __p1)23240 __ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
23241   float32x4_t __ret;
23242   __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
23243   return __ret;
23244 }
23245 #else
vrsqrtsq_f32(float32x4_t __p0,float32x4_t __p1)23246 __ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
23247   float32x4_t __ret;
23248   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23249   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
23250   __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
23251   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23252   return __ret;
23253 }
23254 #endif
23255 
23256 #ifdef __LITTLE_ENDIAN__
vrsqrts_f32(float32x2_t __p0,float32x2_t __p1)23257 __ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
23258   float32x2_t __ret;
23259   __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
23260   return __ret;
23261 }
23262 #else
vrsqrts_f32(float32x2_t __p0,float32x2_t __p1)23263 __ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
23264   float32x2_t __ret;
23265   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23266   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
23267   __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
23268   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23269   return __ret;
23270 }
23271 #endif
23272 
23273 #ifdef __LITTLE_ENDIAN__
23274 #define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23275   uint8x16_t __ret; \
23276   uint8x16_t __s0 = __p0; \
23277   uint8x16_t __s1 = __p1; \
23278   __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
23279   __ret; \
23280 })
23281 #else
23282 #define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23283   uint8x16_t __ret; \
23284   uint8x16_t __s0 = __p0; \
23285   uint8x16_t __s1 = __p1; \
23286   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23287   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23288   __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
23289   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23290   __ret; \
23291 })
23292 #endif
23293 
23294 #ifdef __LITTLE_ENDIAN__
23295 #define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
23296   uint32x4_t __ret; \
23297   uint32x4_t __s0 = __p0; \
23298   uint32x4_t __s1 = __p1; \
23299   __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
23300   __ret; \
23301 })
23302 #else
23303 #define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
23304   uint32x4_t __ret; \
23305   uint32x4_t __s0 = __p0; \
23306   uint32x4_t __s1 = __p1; \
23307   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23308   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23309   __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
23310   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23311   __ret; \
23312 })
23313 #endif
23314 
23315 #ifdef __LITTLE_ENDIAN__
23316 #define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
23317   uint64x2_t __ret; \
23318   uint64x2_t __s0 = __p0; \
23319   uint64x2_t __s1 = __p1; \
23320   __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
23321   __ret; \
23322 })
23323 #else
23324 #define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
23325   uint64x2_t __ret; \
23326   uint64x2_t __s0 = __p0; \
23327   uint64x2_t __s1 = __p1; \
23328   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23329   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23330   __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
23331   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23332   __ret; \
23333 })
23334 #endif
23335 
23336 #ifdef __LITTLE_ENDIAN__
23337 #define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
23338   uint16x8_t __ret; \
23339   uint16x8_t __s0 = __p0; \
23340   uint16x8_t __s1 = __p1; \
23341   __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
23342   __ret; \
23343 })
23344 #else
23345 #define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
23346   uint16x8_t __ret; \
23347   uint16x8_t __s0 = __p0; \
23348   uint16x8_t __s1 = __p1; \
23349   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23350   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23351   __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
23352   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23353   __ret; \
23354 })
23355 #endif
23356 
23357 #ifdef __LITTLE_ENDIAN__
23358 #define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
23359   int8x16_t __ret; \
23360   int8x16_t __s0 = __p0; \
23361   int8x16_t __s1 = __p1; \
23362   __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
23363   __ret; \
23364 })
23365 #else
23366 #define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
23367   int8x16_t __ret; \
23368   int8x16_t __s0 = __p0; \
23369   int8x16_t __s1 = __p1; \
23370   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23371   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23372   __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
23373   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23374   __ret; \
23375 })
23376 #endif
23377 
23378 #ifdef __LITTLE_ENDIAN__
23379 #define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
23380   int32x4_t __ret; \
23381   int32x4_t __s0 = __p0; \
23382   int32x4_t __s1 = __p1; \
23383   __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
23384   __ret; \
23385 })
23386 #else
23387 #define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
23388   int32x4_t __ret; \
23389   int32x4_t __s0 = __p0; \
23390   int32x4_t __s1 = __p1; \
23391   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23392   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23393   __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
23394   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23395   __ret; \
23396 })
23397 #endif
23398 
23399 #ifdef __LITTLE_ENDIAN__
23400 #define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
23401   int64x2_t __ret; \
23402   int64x2_t __s0 = __p0; \
23403   int64x2_t __s1 = __p1; \
23404   __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
23405   __ret; \
23406 })
23407 #else
23408 #define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
23409   int64x2_t __ret; \
23410   int64x2_t __s0 = __p0; \
23411   int64x2_t __s1 = __p1; \
23412   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23413   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23414   __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
23415   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23416   __ret; \
23417 })
23418 #endif
23419 
23420 #ifdef __LITTLE_ENDIAN__
23421 #define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
23422   int16x8_t __ret; \
23423   int16x8_t __s0 = __p0; \
23424   int16x8_t __s1 = __p1; \
23425   __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
23426   __ret; \
23427 })
23428 #else
23429 #define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
23430   int16x8_t __ret; \
23431   int16x8_t __s0 = __p0; \
23432   int16x8_t __s1 = __p1; \
23433   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23434   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23435   __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
23436   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23437   __ret; \
23438 })
23439 #endif
23440 
23441 #ifdef __LITTLE_ENDIAN__
23442 #define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
23443   uint8x8_t __ret; \
23444   uint8x8_t __s0 = __p0; \
23445   uint8x8_t __s1 = __p1; \
23446   __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
23447   __ret; \
23448 })
23449 #else
23450 #define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
23451   uint8x8_t __ret; \
23452   uint8x8_t __s0 = __p0; \
23453   uint8x8_t __s1 = __p1; \
23454   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23455   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23456   __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
23457   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23458   __ret; \
23459 })
23460 #endif
23461 
23462 #ifdef __LITTLE_ENDIAN__
23463 #define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
23464   uint32x2_t __ret; \
23465   uint32x2_t __s0 = __p0; \
23466   uint32x2_t __s1 = __p1; \
23467   __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
23468   __ret; \
23469 })
23470 #else
23471 #define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
23472   uint32x2_t __ret; \
23473   uint32x2_t __s0 = __p0; \
23474   uint32x2_t __s1 = __p1; \
23475   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23476   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23477   __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
23478   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23479   __ret; \
23480 })
23481 #endif
23482 
23483 #define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
23484   uint64x1_t __ret; \
23485   uint64x1_t __s0 = __p0; \
23486   uint64x1_t __s1 = __p1; \
23487   __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
23488   __ret; \
23489 })
23490 #ifdef __LITTLE_ENDIAN__
23491 #define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
23492   uint16x4_t __ret; \
23493   uint16x4_t __s0 = __p0; \
23494   uint16x4_t __s1 = __p1; \
23495   __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
23496   __ret; \
23497 })
23498 #else
23499 #define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
23500   uint16x4_t __ret; \
23501   uint16x4_t __s0 = __p0; \
23502   uint16x4_t __s1 = __p1; \
23503   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23504   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23505   __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
23506   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23507   __ret; \
23508 })
23509 #endif
23510 
23511 #ifdef __LITTLE_ENDIAN__
23512 #define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
23513   int8x8_t __ret; \
23514   int8x8_t __s0 = __p0; \
23515   int8x8_t __s1 = __p1; \
23516   __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
23517   __ret; \
23518 })
23519 #else
23520 #define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
23521   int8x8_t __ret; \
23522   int8x8_t __s0 = __p0; \
23523   int8x8_t __s1 = __p1; \
23524   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23525   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23526   __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
23527   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23528   __ret; \
23529 })
23530 #endif
23531 
23532 #ifdef __LITTLE_ENDIAN__
23533 #define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
23534   int32x2_t __ret; \
23535   int32x2_t __s0 = __p0; \
23536   int32x2_t __s1 = __p1; \
23537   __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
23538   __ret; \
23539 })
23540 #else
23541 #define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
23542   int32x2_t __ret; \
23543   int32x2_t __s0 = __p0; \
23544   int32x2_t __s1 = __p1; \
23545   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23546   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23547   __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
23548   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23549   __ret; \
23550 })
23551 #endif
23552 
23553 #define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
23554   int64x1_t __ret; \
23555   int64x1_t __s0 = __p0; \
23556   int64x1_t __s1 = __p1; \
23557   __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
23558   __ret; \
23559 })
23560 #ifdef __LITTLE_ENDIAN__
23561 #define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
23562   int16x4_t __ret; \
23563   int16x4_t __s0 = __p0; \
23564   int16x4_t __s1 = __p1; \
23565   __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
23566   __ret; \
23567 })
23568 #else
23569 #define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
23570   int16x4_t __ret; \
23571   int16x4_t __s0 = __p0; \
23572   int16x4_t __s1 = __p1; \
23573   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23574   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23575   __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
23576   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23577   __ret; \
23578 })
23579 #endif
23580 
23581 #ifdef __LITTLE_ENDIAN__
vrsubhn_u32(uint32x4_t __p0,uint32x4_t __p1)23582 __ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
23583   uint16x4_t __ret;
23584   __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
23585   return __ret;
23586 }
23587 #else
vrsubhn_u32(uint32x4_t __p0,uint32x4_t __p1)23588 __ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
23589   uint16x4_t __ret;
23590   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23591   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
23592   __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
23593   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23594   return __ret;
23595 }
__noswap_vrsubhn_u32(uint32x4_t __p0,uint32x4_t __p1)23596 __ai uint16x4_t __noswap_vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
23597   uint16x4_t __ret;
23598   __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
23599   return __ret;
23600 }
23601 #endif
23602 
23603 #ifdef __LITTLE_ENDIAN__
vrsubhn_u64(uint64x2_t __p0,uint64x2_t __p1)23604 __ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
23605   uint32x2_t __ret;
23606   __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
23607   return __ret;
23608 }
23609 #else
vrsubhn_u64(uint64x2_t __p0,uint64x2_t __p1)23610 __ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
23611   uint32x2_t __ret;
23612   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23613   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
23614   __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
23615   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23616   return __ret;
23617 }
__noswap_vrsubhn_u64(uint64x2_t __p0,uint64x2_t __p1)23618 __ai uint32x2_t __noswap_vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
23619   uint32x2_t __ret;
23620   __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
23621   return __ret;
23622 }
23623 #endif
23624 
23625 #ifdef __LITTLE_ENDIAN__
vrsubhn_u16(uint16x8_t __p0,uint16x8_t __p1)23626 __ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
23627   uint8x8_t __ret;
23628   __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
23629   return __ret;
23630 }
23631 #else
vrsubhn_u16(uint16x8_t __p0,uint16x8_t __p1)23632 __ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
23633   uint8x8_t __ret;
23634   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
23635   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
23636   __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
23637   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
23638   return __ret;
23639 }
__noswap_vrsubhn_u16(uint16x8_t __p0,uint16x8_t __p1)23640 __ai uint8x8_t __noswap_vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
23641   uint8x8_t __ret;
23642   __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
23643   return __ret;
23644 }
23645 #endif
23646 
23647 #ifdef __LITTLE_ENDIAN__
vrsubhn_s32(int32x4_t __p0,int32x4_t __p1)23648 __ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
23649   int16x4_t __ret;
23650   __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
23651   return __ret;
23652 }
23653 #else
vrsubhn_s32(int32x4_t __p0,int32x4_t __p1)23654 __ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
23655   int16x4_t __ret;
23656   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23657   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
23658   __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
23659   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23660   return __ret;
23661 }
__noswap_vrsubhn_s32(int32x4_t __p0,int32x4_t __p1)23662 __ai int16x4_t __noswap_vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
23663   int16x4_t __ret;
23664   __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
23665   return __ret;
23666 }
23667 #endif
23668 
23669 #ifdef __LITTLE_ENDIAN__
vrsubhn_s64(int64x2_t __p0,int64x2_t __p1)23670 __ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
23671   int32x2_t __ret;
23672   __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
23673   return __ret;
23674 }
23675 #else
vrsubhn_s64(int64x2_t __p0,int64x2_t __p1)23676 __ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
23677   int32x2_t __ret;
23678   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23679   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
23680   __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
23681   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23682   return __ret;
23683 }
__noswap_vrsubhn_s64(int64x2_t __p0,int64x2_t __p1)23684 __ai int32x2_t __noswap_vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
23685   int32x2_t __ret;
23686   __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
23687   return __ret;
23688 }
23689 #endif
23690 
23691 #ifdef __LITTLE_ENDIAN__
vrsubhn_s16(int16x8_t __p0,int16x8_t __p1)23692 __ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
23693   int8x8_t __ret;
23694   __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
23695   return __ret;
23696 }
23697 #else
vrsubhn_s16(int16x8_t __p0,int16x8_t __p1)23698 __ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
23699   int8x8_t __ret;
23700   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
23701   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
23702   __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
23703   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
23704   return __ret;
23705 }
__noswap_vrsubhn_s16(int16x8_t __p0,int16x8_t __p1)23706 __ai int8x8_t __noswap_vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
23707   int8x8_t __ret;
23708   __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
23709   return __ret;
23710 }
23711 #endif
23712 
23713 #ifdef __LITTLE_ENDIAN__
23714 #define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23715   poly8x8_t __ret; \
23716   poly8_t __s0 = __p0; \
23717   poly8x8_t __s1 = __p1; \
23718   __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (poly8x8_t)__s1, __p2); \
23719   __ret; \
23720 })
23721 #else
23722 #define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23723   poly8x8_t __ret; \
23724   poly8_t __s0 = __p0; \
23725   poly8x8_t __s1 = __p1; \
23726   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23727   __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (poly8x8_t)__rev1, __p2); \
23728   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23729   __ret; \
23730 })
23731 #define __noswap_vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23732   poly8x8_t __ret; \
23733   poly8_t __s0 = __p0; \
23734   poly8x8_t __s1 = __p1; \
23735   __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (poly8x8_t)__s1, __p2); \
23736   __ret; \
23737 })
23738 #endif
23739 
23740 #ifdef __LITTLE_ENDIAN__
23741 #define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23742   poly16x4_t __ret; \
23743   poly16_t __s0 = __p0; \
23744   poly16x4_t __s1 = __p1; \
23745   __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (poly16x4_t)__s1, __p2); \
23746   __ret; \
23747 })
23748 #else
23749 #define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23750   poly16x4_t __ret; \
23751   poly16_t __s0 = __p0; \
23752   poly16x4_t __s1 = __p1; \
23753   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23754   __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (poly16x4_t)__rev1, __p2); \
23755   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23756   __ret; \
23757 })
23758 #define __noswap_vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23759   poly16x4_t __ret; \
23760   poly16_t __s0 = __p0; \
23761   poly16x4_t __s1 = __p1; \
23762   __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (poly16x4_t)__s1, __p2); \
23763   __ret; \
23764 })
23765 #endif
23766 
23767 #ifdef __LITTLE_ENDIAN__
23768 #define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23769   poly8x16_t __ret; \
23770   poly8_t __s0 = __p0; \
23771   poly8x16_t __s1 = __p1; \
23772   __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (poly8x16_t)__s1, __p2); \
23773   __ret; \
23774 })
23775 #else
23776 #define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23777   poly8x16_t __ret; \
23778   poly8_t __s0 = __p0; \
23779   poly8x16_t __s1 = __p1; \
23780   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23781   __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (poly8x16_t)__rev1, __p2); \
23782   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23783   __ret; \
23784 })
23785 #define __noswap_vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23786   poly8x16_t __ret; \
23787   poly8_t __s0 = __p0; \
23788   poly8x16_t __s1 = __p1; \
23789   __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (poly8x16_t)__s1, __p2); \
23790   __ret; \
23791 })
23792 #endif
23793 
23794 #ifdef __LITTLE_ENDIAN__
23795 #define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23796   poly16x8_t __ret; \
23797   poly16_t __s0 = __p0; \
23798   poly16x8_t __s1 = __p1; \
23799   __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (poly16x8_t)__s1, __p2); \
23800   __ret; \
23801 })
23802 #else
23803 #define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23804   poly16x8_t __ret; \
23805   poly16_t __s0 = __p0; \
23806   poly16x8_t __s1 = __p1; \
23807   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23808   __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (poly16x8_t)__rev1, __p2); \
23809   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23810   __ret; \
23811 })
23812 #define __noswap_vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23813   poly16x8_t __ret; \
23814   poly16_t __s0 = __p0; \
23815   poly16x8_t __s1 = __p1; \
23816   __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (poly16x8_t)__s1, __p2); \
23817   __ret; \
23818 })
23819 #endif
23820 
23821 #ifdef __LITTLE_ENDIAN__
23822 #define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
23823   uint8x16_t __ret; \
23824   uint8_t __s0 = __p0; \
23825   uint8x16_t __s1 = __p1; \
23826   __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
23827   __ret; \
23828 })
23829 #else
23830 #define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
23831   uint8x16_t __ret; \
23832   uint8_t __s0 = __p0; \
23833   uint8x16_t __s1 = __p1; \
23834   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23835   __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
23836   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23837   __ret; \
23838 })
23839 #define __noswap_vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
23840   uint8x16_t __ret; \
23841   uint8_t __s0 = __p0; \
23842   uint8x16_t __s1 = __p1; \
23843   __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
23844   __ret; \
23845 })
23846 #endif
23847 
23848 #ifdef __LITTLE_ENDIAN__
23849 #define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
23850   uint32x4_t __ret; \
23851   uint32_t __s0 = __p0; \
23852   uint32x4_t __s1 = __p1; \
23853   __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__s1, __p2); \
23854   __ret; \
23855 })
23856 #else
23857 #define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
23858   uint32x4_t __ret; \
23859   uint32_t __s0 = __p0; \
23860   uint32x4_t __s1 = __p1; \
23861   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23862   __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__rev1, __p2); \
23863   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23864   __ret; \
23865 })
23866 #define __noswap_vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
23867   uint32x4_t __ret; \
23868   uint32_t __s0 = __p0; \
23869   uint32x4_t __s1 = __p1; \
23870   __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__s1, __p2); \
23871   __ret; \
23872 })
23873 #endif
23874 
23875 #ifdef __LITTLE_ENDIAN__
23876 #define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
23877   uint64x2_t __ret; \
23878   uint64_t __s0 = __p0; \
23879   uint64x2_t __s1 = __p1; \
23880   __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__s1, __p2); \
23881   __ret; \
23882 })
23883 #else
23884 #define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
23885   uint64x2_t __ret; \
23886   uint64_t __s0 = __p0; \
23887   uint64x2_t __s1 = __p1; \
23888   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23889   __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__rev1, __p2); \
23890   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23891   __ret; \
23892 })
23893 #define __noswap_vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
23894   uint64x2_t __ret; \
23895   uint64_t __s0 = __p0; \
23896   uint64x2_t __s1 = __p1; \
23897   __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__s1, __p2); \
23898   __ret; \
23899 })
23900 #endif
23901 
23902 #ifdef __LITTLE_ENDIAN__
23903 #define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
23904   uint16x8_t __ret; \
23905   uint16_t __s0 = __p0; \
23906   uint16x8_t __s1 = __p1; \
23907   __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__s1, __p2); \
23908   __ret; \
23909 })
23910 #else
23911 #define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
23912   uint16x8_t __ret; \
23913   uint16_t __s0 = __p0; \
23914   uint16x8_t __s1 = __p1; \
23915   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23916   __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__rev1, __p2); \
23917   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23918   __ret; \
23919 })
23920 #define __noswap_vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
23921   uint16x8_t __ret; \
23922   uint16_t __s0 = __p0; \
23923   uint16x8_t __s1 = __p1; \
23924   __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__s1, __p2); \
23925   __ret; \
23926 })
23927 #endif
23928 
23929 #ifdef __LITTLE_ENDIAN__
23930 #define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
23931   int8x16_t __ret; \
23932   int8_t __s0 = __p0; \
23933   int8x16_t __s1 = __p1; \
23934   __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
23935   __ret; \
23936 })
23937 #else
23938 #define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
23939   int8x16_t __ret; \
23940   int8_t __s0 = __p0; \
23941   int8x16_t __s1 = __p1; \
23942   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23943   __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
23944   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23945   __ret; \
23946 })
23947 #define __noswap_vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
23948   int8x16_t __ret; \
23949   int8_t __s0 = __p0; \
23950   int8x16_t __s1 = __p1; \
23951   __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
23952   __ret; \
23953 })
23954 #endif
23955 
23956 #ifdef __LITTLE_ENDIAN__
23957 #define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
23958   float32x4_t __ret; \
23959   float32_t __s0 = __p0; \
23960   float32x4_t __s1 = __p1; \
23961   __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (float32x4_t)__s1, __p2); \
23962   __ret; \
23963 })
23964 #else
23965 #define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
23966   float32x4_t __ret; \
23967   float32_t __s0 = __p0; \
23968   float32x4_t __s1 = __p1; \
23969   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23970   __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (float32x4_t)__rev1, __p2); \
23971   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23972   __ret; \
23973 })
23974 #define __noswap_vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
23975   float32x4_t __ret; \
23976   float32_t __s0 = __p0; \
23977   float32x4_t __s1 = __p1; \
23978   __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (float32x4_t)__s1, __p2); \
23979   __ret; \
23980 })
23981 #endif
23982 
23983 #ifdef __LITTLE_ENDIAN__
23984 #define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
23985   int32x4_t __ret; \
23986   int32_t __s0 = __p0; \
23987   int32x4_t __s1 = __p1; \
23988   __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__s1, __p2); \
23989   __ret; \
23990 })
23991 #else
23992 #define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
23993   int32x4_t __ret; \
23994   int32_t __s0 = __p0; \
23995   int32x4_t __s1 = __p1; \
23996   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23997   __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__rev1, __p2); \
23998   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23999   __ret; \
24000 })
24001 #define __noswap_vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24002   int32x4_t __ret; \
24003   int32_t __s0 = __p0; \
24004   int32x4_t __s1 = __p1; \
24005   __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__s1, __p2); \
24006   __ret; \
24007 })
24008 #endif
24009 
24010 #ifdef __LITTLE_ENDIAN__
24011 #define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
24012   int64x2_t __ret; \
24013   int64_t __s0 = __p0; \
24014   int64x2_t __s1 = __p1; \
24015   __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__s1, __p2); \
24016   __ret; \
24017 })
24018 #else
24019 #define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
24020   int64x2_t __ret; \
24021   int64_t __s0 = __p0; \
24022   int64x2_t __s1 = __p1; \
24023   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24024   __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__rev1, __p2); \
24025   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24026   __ret; \
24027 })
24028 #define __noswap_vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
24029   int64x2_t __ret; \
24030   int64_t __s0 = __p0; \
24031   int64x2_t __s1 = __p1; \
24032   __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__s1, __p2); \
24033   __ret; \
24034 })
24035 #endif
24036 
24037 #ifdef __LITTLE_ENDIAN__
24038 #define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24039   int16x8_t __ret; \
24040   int16_t __s0 = __p0; \
24041   int16x8_t __s1 = __p1; \
24042   __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__s1, __p2); \
24043   __ret; \
24044 })
24045 #else
24046 #define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24047   int16x8_t __ret; \
24048   int16_t __s0 = __p0; \
24049   int16x8_t __s1 = __p1; \
24050   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24051   __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__rev1, __p2); \
24052   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24053   __ret; \
24054 })
24055 #define __noswap_vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24056   int16x8_t __ret; \
24057   int16_t __s0 = __p0; \
24058   int16x8_t __s1 = __p1; \
24059   __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__s1, __p2); \
24060   __ret; \
24061 })
24062 #endif
24063 
24064 #ifdef __LITTLE_ENDIAN__
24065 #define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
24066   uint8x8_t __ret; \
24067   uint8_t __s0 = __p0; \
24068   uint8x8_t __s1 = __p1; \
24069   __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
24070   __ret; \
24071 })
24072 #else
24073 #define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
24074   uint8x8_t __ret; \
24075   uint8_t __s0 = __p0; \
24076   uint8x8_t __s1 = __p1; \
24077   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24078   __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
24079   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24080   __ret; \
24081 })
24082 #define __noswap_vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
24083   uint8x8_t __ret; \
24084   uint8_t __s0 = __p0; \
24085   uint8x8_t __s1 = __p1; \
24086   __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
24087   __ret; \
24088 })
24089 #endif
24090 
24091 #ifdef __LITTLE_ENDIAN__
24092 #define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
24093   uint32x2_t __ret; \
24094   uint32_t __s0 = __p0; \
24095   uint32x2_t __s1 = __p1; \
24096   __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__s1, __p2); \
24097   __ret; \
24098 })
24099 #else
24100 #define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
24101   uint32x2_t __ret; \
24102   uint32_t __s0 = __p0; \
24103   uint32x2_t __s1 = __p1; \
24104   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24105   __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__rev1, __p2); \
24106   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24107   __ret; \
24108 })
24109 #define __noswap_vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
24110   uint32x2_t __ret; \
24111   uint32_t __s0 = __p0; \
24112   uint32x2_t __s1 = __p1; \
24113   __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__s1, __p2); \
24114   __ret; \
24115 })
24116 #endif
24117 
24118 #define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
24119   uint64x1_t __ret; \
24120   uint64_t __s0 = __p0; \
24121   uint64x1_t __s1 = __p1; \
24122   __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int64x1_t)__s1, __p2); \
24123   __ret; \
24124 })
24125 #ifdef __LITTLE_ENDIAN__
24126 #define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
24127   uint16x4_t __ret; \
24128   uint16_t __s0 = __p0; \
24129   uint16x4_t __s1 = __p1; \
24130   __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__s1, __p2); \
24131   __ret; \
24132 })
24133 #else
24134 #define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
24135   uint16x4_t __ret; \
24136   uint16_t __s0 = __p0; \
24137   uint16x4_t __s1 = __p1; \
24138   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24139   __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__rev1, __p2); \
24140   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24141   __ret; \
24142 })
24143 #define __noswap_vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
24144   uint16x4_t __ret; \
24145   uint16_t __s0 = __p0; \
24146   uint16x4_t __s1 = __p1; \
24147   __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__s1, __p2); \
24148   __ret; \
24149 })
24150 #endif
24151 
24152 #ifdef __LITTLE_ENDIAN__
24153 #define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
24154   int8x8_t __ret; \
24155   int8_t __s0 = __p0; \
24156   int8x8_t __s1 = __p1; \
24157   __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
24158   __ret; \
24159 })
24160 #else
24161 #define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
24162   int8x8_t __ret; \
24163   int8_t __s0 = __p0; \
24164   int8x8_t __s1 = __p1; \
24165   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24166   __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
24167   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24168   __ret; \
24169 })
24170 #define __noswap_vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
24171   int8x8_t __ret; \
24172   int8_t __s0 = __p0; \
24173   int8x8_t __s1 = __p1; \
24174   __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
24175   __ret; \
24176 })
24177 #endif
24178 
24179 #ifdef __LITTLE_ENDIAN__
24180 #define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
24181   float32x2_t __ret; \
24182   float32_t __s0 = __p0; \
24183   float32x2_t __s1 = __p1; \
24184   __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (float32x2_t)__s1, __p2); \
24185   __ret; \
24186 })
24187 #else
24188 #define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
24189   float32x2_t __ret; \
24190   float32_t __s0 = __p0; \
24191   float32x2_t __s1 = __p1; \
24192   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24193   __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (float32x2_t)__rev1, __p2); \
24194   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24195   __ret; \
24196 })
24197 #define __noswap_vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
24198   float32x2_t __ret; \
24199   float32_t __s0 = __p0; \
24200   float32x2_t __s1 = __p1; \
24201   __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (float32x2_t)__s1, __p2); \
24202   __ret; \
24203 })
24204 #endif
24205 
24206 #ifdef __LITTLE_ENDIAN__
24207 #define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24208   int32x2_t __ret; \
24209   int32_t __s0 = __p0; \
24210   int32x2_t __s1 = __p1; \
24211   __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__s1, __p2); \
24212   __ret; \
24213 })
24214 #else
24215 #define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24216   int32x2_t __ret; \
24217   int32_t __s0 = __p0; \
24218   int32x2_t __s1 = __p1; \
24219   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24220   __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__rev1, __p2); \
24221   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24222   __ret; \
24223 })
24224 #define __noswap_vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24225   int32x2_t __ret; \
24226   int32_t __s0 = __p0; \
24227   int32x2_t __s1 = __p1; \
24228   __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__s1, __p2); \
24229   __ret; \
24230 })
24231 #endif
24232 
24233 #define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
24234   int64x1_t __ret; \
24235   int64_t __s0 = __p0; \
24236   int64x1_t __s1 = __p1; \
24237   __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int64x1_t)__s1, __p2); \
24238   __ret; \
24239 })
24240 #ifdef __LITTLE_ENDIAN__
24241 #define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24242   int16x4_t __ret; \
24243   int16_t __s0 = __p0; \
24244   int16x4_t __s1 = __p1; \
24245   __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__s1, __p2); \
24246   __ret; \
24247 })
24248 #else
24249 #define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24250   int16x4_t __ret; \
24251   int16_t __s0 = __p0; \
24252   int16x4_t __s1 = __p1; \
24253   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24254   __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__rev1, __p2); \
24255   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24256   __ret; \
24257 })
24258 #define __noswap_vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24259   int16x4_t __ret; \
24260   int16_t __s0 = __p0; \
24261   int16x4_t __s1 = __p1; \
24262   __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__s1, __p2); \
24263   __ret; \
24264 })
24265 #endif
24266 
24267 #ifdef __LITTLE_ENDIAN__
vshlq_u8(uint8x16_t __p0,int8x16_t __p1)24268 __ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
24269   uint8x16_t __ret;
24270   __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
24271   return __ret;
24272 }
24273 #else
vshlq_u8(uint8x16_t __p0,int8x16_t __p1)24274 __ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
24275   uint8x16_t __ret;
24276   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24277   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24278   __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
24279   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24280   return __ret;
24281 }
24282 #endif
24283 
24284 #ifdef __LITTLE_ENDIAN__
vshlq_u32(uint32x4_t __p0,int32x4_t __p1)24285 __ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
24286   uint32x4_t __ret;
24287   __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
24288   return __ret;
24289 }
24290 #else
vshlq_u32(uint32x4_t __p0,int32x4_t __p1)24291 __ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
24292   uint32x4_t __ret;
24293   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
24294   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
24295   __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
24296   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
24297   return __ret;
24298 }
24299 #endif
24300 
24301 #ifdef __LITTLE_ENDIAN__
vshlq_u64(uint64x2_t __p0,int64x2_t __p1)24302 __ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
24303   uint64x2_t __ret;
24304   __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
24305   return __ret;
24306 }
24307 #else
vshlq_u64(uint64x2_t __p0,int64x2_t __p1)24308 __ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
24309   uint64x2_t __ret;
24310   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
24311   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
24312   __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
24313   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
24314   return __ret;
24315 }
24316 #endif
24317 
24318 #ifdef __LITTLE_ENDIAN__
vshlq_u16(uint16x8_t __p0,int16x8_t __p1)24319 __ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
24320   uint16x8_t __ret;
24321   __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
24322   return __ret;
24323 }
24324 #else
vshlq_u16(uint16x8_t __p0,int16x8_t __p1)24325 __ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
24326   uint16x8_t __ret;
24327   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
24328   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
24329   __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
24330   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
24331   return __ret;
24332 }
24333 #endif
24334 
24335 #ifdef __LITTLE_ENDIAN__
vshlq_s8(int8x16_t __p0,int8x16_t __p1)24336 __ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
24337   int8x16_t __ret;
24338   __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
24339   return __ret;
24340 }
24341 #else
vshlq_s8(int8x16_t __p0,int8x16_t __p1)24342 __ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
24343   int8x16_t __ret;
24344   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24345   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24346   __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
24347   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24348   return __ret;
24349 }
24350 #endif
24351 
24352 #ifdef __LITTLE_ENDIAN__
vshlq_s32(int32x4_t __p0,int32x4_t __p1)24353 __ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
24354   int32x4_t __ret;
24355   __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
24356   return __ret;
24357 }
24358 #else
vshlq_s32(int32x4_t __p0,int32x4_t __p1)24359 __ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
24360   int32x4_t __ret;
24361   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
24362   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
24363   __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
24364   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
24365   return __ret;
24366 }
24367 #endif
24368 
24369 #ifdef __LITTLE_ENDIAN__
vshlq_s64(int64x2_t __p0,int64x2_t __p1)24370 __ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
24371   int64x2_t __ret;
24372   __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
24373   return __ret;
24374 }
24375 #else
vshlq_s64(int64x2_t __p0,int64x2_t __p1)24376 __ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
24377   int64x2_t __ret;
24378   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
24379   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
24380   __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
24381   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
24382   return __ret;
24383 }
24384 #endif
24385 
24386 #ifdef __LITTLE_ENDIAN__
vshlq_s16(int16x8_t __p0,int16x8_t __p1)24387 __ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
24388   int16x8_t __ret;
24389   __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
24390   return __ret;
24391 }
24392 #else
vshlq_s16(int16x8_t __p0,int16x8_t __p1)24393 __ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
24394   int16x8_t __ret;
24395   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
24396   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
24397   __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
24398   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
24399   return __ret;
24400 }
24401 #endif
24402 
24403 #ifdef __LITTLE_ENDIAN__
vshl_u8(uint8x8_t __p0,int8x8_t __p1)24404 __ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
24405   uint8x8_t __ret;
24406   __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
24407   return __ret;
24408 }
24409 #else
vshl_u8(uint8x8_t __p0,int8x8_t __p1)24410 __ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
24411   uint8x8_t __ret;
24412   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
24413   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
24414   __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
24415   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
24416   return __ret;
24417 }
24418 #endif
24419 
24420 #ifdef __LITTLE_ENDIAN__
vshl_u32(uint32x2_t __p0,int32x2_t __p1)24421 __ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
24422   uint32x2_t __ret;
24423   __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
24424   return __ret;
24425 }
24426 #else
vshl_u32(uint32x2_t __p0,int32x2_t __p1)24427 __ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
24428   uint32x2_t __ret;
24429   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
24430   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
24431   __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
24432   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
24433   return __ret;
24434 }
24435 #endif
24436 
vshl_u64(uint64x1_t __p0,int64x1_t __p1)24437 __ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
24438   uint64x1_t __ret;
24439   __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
24440   return __ret;
24441 }
24442 #ifdef __LITTLE_ENDIAN__
vshl_u16(uint16x4_t __p0,int16x4_t __p1)24443 __ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
24444   uint16x4_t __ret;
24445   __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
24446   return __ret;
24447 }
24448 #else
vshl_u16(uint16x4_t __p0,int16x4_t __p1)24449 __ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
24450   uint16x4_t __ret;
24451   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
24452   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
24453   __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
24454   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
24455   return __ret;
24456 }
24457 #endif
24458 
24459 #ifdef __LITTLE_ENDIAN__
vshl_s8(int8x8_t __p0,int8x8_t __p1)24460 __ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
24461   int8x8_t __ret;
24462   __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
24463   return __ret;
24464 }
24465 #else
vshl_s8(int8x8_t __p0,int8x8_t __p1)24466 __ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
24467   int8x8_t __ret;
24468   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
24469   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
24470   __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
24471   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
24472   return __ret;
24473 }
24474 #endif
24475 
24476 #ifdef __LITTLE_ENDIAN__
vshl_s32(int32x2_t __p0,int32x2_t __p1)24477 __ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
24478   int32x2_t __ret;
24479   __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
24480   return __ret;
24481 }
24482 #else
vshl_s32(int32x2_t __p0,int32x2_t __p1)24483 __ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
24484   int32x2_t __ret;
24485   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
24486   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
24487   __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
24488   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
24489   return __ret;
24490 }
24491 #endif
24492 
vshl_s64(int64x1_t __p0,int64x1_t __p1)24493 __ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
24494   int64x1_t __ret;
24495   __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
24496   return __ret;
24497 }
24498 #ifdef __LITTLE_ENDIAN__
vshl_s16(int16x4_t __p0,int16x4_t __p1)24499 __ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
24500   int16x4_t __ret;
24501   __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
24502   return __ret;
24503 }
24504 #else
vshl_s16(int16x4_t __p0,int16x4_t __p1)24505 __ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
24506   int16x4_t __ret;
24507   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
24508   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
24509   __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
24510   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
24511   return __ret;
24512 }
24513 #endif
24514 
24515 #ifdef __LITTLE_ENDIAN__
24516 #define vshlq_n_u8(__p0, __p1) __extension__ ({ \
24517   uint8x16_t __ret; \
24518   uint8x16_t __s0 = __p0; \
24519   __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 48); \
24520   __ret; \
24521 })
24522 #else
24523 #define vshlq_n_u8(__p0, __p1) __extension__ ({ \
24524   uint8x16_t __ret; \
24525   uint8x16_t __s0 = __p0; \
24526   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24527   __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 48); \
24528   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24529   __ret; \
24530 })
24531 #endif
24532 
24533 #ifdef __LITTLE_ENDIAN__
24534 #define vshlq_n_u32(__p0, __p1) __extension__ ({ \
24535   uint32x4_t __ret; \
24536   uint32x4_t __s0 = __p0; \
24537   __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 50); \
24538   __ret; \
24539 })
24540 #else
24541 #define vshlq_n_u32(__p0, __p1) __extension__ ({ \
24542   uint32x4_t __ret; \
24543   uint32x4_t __s0 = __p0; \
24544   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24545   __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 50); \
24546   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24547   __ret; \
24548 })
24549 #endif
24550 
24551 #ifdef __LITTLE_ENDIAN__
24552 #define vshlq_n_u64(__p0, __p1) __extension__ ({ \
24553   uint64x2_t __ret; \
24554   uint64x2_t __s0 = __p0; \
24555   __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 51); \
24556   __ret; \
24557 })
24558 #else
24559 #define vshlq_n_u64(__p0, __p1) __extension__ ({ \
24560   uint64x2_t __ret; \
24561   uint64x2_t __s0 = __p0; \
24562   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24563   __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 51); \
24564   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24565   __ret; \
24566 })
24567 #endif
24568 
24569 #ifdef __LITTLE_ENDIAN__
24570 #define vshlq_n_u16(__p0, __p1) __extension__ ({ \
24571   uint16x8_t __ret; \
24572   uint16x8_t __s0 = __p0; \
24573   __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 49); \
24574   __ret; \
24575 })
24576 #else
24577 #define vshlq_n_u16(__p0, __p1) __extension__ ({ \
24578   uint16x8_t __ret; \
24579   uint16x8_t __s0 = __p0; \
24580   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24581   __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 49); \
24582   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24583   __ret; \
24584 })
24585 #endif
24586 
24587 #ifdef __LITTLE_ENDIAN__
24588 #define vshlq_n_s8(__p0, __p1) __extension__ ({ \
24589   int8x16_t __ret; \
24590   int8x16_t __s0 = __p0; \
24591   __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 32); \
24592   __ret; \
24593 })
24594 #else
24595 #define vshlq_n_s8(__p0, __p1) __extension__ ({ \
24596   int8x16_t __ret; \
24597   int8x16_t __s0 = __p0; \
24598   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24599   __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 32); \
24600   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24601   __ret; \
24602 })
24603 #endif
24604 
24605 #ifdef __LITTLE_ENDIAN__
24606 #define vshlq_n_s32(__p0, __p1) __extension__ ({ \
24607   int32x4_t __ret; \
24608   int32x4_t __s0 = __p0; \
24609   __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 34); \
24610   __ret; \
24611 })
24612 #else
24613 #define vshlq_n_s32(__p0, __p1) __extension__ ({ \
24614   int32x4_t __ret; \
24615   int32x4_t __s0 = __p0; \
24616   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24617   __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 34); \
24618   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24619   __ret; \
24620 })
24621 #endif
24622 
24623 #ifdef __LITTLE_ENDIAN__
24624 #define vshlq_n_s64(__p0, __p1) __extension__ ({ \
24625   int64x2_t __ret; \
24626   int64x2_t __s0 = __p0; \
24627   __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 35); \
24628   __ret; \
24629 })
24630 #else
24631 #define vshlq_n_s64(__p0, __p1) __extension__ ({ \
24632   int64x2_t __ret; \
24633   int64x2_t __s0 = __p0; \
24634   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24635   __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 35); \
24636   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24637   __ret; \
24638 })
24639 #endif
24640 
24641 #ifdef __LITTLE_ENDIAN__
24642 #define vshlq_n_s16(__p0, __p1) __extension__ ({ \
24643   int16x8_t __ret; \
24644   int16x8_t __s0 = __p0; \
24645   __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 33); \
24646   __ret; \
24647 })
24648 #else
24649 #define vshlq_n_s16(__p0, __p1) __extension__ ({ \
24650   int16x8_t __ret; \
24651   int16x8_t __s0 = __p0; \
24652   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24653   __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 33); \
24654   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24655   __ret; \
24656 })
24657 #endif
24658 
24659 #ifdef __LITTLE_ENDIAN__
24660 #define vshl_n_u8(__p0, __p1) __extension__ ({ \
24661   uint8x8_t __ret; \
24662   uint8x8_t __s0 = __p0; \
24663   __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 16); \
24664   __ret; \
24665 })
24666 #else
24667 #define vshl_n_u8(__p0, __p1) __extension__ ({ \
24668   uint8x8_t __ret; \
24669   uint8x8_t __s0 = __p0; \
24670   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24671   __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 16); \
24672   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24673   __ret; \
24674 })
24675 #endif
24676 
24677 #ifdef __LITTLE_ENDIAN__
24678 #define vshl_n_u32(__p0, __p1) __extension__ ({ \
24679   uint32x2_t __ret; \
24680   uint32x2_t __s0 = __p0; \
24681   __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 18); \
24682   __ret; \
24683 })
24684 #else
24685 #define vshl_n_u32(__p0, __p1) __extension__ ({ \
24686   uint32x2_t __ret; \
24687   uint32x2_t __s0 = __p0; \
24688   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24689   __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 18); \
24690   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24691   __ret; \
24692 })
24693 #endif
24694 
24695 #define vshl_n_u64(__p0, __p1) __extension__ ({ \
24696   uint64x1_t __ret; \
24697   uint64x1_t __s0 = __p0; \
24698   __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
24699   __ret; \
24700 })
24701 #ifdef __LITTLE_ENDIAN__
24702 #define vshl_n_u16(__p0, __p1) __extension__ ({ \
24703   uint16x4_t __ret; \
24704   uint16x4_t __s0 = __p0; \
24705   __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 17); \
24706   __ret; \
24707 })
24708 #else
24709 #define vshl_n_u16(__p0, __p1) __extension__ ({ \
24710   uint16x4_t __ret; \
24711   uint16x4_t __s0 = __p0; \
24712   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24713   __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 17); \
24714   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24715   __ret; \
24716 })
24717 #endif
24718 
24719 #ifdef __LITTLE_ENDIAN__
24720 #define vshl_n_s8(__p0, __p1) __extension__ ({ \
24721   int8x8_t __ret; \
24722   int8x8_t __s0 = __p0; \
24723   __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 0); \
24724   __ret; \
24725 })
24726 #else
24727 #define vshl_n_s8(__p0, __p1) __extension__ ({ \
24728   int8x8_t __ret; \
24729   int8x8_t __s0 = __p0; \
24730   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24731   __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 0); \
24732   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24733   __ret; \
24734 })
24735 #endif
24736 
24737 #ifdef __LITTLE_ENDIAN__
24738 #define vshl_n_s32(__p0, __p1) __extension__ ({ \
24739   int32x2_t __ret; \
24740   int32x2_t __s0 = __p0; \
24741   __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 2); \
24742   __ret; \
24743 })
24744 #else
24745 #define vshl_n_s32(__p0, __p1) __extension__ ({ \
24746   int32x2_t __ret; \
24747   int32x2_t __s0 = __p0; \
24748   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24749   __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 2); \
24750   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24751   __ret; \
24752 })
24753 #endif
24754 
24755 #define vshl_n_s64(__p0, __p1) __extension__ ({ \
24756   int64x1_t __ret; \
24757   int64x1_t __s0 = __p0; \
24758   __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
24759   __ret; \
24760 })
24761 #ifdef __LITTLE_ENDIAN__
24762 #define vshl_n_s16(__p0, __p1) __extension__ ({ \
24763   int16x4_t __ret; \
24764   int16x4_t __s0 = __p0; \
24765   __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 1); \
24766   __ret; \
24767 })
24768 #else
24769 #define vshl_n_s16(__p0, __p1) __extension__ ({ \
24770   int16x4_t __ret; \
24771   int16x4_t __s0 = __p0; \
24772   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24773   __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 1); \
24774   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24775   __ret; \
24776 })
24777 #endif
24778 
24779 #ifdef __LITTLE_ENDIAN__
24780 #define vshll_n_u8(__p0, __p1) __extension__ ({ \
24781   uint16x8_t __ret; \
24782   uint8x8_t __s0 = __p0; \
24783   __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
24784   __ret; \
24785 })
24786 #else
24787 #define vshll_n_u8(__p0, __p1) __extension__ ({ \
24788   uint16x8_t __ret; \
24789   uint8x8_t __s0 = __p0; \
24790   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24791   __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 49); \
24792   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24793   __ret; \
24794 })
24795 #define __noswap_vshll_n_u8(__p0, __p1) __extension__ ({ \
24796   uint16x8_t __ret; \
24797   uint8x8_t __s0 = __p0; \
24798   __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
24799   __ret; \
24800 })
24801 #endif
24802 
24803 #ifdef __LITTLE_ENDIAN__
24804 #define vshll_n_u32(__p0, __p1) __extension__ ({ \
24805   uint64x2_t __ret; \
24806   uint32x2_t __s0 = __p0; \
24807   __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
24808   __ret; \
24809 })
24810 #else
24811 #define vshll_n_u32(__p0, __p1) __extension__ ({ \
24812   uint64x2_t __ret; \
24813   uint32x2_t __s0 = __p0; \
24814   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24815   __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 51); \
24816   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24817   __ret; \
24818 })
24819 #define __noswap_vshll_n_u32(__p0, __p1) __extension__ ({ \
24820   uint64x2_t __ret; \
24821   uint32x2_t __s0 = __p0; \
24822   __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
24823   __ret; \
24824 })
24825 #endif
24826 
24827 #ifdef __LITTLE_ENDIAN__
24828 #define vshll_n_u16(__p0, __p1) __extension__ ({ \
24829   uint32x4_t __ret; \
24830   uint16x4_t __s0 = __p0; \
24831   __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
24832   __ret; \
24833 })
24834 #else
24835 #define vshll_n_u16(__p0, __p1) __extension__ ({ \
24836   uint32x4_t __ret; \
24837   uint16x4_t __s0 = __p0; \
24838   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24839   __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 50); \
24840   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24841   __ret; \
24842 })
24843 #define __noswap_vshll_n_u16(__p0, __p1) __extension__ ({ \
24844   uint32x4_t __ret; \
24845   uint16x4_t __s0 = __p0; \
24846   __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
24847   __ret; \
24848 })
24849 #endif
24850 
24851 #ifdef __LITTLE_ENDIAN__
24852 #define vshll_n_s8(__p0, __p1) __extension__ ({ \
24853   int16x8_t __ret; \
24854   int8x8_t __s0 = __p0; \
24855   __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
24856   __ret; \
24857 })
24858 #else
24859 #define vshll_n_s8(__p0, __p1) __extension__ ({ \
24860   int16x8_t __ret; \
24861   int8x8_t __s0 = __p0; \
24862   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24863   __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 33); \
24864   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24865   __ret; \
24866 })
24867 #define __noswap_vshll_n_s8(__p0, __p1) __extension__ ({ \
24868   int16x8_t __ret; \
24869   int8x8_t __s0 = __p0; \
24870   __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
24871   __ret; \
24872 })
24873 #endif
24874 
24875 #ifdef __LITTLE_ENDIAN__
24876 #define vshll_n_s32(__p0, __p1) __extension__ ({ \
24877   int64x2_t __ret; \
24878   int32x2_t __s0 = __p0; \
24879   __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
24880   __ret; \
24881 })
24882 #else
24883 #define vshll_n_s32(__p0, __p1) __extension__ ({ \
24884   int64x2_t __ret; \
24885   int32x2_t __s0 = __p0; \
24886   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24887   __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 35); \
24888   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24889   __ret; \
24890 })
24891 #define __noswap_vshll_n_s32(__p0, __p1) __extension__ ({ \
24892   int64x2_t __ret; \
24893   int32x2_t __s0 = __p0; \
24894   __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
24895   __ret; \
24896 })
24897 #endif
24898 
24899 #ifdef __LITTLE_ENDIAN__
24900 #define vshll_n_s16(__p0, __p1) __extension__ ({ \
24901   int32x4_t __ret; \
24902   int16x4_t __s0 = __p0; \
24903   __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
24904   __ret; \
24905 })
24906 #else
24907 #define vshll_n_s16(__p0, __p1) __extension__ ({ \
24908   int32x4_t __ret; \
24909   int16x4_t __s0 = __p0; \
24910   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24911   __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 34); \
24912   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24913   __ret; \
24914 })
24915 #define __noswap_vshll_n_s16(__p0, __p1) __extension__ ({ \
24916   int32x4_t __ret; \
24917   int16x4_t __s0 = __p0; \
24918   __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
24919   __ret; \
24920 })
24921 #endif
24922 
24923 #ifdef __LITTLE_ENDIAN__
24924 #define vshrq_n_u8(__p0, __p1) __extension__ ({ \
24925   uint8x16_t __ret; \
24926   uint8x16_t __s0 = __p0; \
24927   __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 48); \
24928   __ret; \
24929 })
24930 #else
24931 #define vshrq_n_u8(__p0, __p1) __extension__ ({ \
24932   uint8x16_t __ret; \
24933   uint8x16_t __s0 = __p0; \
24934   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24935   __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 48); \
24936   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24937   __ret; \
24938 })
24939 #endif
24940 
24941 #ifdef __LITTLE_ENDIAN__
24942 #define vshrq_n_u32(__p0, __p1) __extension__ ({ \
24943   uint32x4_t __ret; \
24944   uint32x4_t __s0 = __p0; \
24945   __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 50); \
24946   __ret; \
24947 })
24948 #else
24949 #define vshrq_n_u32(__p0, __p1) __extension__ ({ \
24950   uint32x4_t __ret; \
24951   uint32x4_t __s0 = __p0; \
24952   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24953   __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 50); \
24954   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24955   __ret; \
24956 })
24957 #endif
24958 
24959 #ifdef __LITTLE_ENDIAN__
24960 #define vshrq_n_u64(__p0, __p1) __extension__ ({ \
24961   uint64x2_t __ret; \
24962   uint64x2_t __s0 = __p0; \
24963   __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 51); \
24964   __ret; \
24965 })
24966 #else
24967 #define vshrq_n_u64(__p0, __p1) __extension__ ({ \
24968   uint64x2_t __ret; \
24969   uint64x2_t __s0 = __p0; \
24970   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24971   __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 51); \
24972   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24973   __ret; \
24974 })
24975 #endif
24976 
24977 #ifdef __LITTLE_ENDIAN__
24978 #define vshrq_n_u16(__p0, __p1) __extension__ ({ \
24979   uint16x8_t __ret; \
24980   uint16x8_t __s0 = __p0; \
24981   __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 49); \
24982   __ret; \
24983 })
24984 #else
24985 #define vshrq_n_u16(__p0, __p1) __extension__ ({ \
24986   uint16x8_t __ret; \
24987   uint16x8_t __s0 = __p0; \
24988   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24989   __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 49); \
24990   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24991   __ret; \
24992 })
24993 #endif
24994 
24995 #ifdef __LITTLE_ENDIAN__
24996 #define vshrq_n_s8(__p0, __p1) __extension__ ({ \
24997   int8x16_t __ret; \
24998   int8x16_t __s0 = __p0; \
24999   __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 32); \
25000   __ret; \
25001 })
25002 #else
25003 #define vshrq_n_s8(__p0, __p1) __extension__ ({ \
25004   int8x16_t __ret; \
25005   int8x16_t __s0 = __p0; \
25006   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25007   __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 32); \
25008   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25009   __ret; \
25010 })
25011 #endif
25012 
25013 #ifdef __LITTLE_ENDIAN__
25014 #define vshrq_n_s32(__p0, __p1) __extension__ ({ \
25015   int32x4_t __ret; \
25016   int32x4_t __s0 = __p0; \
25017   __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 34); \
25018   __ret; \
25019 })
25020 #else
25021 #define vshrq_n_s32(__p0, __p1) __extension__ ({ \
25022   int32x4_t __ret; \
25023   int32x4_t __s0 = __p0; \
25024   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25025   __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 34); \
25026   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25027   __ret; \
25028 })
25029 #endif
25030 
25031 #ifdef __LITTLE_ENDIAN__
25032 #define vshrq_n_s64(__p0, __p1) __extension__ ({ \
25033   int64x2_t __ret; \
25034   int64x2_t __s0 = __p0; \
25035   __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 35); \
25036   __ret; \
25037 })
25038 #else
25039 #define vshrq_n_s64(__p0, __p1) __extension__ ({ \
25040   int64x2_t __ret; \
25041   int64x2_t __s0 = __p0; \
25042   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25043   __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 35); \
25044   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25045   __ret; \
25046 })
25047 #endif
25048 
25049 #ifdef __LITTLE_ENDIAN__
25050 #define vshrq_n_s16(__p0, __p1) __extension__ ({ \
25051   int16x8_t __ret; \
25052   int16x8_t __s0 = __p0; \
25053   __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 33); \
25054   __ret; \
25055 })
25056 #else
25057 #define vshrq_n_s16(__p0, __p1) __extension__ ({ \
25058   int16x8_t __ret; \
25059   int16x8_t __s0 = __p0; \
25060   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25061   __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 33); \
25062   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25063   __ret; \
25064 })
25065 #endif
25066 
25067 #ifdef __LITTLE_ENDIAN__
25068 #define vshr_n_u8(__p0, __p1) __extension__ ({ \
25069   uint8x8_t __ret; \
25070   uint8x8_t __s0 = __p0; \
25071   __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 16); \
25072   __ret; \
25073 })
25074 #else
25075 #define vshr_n_u8(__p0, __p1) __extension__ ({ \
25076   uint8x8_t __ret; \
25077   uint8x8_t __s0 = __p0; \
25078   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25079   __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 16); \
25080   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25081   __ret; \
25082 })
25083 #endif
25084 
25085 #ifdef __LITTLE_ENDIAN__
25086 #define vshr_n_u32(__p0, __p1) __extension__ ({ \
25087   uint32x2_t __ret; \
25088   uint32x2_t __s0 = __p0; \
25089   __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 18); \
25090   __ret; \
25091 })
25092 #else
25093 #define vshr_n_u32(__p0, __p1) __extension__ ({ \
25094   uint32x2_t __ret; \
25095   uint32x2_t __s0 = __p0; \
25096   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25097   __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 18); \
25098   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25099   __ret; \
25100 })
25101 #endif
25102 
25103 #define vshr_n_u64(__p0, __p1) __extension__ ({ \
25104   uint64x1_t __ret; \
25105   uint64x1_t __s0 = __p0; \
25106   __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
25107   __ret; \
25108 })
25109 #ifdef __LITTLE_ENDIAN__
25110 #define vshr_n_u16(__p0, __p1) __extension__ ({ \
25111   uint16x4_t __ret; \
25112   uint16x4_t __s0 = __p0; \
25113   __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 17); \
25114   __ret; \
25115 })
25116 #else
25117 #define vshr_n_u16(__p0, __p1) __extension__ ({ \
25118   uint16x4_t __ret; \
25119   uint16x4_t __s0 = __p0; \
25120   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25121   __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 17); \
25122   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25123   __ret; \
25124 })
25125 #endif
25126 
25127 #ifdef __LITTLE_ENDIAN__
25128 #define vshr_n_s8(__p0, __p1) __extension__ ({ \
25129   int8x8_t __ret; \
25130   int8x8_t __s0 = __p0; \
25131   __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 0); \
25132   __ret; \
25133 })
25134 #else
25135 #define vshr_n_s8(__p0, __p1) __extension__ ({ \
25136   int8x8_t __ret; \
25137   int8x8_t __s0 = __p0; \
25138   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25139   __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 0); \
25140   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25141   __ret; \
25142 })
25143 #endif
25144 
25145 #ifdef __LITTLE_ENDIAN__
25146 #define vshr_n_s32(__p0, __p1) __extension__ ({ \
25147   int32x2_t __ret; \
25148   int32x2_t __s0 = __p0; \
25149   __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 2); \
25150   __ret; \
25151 })
25152 #else
25153 #define vshr_n_s32(__p0, __p1) __extension__ ({ \
25154   int32x2_t __ret; \
25155   int32x2_t __s0 = __p0; \
25156   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25157   __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 2); \
25158   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25159   __ret; \
25160 })
25161 #endif
25162 
25163 #define vshr_n_s64(__p0, __p1) __extension__ ({ \
25164   int64x1_t __ret; \
25165   int64x1_t __s0 = __p0; \
25166   __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
25167   __ret; \
25168 })
25169 #ifdef __LITTLE_ENDIAN__
25170 #define vshr_n_s16(__p0, __p1) __extension__ ({ \
25171   int16x4_t __ret; \
25172   int16x4_t __s0 = __p0; \
25173   __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 1); \
25174   __ret; \
25175 })
25176 #else
25177 #define vshr_n_s16(__p0, __p1) __extension__ ({ \
25178   int16x4_t __ret; \
25179   int16x4_t __s0 = __p0; \
25180   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25181   __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 1); \
25182   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25183   __ret; \
25184 })
25185 #endif
25186 
25187 #ifdef __LITTLE_ENDIAN__
25188 #define vshrn_n_u32(__p0, __p1) __extension__ ({ \
25189   uint16x4_t __ret; \
25190   uint32x4_t __s0 = __p0; \
25191   __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
25192   __ret; \
25193 })
25194 #else
25195 #define vshrn_n_u32(__p0, __p1) __extension__ ({ \
25196   uint16x4_t __ret; \
25197   uint32x4_t __s0 = __p0; \
25198   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25199   __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 17); \
25200   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25201   __ret; \
25202 })
25203 #define __noswap_vshrn_n_u32(__p0, __p1) __extension__ ({ \
25204   uint16x4_t __ret; \
25205   uint32x4_t __s0 = __p0; \
25206   __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
25207   __ret; \
25208 })
25209 #endif
25210 
25211 #ifdef __LITTLE_ENDIAN__
25212 #define vshrn_n_u64(__p0, __p1) __extension__ ({ \
25213   uint32x2_t __ret; \
25214   uint64x2_t __s0 = __p0; \
25215   __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
25216   __ret; \
25217 })
25218 #else
25219 #define vshrn_n_u64(__p0, __p1) __extension__ ({ \
25220   uint32x2_t __ret; \
25221   uint64x2_t __s0 = __p0; \
25222   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25223   __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 18); \
25224   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25225   __ret; \
25226 })
25227 #define __noswap_vshrn_n_u64(__p0, __p1) __extension__ ({ \
25228   uint32x2_t __ret; \
25229   uint64x2_t __s0 = __p0; \
25230   __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
25231   __ret; \
25232 })
25233 #endif
25234 
25235 #ifdef __LITTLE_ENDIAN__
25236 #define vshrn_n_u16(__p0, __p1) __extension__ ({ \
25237   uint8x8_t __ret; \
25238   uint16x8_t __s0 = __p0; \
25239   __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
25240   __ret; \
25241 })
25242 #else
25243 #define vshrn_n_u16(__p0, __p1) __extension__ ({ \
25244   uint8x8_t __ret; \
25245   uint16x8_t __s0 = __p0; \
25246   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25247   __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 16); \
25248   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25249   __ret; \
25250 })
25251 #define __noswap_vshrn_n_u16(__p0, __p1) __extension__ ({ \
25252   uint8x8_t __ret; \
25253   uint16x8_t __s0 = __p0; \
25254   __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
25255   __ret; \
25256 })
25257 #endif
25258 
25259 #ifdef __LITTLE_ENDIAN__
25260 #define vshrn_n_s32(__p0, __p1) __extension__ ({ \
25261   int16x4_t __ret; \
25262   int32x4_t __s0 = __p0; \
25263   __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
25264   __ret; \
25265 })
25266 #else
25267 #define vshrn_n_s32(__p0, __p1) __extension__ ({ \
25268   int16x4_t __ret; \
25269   int32x4_t __s0 = __p0; \
25270   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25271   __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 1); \
25272   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25273   __ret; \
25274 })
25275 #define __noswap_vshrn_n_s32(__p0, __p1) __extension__ ({ \
25276   int16x4_t __ret; \
25277   int32x4_t __s0 = __p0; \
25278   __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
25279   __ret; \
25280 })
25281 #endif
25282 
25283 #ifdef __LITTLE_ENDIAN__
25284 #define vshrn_n_s64(__p0, __p1) __extension__ ({ \
25285   int32x2_t __ret; \
25286   int64x2_t __s0 = __p0; \
25287   __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
25288   __ret; \
25289 })
25290 #else
25291 #define vshrn_n_s64(__p0, __p1) __extension__ ({ \
25292   int32x2_t __ret; \
25293   int64x2_t __s0 = __p0; \
25294   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25295   __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 2); \
25296   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25297   __ret; \
25298 })
25299 #define __noswap_vshrn_n_s64(__p0, __p1) __extension__ ({ \
25300   int32x2_t __ret; \
25301   int64x2_t __s0 = __p0; \
25302   __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
25303   __ret; \
25304 })
25305 #endif
25306 
25307 #ifdef __LITTLE_ENDIAN__
25308 #define vshrn_n_s16(__p0, __p1) __extension__ ({ \
25309   int8x8_t __ret; \
25310   int16x8_t __s0 = __p0; \
25311   __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
25312   __ret; \
25313 })
25314 #else
25315 #define vshrn_n_s16(__p0, __p1) __extension__ ({ \
25316   int8x8_t __ret; \
25317   int16x8_t __s0 = __p0; \
25318   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25319   __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 0); \
25320   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25321   __ret; \
25322 })
25323 #define __noswap_vshrn_n_s16(__p0, __p1) __extension__ ({ \
25324   int8x8_t __ret; \
25325   int16x8_t __s0 = __p0; \
25326   __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
25327   __ret; \
25328 })
25329 #endif
25330 
25331 #ifdef __LITTLE_ENDIAN__
25332 #define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
25333   poly8x8_t __ret; \
25334   poly8x8_t __s0 = __p0; \
25335   poly8x8_t __s1 = __p1; \
25336   __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
25337   __ret; \
25338 })
25339 #else
25340 #define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
25341   poly8x8_t __ret; \
25342   poly8x8_t __s0 = __p0; \
25343   poly8x8_t __s1 = __p1; \
25344   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25345   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25346   __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
25347   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25348   __ret; \
25349 })
25350 #endif
25351 
25352 #ifdef __LITTLE_ENDIAN__
25353 #define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
25354   poly16x4_t __ret; \
25355   poly16x4_t __s0 = __p0; \
25356   poly16x4_t __s1 = __p1; \
25357   __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
25358   __ret; \
25359 })
25360 #else
25361 #define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
25362   poly16x4_t __ret; \
25363   poly16x4_t __s0 = __p0; \
25364   poly16x4_t __s1 = __p1; \
25365   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25366   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25367   __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
25368   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25369   __ret; \
25370 })
25371 #endif
25372 
25373 #ifdef __LITTLE_ENDIAN__
25374 #define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
25375   poly8x16_t __ret; \
25376   poly8x16_t __s0 = __p0; \
25377   poly8x16_t __s1 = __p1; \
25378   __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
25379   __ret; \
25380 })
25381 #else
25382 #define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
25383   poly8x16_t __ret; \
25384   poly8x16_t __s0 = __p0; \
25385   poly8x16_t __s1 = __p1; \
25386   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25387   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25388   __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
25389   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25390   __ret; \
25391 })
25392 #endif
25393 
25394 #ifdef __LITTLE_ENDIAN__
25395 #define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
25396   poly16x8_t __ret; \
25397   poly16x8_t __s0 = __p0; \
25398   poly16x8_t __s1 = __p1; \
25399   __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
25400   __ret; \
25401 })
25402 #else
25403 #define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
25404   poly16x8_t __ret; \
25405   poly16x8_t __s0 = __p0; \
25406   poly16x8_t __s1 = __p1; \
25407   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25408   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25409   __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
25410   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25411   __ret; \
25412 })
25413 #endif
25414 
25415 #ifdef __LITTLE_ENDIAN__
25416 #define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
25417   uint8x16_t __ret; \
25418   uint8x16_t __s0 = __p0; \
25419   uint8x16_t __s1 = __p1; \
25420   __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
25421   __ret; \
25422 })
25423 #else
25424 #define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
25425   uint8x16_t __ret; \
25426   uint8x16_t __s0 = __p0; \
25427   uint8x16_t __s1 = __p1; \
25428   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25429   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25430   __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
25431   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25432   __ret; \
25433 })
25434 #endif
25435 
25436 #ifdef __LITTLE_ENDIAN__
25437 #define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
25438   uint32x4_t __ret; \
25439   uint32x4_t __s0 = __p0; \
25440   uint32x4_t __s1 = __p1; \
25441   __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
25442   __ret; \
25443 })
25444 #else
25445 #define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
25446   uint32x4_t __ret; \
25447   uint32x4_t __s0 = __p0; \
25448   uint32x4_t __s1 = __p1; \
25449   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25450   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25451   __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
25452   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25453   __ret; \
25454 })
25455 #endif
25456 
25457 #ifdef __LITTLE_ENDIAN__
25458 #define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
25459   uint64x2_t __ret; \
25460   uint64x2_t __s0 = __p0; \
25461   uint64x2_t __s1 = __p1; \
25462   __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
25463   __ret; \
25464 })
25465 #else
25466 #define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
25467   uint64x2_t __ret; \
25468   uint64x2_t __s0 = __p0; \
25469   uint64x2_t __s1 = __p1; \
25470   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25471   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25472   __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
25473   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25474   __ret; \
25475 })
25476 #endif
25477 
25478 #ifdef __LITTLE_ENDIAN__
25479 #define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
25480   uint16x8_t __ret; \
25481   uint16x8_t __s0 = __p0; \
25482   uint16x8_t __s1 = __p1; \
25483   __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
25484   __ret; \
25485 })
25486 #else
25487 #define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
25488   uint16x8_t __ret; \
25489   uint16x8_t __s0 = __p0; \
25490   uint16x8_t __s1 = __p1; \
25491   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25492   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25493   __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
25494   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25495   __ret; \
25496 })
25497 #endif
25498 
25499 #ifdef __LITTLE_ENDIAN__
25500 #define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
25501   int8x16_t __ret; \
25502   int8x16_t __s0 = __p0; \
25503   int8x16_t __s1 = __p1; \
25504   __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
25505   __ret; \
25506 })
25507 #else
25508 #define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
25509   int8x16_t __ret; \
25510   int8x16_t __s0 = __p0; \
25511   int8x16_t __s1 = __p1; \
25512   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25513   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25514   __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
25515   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25516   __ret; \
25517 })
25518 #endif
25519 
25520 #ifdef __LITTLE_ENDIAN__
25521 #define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
25522   int32x4_t __ret; \
25523   int32x4_t __s0 = __p0; \
25524   int32x4_t __s1 = __p1; \
25525   __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
25526   __ret; \
25527 })
25528 #else
25529 #define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
25530   int32x4_t __ret; \
25531   int32x4_t __s0 = __p0; \
25532   int32x4_t __s1 = __p1; \
25533   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25534   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25535   __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
25536   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25537   __ret; \
25538 })
25539 #endif
25540 
25541 #ifdef __LITTLE_ENDIAN__
25542 #define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
25543   int64x2_t __ret; \
25544   int64x2_t __s0 = __p0; \
25545   int64x2_t __s1 = __p1; \
25546   __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
25547   __ret; \
25548 })
25549 #else
25550 #define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
25551   int64x2_t __ret; \
25552   int64x2_t __s0 = __p0; \
25553   int64x2_t __s1 = __p1; \
25554   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25555   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25556   __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
25557   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25558   __ret; \
25559 })
25560 #endif
25561 
25562 #ifdef __LITTLE_ENDIAN__
25563 #define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
25564   int16x8_t __ret; \
25565   int16x8_t __s0 = __p0; \
25566   int16x8_t __s1 = __p1; \
25567   __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
25568   __ret; \
25569 })
25570 #else
25571 #define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
25572   int16x8_t __ret; \
25573   int16x8_t __s0 = __p0; \
25574   int16x8_t __s1 = __p1; \
25575   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25576   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25577   __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
25578   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25579   __ret; \
25580 })
25581 #endif
25582 
25583 #ifdef __LITTLE_ENDIAN__
25584 #define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
25585   uint8x8_t __ret; \
25586   uint8x8_t __s0 = __p0; \
25587   uint8x8_t __s1 = __p1; \
25588   __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
25589   __ret; \
25590 })
25591 #else
25592 #define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
25593   uint8x8_t __ret; \
25594   uint8x8_t __s0 = __p0; \
25595   uint8x8_t __s1 = __p1; \
25596   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25597   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25598   __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
25599   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25600   __ret; \
25601 })
25602 #endif
25603 
25604 #ifdef __LITTLE_ENDIAN__
25605 #define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
25606   uint32x2_t __ret; \
25607   uint32x2_t __s0 = __p0; \
25608   uint32x2_t __s1 = __p1; \
25609   __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
25610   __ret; \
25611 })
25612 #else
25613 #define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
25614   uint32x2_t __ret; \
25615   uint32x2_t __s0 = __p0; \
25616   uint32x2_t __s1 = __p1; \
25617   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25618   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25619   __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
25620   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25621   __ret; \
25622 })
25623 #endif
25624 
25625 #define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
25626   uint64x1_t __ret; \
25627   uint64x1_t __s0 = __p0; \
25628   uint64x1_t __s1 = __p1; \
25629   __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
25630   __ret; \
25631 })
25632 #ifdef __LITTLE_ENDIAN__
25633 #define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
25634   uint16x4_t __ret; \
25635   uint16x4_t __s0 = __p0; \
25636   uint16x4_t __s1 = __p1; \
25637   __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
25638   __ret; \
25639 })
25640 #else
25641 #define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
25642   uint16x4_t __ret; \
25643   uint16x4_t __s0 = __p0; \
25644   uint16x4_t __s1 = __p1; \
25645   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25646   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25647   __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
25648   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25649   __ret; \
25650 })
25651 #endif
25652 
25653 #ifdef __LITTLE_ENDIAN__
25654 #define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
25655   int8x8_t __ret; \
25656   int8x8_t __s0 = __p0; \
25657   int8x8_t __s1 = __p1; \
25658   __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
25659   __ret; \
25660 })
25661 #else
25662 #define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
25663   int8x8_t __ret; \
25664   int8x8_t __s0 = __p0; \
25665   int8x8_t __s1 = __p1; \
25666   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25667   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25668   __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
25669   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25670   __ret; \
25671 })
25672 #endif
25673 
25674 #ifdef __LITTLE_ENDIAN__
25675 #define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
25676   int32x2_t __ret; \
25677   int32x2_t __s0 = __p0; \
25678   int32x2_t __s1 = __p1; \
25679   __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
25680   __ret; \
25681 })
25682 #else
25683 #define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
25684   int32x2_t __ret; \
25685   int32x2_t __s0 = __p0; \
25686   int32x2_t __s1 = __p1; \
25687   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25688   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25689   __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
25690   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25691   __ret; \
25692 })
25693 #endif
25694 
25695 #define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
25696   int64x1_t __ret; \
25697   int64x1_t __s0 = __p0; \
25698   int64x1_t __s1 = __p1; \
25699   __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
25700   __ret; \
25701 })
25702 #ifdef __LITTLE_ENDIAN__
25703 #define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
25704   int16x4_t __ret; \
25705   int16x4_t __s0 = __p0; \
25706   int16x4_t __s1 = __p1; \
25707   __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
25708   __ret; \
25709 })
25710 #else
25711 #define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
25712   int16x4_t __ret; \
25713   int16x4_t __s0 = __p0; \
25714   int16x4_t __s1 = __p1; \
25715   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25716   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25717   __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
25718   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25719   __ret; \
25720 })
25721 #endif
25722 
25723 #ifdef __LITTLE_ENDIAN__
25724 #define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
25725   uint8x16_t __ret; \
25726   uint8x16_t __s0 = __p0; \
25727   uint8x16_t __s1 = __p1; \
25728   __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
25729   __ret; \
25730 })
25731 #else
25732 #define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
25733   uint8x16_t __ret; \
25734   uint8x16_t __s0 = __p0; \
25735   uint8x16_t __s1 = __p1; \
25736   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25737   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25738   __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
25739   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25740   __ret; \
25741 })
25742 #endif
25743 
25744 #ifdef __LITTLE_ENDIAN__
25745 #define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
25746   uint32x4_t __ret; \
25747   uint32x4_t __s0 = __p0; \
25748   uint32x4_t __s1 = __p1; \
25749   __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
25750   __ret; \
25751 })
25752 #else
25753 #define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
25754   uint32x4_t __ret; \
25755   uint32x4_t __s0 = __p0; \
25756   uint32x4_t __s1 = __p1; \
25757   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25758   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25759   __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
25760   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25761   __ret; \
25762 })
25763 #endif
25764 
25765 #ifdef __LITTLE_ENDIAN__
25766 #define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
25767   uint64x2_t __ret; \
25768   uint64x2_t __s0 = __p0; \
25769   uint64x2_t __s1 = __p1; \
25770   __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
25771   __ret; \
25772 })
25773 #else
25774 #define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
25775   uint64x2_t __ret; \
25776   uint64x2_t __s0 = __p0; \
25777   uint64x2_t __s1 = __p1; \
25778   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25779   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25780   __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
25781   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25782   __ret; \
25783 })
25784 #endif
25785 
25786 #ifdef __LITTLE_ENDIAN__
25787 #define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
25788   uint16x8_t __ret; \
25789   uint16x8_t __s0 = __p0; \
25790   uint16x8_t __s1 = __p1; \
25791   __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
25792   __ret; \
25793 })
25794 #else
25795 #define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
25796   uint16x8_t __ret; \
25797   uint16x8_t __s0 = __p0; \
25798   uint16x8_t __s1 = __p1; \
25799   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25800   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25801   __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
25802   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25803   __ret; \
25804 })
25805 #endif
25806 
25807 #ifdef __LITTLE_ENDIAN__
25808 #define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
25809   int8x16_t __ret; \
25810   int8x16_t __s0 = __p0; \
25811   int8x16_t __s1 = __p1; \
25812   __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
25813   __ret; \
25814 })
25815 #else
25816 #define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
25817   int8x16_t __ret; \
25818   int8x16_t __s0 = __p0; \
25819   int8x16_t __s1 = __p1; \
25820   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25821   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25822   __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
25823   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25824   __ret; \
25825 })
25826 #endif
25827 
25828 #ifdef __LITTLE_ENDIAN__
25829 #define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
25830   int32x4_t __ret; \
25831   int32x4_t __s0 = __p0; \
25832   int32x4_t __s1 = __p1; \
25833   __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
25834   __ret; \
25835 })
25836 #else
25837 #define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
25838   int32x4_t __ret; \
25839   int32x4_t __s0 = __p0; \
25840   int32x4_t __s1 = __p1; \
25841   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25842   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25843   __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
25844   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25845   __ret; \
25846 })
25847 #endif
25848 
25849 #ifdef __LITTLE_ENDIAN__
25850 #define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
25851   int64x2_t __ret; \
25852   int64x2_t __s0 = __p0; \
25853   int64x2_t __s1 = __p1; \
25854   __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
25855   __ret; \
25856 })
25857 #else
25858 #define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
25859   int64x2_t __ret; \
25860   int64x2_t __s0 = __p0; \
25861   int64x2_t __s1 = __p1; \
25862   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25863   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25864   __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
25865   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25866   __ret; \
25867 })
25868 #endif
25869 
25870 #ifdef __LITTLE_ENDIAN__
25871 #define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
25872   int16x8_t __ret; \
25873   int16x8_t __s0 = __p0; \
25874   int16x8_t __s1 = __p1; \
25875   __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
25876   __ret; \
25877 })
25878 #else
25879 #define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
25880   int16x8_t __ret; \
25881   int16x8_t __s0 = __p0; \
25882   int16x8_t __s1 = __p1; \
25883   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25884   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25885   __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
25886   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25887   __ret; \
25888 })
25889 #endif
25890 
25891 #ifdef __LITTLE_ENDIAN__
25892 #define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
25893   uint8x8_t __ret; \
25894   uint8x8_t __s0 = __p0; \
25895   uint8x8_t __s1 = __p1; \
25896   __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
25897   __ret; \
25898 })
25899 #else
25900 #define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
25901   uint8x8_t __ret; \
25902   uint8x8_t __s0 = __p0; \
25903   uint8x8_t __s1 = __p1; \
25904   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25905   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25906   __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
25907   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25908   __ret; \
25909 })
25910 #endif
25911 
25912 #ifdef __LITTLE_ENDIAN__
25913 #define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
25914   uint32x2_t __ret; \
25915   uint32x2_t __s0 = __p0; \
25916   uint32x2_t __s1 = __p1; \
25917   __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
25918   __ret; \
25919 })
25920 #else
25921 #define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
25922   uint32x2_t __ret; \
25923   uint32x2_t __s0 = __p0; \
25924   uint32x2_t __s1 = __p1; \
25925   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25926   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25927   __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
25928   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25929   __ret; \
25930 })
25931 #endif
25932 
25933 #define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
25934   uint64x1_t __ret; \
25935   uint64x1_t __s0 = __p0; \
25936   uint64x1_t __s1 = __p1; \
25937   __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
25938   __ret; \
25939 })
25940 #ifdef __LITTLE_ENDIAN__
25941 #define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
25942   uint16x4_t __ret; \
25943   uint16x4_t __s0 = __p0; \
25944   uint16x4_t __s1 = __p1; \
25945   __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
25946   __ret; \
25947 })
25948 #else
25949 #define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
25950   uint16x4_t __ret; \
25951   uint16x4_t __s0 = __p0; \
25952   uint16x4_t __s1 = __p1; \
25953   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25954   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25955   __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
25956   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25957   __ret; \
25958 })
25959 #endif
25960 
25961 #ifdef __LITTLE_ENDIAN__
25962 #define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
25963   int8x8_t __ret; \
25964   int8x8_t __s0 = __p0; \
25965   int8x8_t __s1 = __p1; \
25966   __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
25967   __ret; \
25968 })
25969 #else
25970 #define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
25971   int8x8_t __ret; \
25972   int8x8_t __s0 = __p0; \
25973   int8x8_t __s1 = __p1; \
25974   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25975   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25976   __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
25977   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25978   __ret; \
25979 })
25980 #endif
25981 
25982 #ifdef __LITTLE_ENDIAN__
25983 #define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
25984   int32x2_t __ret; \
25985   int32x2_t __s0 = __p0; \
25986   int32x2_t __s1 = __p1; \
25987   __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
25988   __ret; \
25989 })
25990 #else
25991 #define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
25992   int32x2_t __ret; \
25993   int32x2_t __s0 = __p0; \
25994   int32x2_t __s1 = __p1; \
25995   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25996   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25997   __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
25998   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25999   __ret; \
26000 })
26001 #endif
26002 
26003 #define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
26004   int64x1_t __ret; \
26005   int64x1_t __s0 = __p0; \
26006   int64x1_t __s1 = __p1; \
26007   __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
26008   __ret; \
26009 })
26010 #ifdef __LITTLE_ENDIAN__
26011 #define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
26012   int16x4_t __ret; \
26013   int16x4_t __s0 = __p0; \
26014   int16x4_t __s1 = __p1; \
26015   __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
26016   __ret; \
26017 })
26018 #else
26019 #define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
26020   int16x4_t __ret; \
26021   int16x4_t __s0 = __p0; \
26022   int16x4_t __s1 = __p1; \
26023   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26024   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26025   __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
26026   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26027   __ret; \
26028 })
26029 #endif
26030 
26031 #ifdef __LITTLE_ENDIAN__
26032 #define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
26033   poly8x8_t __ret; \
26034   poly8x8_t __s0 = __p0; \
26035   poly8x8_t __s1 = __p1; \
26036   __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
26037   __ret; \
26038 })
26039 #else
26040 #define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
26041   poly8x8_t __ret; \
26042   poly8x8_t __s0 = __p0; \
26043   poly8x8_t __s1 = __p1; \
26044   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26045   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26046   __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
26047   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26048   __ret; \
26049 })
26050 #endif
26051 
26052 #ifdef __LITTLE_ENDIAN__
26053 #define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
26054   poly16x4_t __ret; \
26055   poly16x4_t __s0 = __p0; \
26056   poly16x4_t __s1 = __p1; \
26057   __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
26058   __ret; \
26059 })
26060 #else
26061 #define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
26062   poly16x4_t __ret; \
26063   poly16x4_t __s0 = __p0; \
26064   poly16x4_t __s1 = __p1; \
26065   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26066   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26067   __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
26068   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26069   __ret; \
26070 })
26071 #endif
26072 
26073 #ifdef __LITTLE_ENDIAN__
26074 #define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
26075   poly8x16_t __ret; \
26076   poly8x16_t __s0 = __p0; \
26077   poly8x16_t __s1 = __p1; \
26078   __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
26079   __ret; \
26080 })
26081 #else
26082 #define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
26083   poly8x16_t __ret; \
26084   poly8x16_t __s0 = __p0; \
26085   poly8x16_t __s1 = __p1; \
26086   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26087   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26088   __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
26089   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26090   __ret; \
26091 })
26092 #endif
26093 
26094 #ifdef __LITTLE_ENDIAN__
26095 #define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
26096   poly16x8_t __ret; \
26097   poly16x8_t __s0 = __p0; \
26098   poly16x8_t __s1 = __p1; \
26099   __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
26100   __ret; \
26101 })
26102 #else
26103 #define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
26104   poly16x8_t __ret; \
26105   poly16x8_t __s0 = __p0; \
26106   poly16x8_t __s1 = __p1; \
26107   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26108   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26109   __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
26110   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26111   __ret; \
26112 })
26113 #endif
26114 
26115 #ifdef __LITTLE_ENDIAN__
26116 #define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
26117   uint8x16_t __ret; \
26118   uint8x16_t __s0 = __p0; \
26119   uint8x16_t __s1 = __p1; \
26120   __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
26121   __ret; \
26122 })
26123 #else
26124 #define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
26125   uint8x16_t __ret; \
26126   uint8x16_t __s0 = __p0; \
26127   uint8x16_t __s1 = __p1; \
26128   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26129   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26130   __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
26131   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26132   __ret; \
26133 })
26134 #endif
26135 
26136 #ifdef __LITTLE_ENDIAN__
26137 #define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
26138   uint32x4_t __ret; \
26139   uint32x4_t __s0 = __p0; \
26140   uint32x4_t __s1 = __p1; \
26141   __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
26142   __ret; \
26143 })
26144 #else
26145 #define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
26146   uint32x4_t __ret; \
26147   uint32x4_t __s0 = __p0; \
26148   uint32x4_t __s1 = __p1; \
26149   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26150   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26151   __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
26152   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26153   __ret; \
26154 })
26155 #endif
26156 
26157 #ifdef __LITTLE_ENDIAN__
26158 #define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
26159   uint64x2_t __ret; \
26160   uint64x2_t __s0 = __p0; \
26161   uint64x2_t __s1 = __p1; \
26162   __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
26163   __ret; \
26164 })
26165 #else
26166 #define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
26167   uint64x2_t __ret; \
26168   uint64x2_t __s0 = __p0; \
26169   uint64x2_t __s1 = __p1; \
26170   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26171   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26172   __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
26173   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26174   __ret; \
26175 })
26176 #endif
26177 
26178 #ifdef __LITTLE_ENDIAN__
26179 #define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
26180   uint16x8_t __ret; \
26181   uint16x8_t __s0 = __p0; \
26182   uint16x8_t __s1 = __p1; \
26183   __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
26184   __ret; \
26185 })
26186 #else
26187 #define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
26188   uint16x8_t __ret; \
26189   uint16x8_t __s0 = __p0; \
26190   uint16x8_t __s1 = __p1; \
26191   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26192   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26193   __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
26194   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26195   __ret; \
26196 })
26197 #endif
26198 
26199 #ifdef __LITTLE_ENDIAN__
26200 #define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
26201   int8x16_t __ret; \
26202   int8x16_t __s0 = __p0; \
26203   int8x16_t __s1 = __p1; \
26204   __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
26205   __ret; \
26206 })
26207 #else
26208 #define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
26209   int8x16_t __ret; \
26210   int8x16_t __s0 = __p0; \
26211   int8x16_t __s1 = __p1; \
26212   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26213   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26214   __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
26215   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26216   __ret; \
26217 })
26218 #endif
26219 
26220 #ifdef __LITTLE_ENDIAN__
26221 #define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
26222   int32x4_t __ret; \
26223   int32x4_t __s0 = __p0; \
26224   int32x4_t __s1 = __p1; \
26225   __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
26226   __ret; \
26227 })
26228 #else
26229 #define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
26230   int32x4_t __ret; \
26231   int32x4_t __s0 = __p0; \
26232   int32x4_t __s1 = __p1; \
26233   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26234   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26235   __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
26236   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26237   __ret; \
26238 })
26239 #endif
26240 
26241 #ifdef __LITTLE_ENDIAN__
26242 #define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
26243   int64x2_t __ret; \
26244   int64x2_t __s0 = __p0; \
26245   int64x2_t __s1 = __p1; \
26246   __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
26247   __ret; \
26248 })
26249 #else
26250 #define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
26251   int64x2_t __ret; \
26252   int64x2_t __s0 = __p0; \
26253   int64x2_t __s1 = __p1; \
26254   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26255   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26256   __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
26257   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26258   __ret; \
26259 })
26260 #endif
26261 
26262 #ifdef __LITTLE_ENDIAN__
26263 #define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
26264   int16x8_t __ret; \
26265   int16x8_t __s0 = __p0; \
26266   int16x8_t __s1 = __p1; \
26267   __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
26268   __ret; \
26269 })
26270 #else
26271 #define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
26272   int16x8_t __ret; \
26273   int16x8_t __s0 = __p0; \
26274   int16x8_t __s1 = __p1; \
26275   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26276   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26277   __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
26278   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26279   __ret; \
26280 })
26281 #endif
26282 
26283 #ifdef __LITTLE_ENDIAN__
26284 #define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
26285   uint8x8_t __ret; \
26286   uint8x8_t __s0 = __p0; \
26287   uint8x8_t __s1 = __p1; \
26288   __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
26289   __ret; \
26290 })
26291 #else
26292 #define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
26293   uint8x8_t __ret; \
26294   uint8x8_t __s0 = __p0; \
26295   uint8x8_t __s1 = __p1; \
26296   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26297   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26298   __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
26299   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26300   __ret; \
26301 })
26302 #endif
26303 
26304 #ifdef __LITTLE_ENDIAN__
26305 #define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
26306   uint32x2_t __ret; \
26307   uint32x2_t __s0 = __p0; \
26308   uint32x2_t __s1 = __p1; \
26309   __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
26310   __ret; \
26311 })
26312 #else
26313 #define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
26314   uint32x2_t __ret; \
26315   uint32x2_t __s0 = __p0; \
26316   uint32x2_t __s1 = __p1; \
26317   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26318   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26319   __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
26320   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26321   __ret; \
26322 })
26323 #endif
26324 
26325 #define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
26326   uint64x1_t __ret; \
26327   uint64x1_t __s0 = __p0; \
26328   uint64x1_t __s1 = __p1; \
26329   __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
26330   __ret; \
26331 })
26332 #ifdef __LITTLE_ENDIAN__
26333 #define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
26334   uint16x4_t __ret; \
26335   uint16x4_t __s0 = __p0; \
26336   uint16x4_t __s1 = __p1; \
26337   __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
26338   __ret; \
26339 })
26340 #else
26341 #define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
26342   uint16x4_t __ret; \
26343   uint16x4_t __s0 = __p0; \
26344   uint16x4_t __s1 = __p1; \
26345   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26346   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26347   __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
26348   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26349   __ret; \
26350 })
26351 #endif
26352 
26353 #ifdef __LITTLE_ENDIAN__
26354 #define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
26355   int8x8_t __ret; \
26356   int8x8_t __s0 = __p0; \
26357   int8x8_t __s1 = __p1; \
26358   __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
26359   __ret; \
26360 })
26361 #else
26362 #define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
26363   int8x8_t __ret; \
26364   int8x8_t __s0 = __p0; \
26365   int8x8_t __s1 = __p1; \
26366   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26367   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26368   __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
26369   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26370   __ret; \
26371 })
26372 #endif
26373 
26374 #ifdef __LITTLE_ENDIAN__
26375 #define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
26376   int32x2_t __ret; \
26377   int32x2_t __s0 = __p0; \
26378   int32x2_t __s1 = __p1; \
26379   __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
26380   __ret; \
26381 })
26382 #else
26383 #define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
26384   int32x2_t __ret; \
26385   int32x2_t __s0 = __p0; \
26386   int32x2_t __s1 = __p1; \
26387   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26388   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26389   __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
26390   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26391   __ret; \
26392 })
26393 #endif
26394 
26395 #define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
26396   int64x1_t __ret; \
26397   int64x1_t __s0 = __p0; \
26398   int64x1_t __s1 = __p1; \
26399   __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
26400   __ret; \
26401 })
26402 #ifdef __LITTLE_ENDIAN__
26403 #define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
26404   int16x4_t __ret; \
26405   int16x4_t __s0 = __p0; \
26406   int16x4_t __s1 = __p1; \
26407   __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
26408   __ret; \
26409 })
26410 #else
26411 #define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
26412   int16x4_t __ret; \
26413   int16x4_t __s0 = __p0; \
26414   int16x4_t __s1 = __p1; \
26415   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26416   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26417   __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
26418   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26419   __ret; \
26420 })
26421 #endif
26422 
26423 #ifdef __LITTLE_ENDIAN__
26424 #define vst1_p8(__p0, __p1) __extension__ ({ \
26425   poly8x8_t __s1 = __p1; \
26426   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 4); \
26427 })
26428 #else
26429 #define vst1_p8(__p0, __p1) __extension__ ({ \
26430   poly8x8_t __s1 = __p1; \
26431   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26432   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 4); \
26433 })
26434 #endif
26435 
26436 #ifdef __LITTLE_ENDIAN__
26437 #define vst1_p16(__p0, __p1) __extension__ ({ \
26438   poly16x4_t __s1 = __p1; \
26439   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 5); \
26440 })
26441 #else
26442 #define vst1_p16(__p0, __p1) __extension__ ({ \
26443   poly16x4_t __s1 = __p1; \
26444   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26445   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 5); \
26446 })
26447 #endif
26448 
26449 #ifdef __LITTLE_ENDIAN__
26450 #define vst1q_p8(__p0, __p1) __extension__ ({ \
26451   poly8x16_t __s1 = __p1; \
26452   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 36); \
26453 })
26454 #else
26455 #define vst1q_p8(__p0, __p1) __extension__ ({ \
26456   poly8x16_t __s1 = __p1; \
26457   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26458   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 36); \
26459 })
26460 #endif
26461 
26462 #ifdef __LITTLE_ENDIAN__
26463 #define vst1q_p16(__p0, __p1) __extension__ ({ \
26464   poly16x8_t __s1 = __p1; \
26465   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 37); \
26466 })
26467 #else
26468 #define vst1q_p16(__p0, __p1) __extension__ ({ \
26469   poly16x8_t __s1 = __p1; \
26470   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26471   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 37); \
26472 })
26473 #endif
26474 
26475 #ifdef __LITTLE_ENDIAN__
26476 #define vst1q_u8(__p0, __p1) __extension__ ({ \
26477   uint8x16_t __s1 = __p1; \
26478   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 48); \
26479 })
26480 #else
26481 #define vst1q_u8(__p0, __p1) __extension__ ({ \
26482   uint8x16_t __s1 = __p1; \
26483   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26484   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 48); \
26485 })
26486 #endif
26487 
26488 #ifdef __LITTLE_ENDIAN__
26489 #define vst1q_u32(__p0, __p1) __extension__ ({ \
26490   uint32x4_t __s1 = __p1; \
26491   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 50); \
26492 })
26493 #else
26494 #define vst1q_u32(__p0, __p1) __extension__ ({ \
26495   uint32x4_t __s1 = __p1; \
26496   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26497   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 50); \
26498 })
26499 #endif
26500 
26501 #ifdef __LITTLE_ENDIAN__
26502 #define vst1q_u64(__p0, __p1) __extension__ ({ \
26503   uint64x2_t __s1 = __p1; \
26504   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 51); \
26505 })
26506 #else
26507 #define vst1q_u64(__p0, __p1) __extension__ ({ \
26508   uint64x2_t __s1 = __p1; \
26509   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26510   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 51); \
26511 })
26512 #endif
26513 
26514 #ifdef __LITTLE_ENDIAN__
26515 #define vst1q_u16(__p0, __p1) __extension__ ({ \
26516   uint16x8_t __s1 = __p1; \
26517   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 49); \
26518 })
26519 #else
26520 #define vst1q_u16(__p0, __p1) __extension__ ({ \
26521   uint16x8_t __s1 = __p1; \
26522   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26523   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 49); \
26524 })
26525 #endif
26526 
26527 #ifdef __LITTLE_ENDIAN__
26528 #define vst1q_s8(__p0, __p1) __extension__ ({ \
26529   int8x16_t __s1 = __p1; \
26530   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 32); \
26531 })
26532 #else
26533 #define vst1q_s8(__p0, __p1) __extension__ ({ \
26534   int8x16_t __s1 = __p1; \
26535   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26536   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 32); \
26537 })
26538 #endif
26539 
26540 #ifdef __LITTLE_ENDIAN__
26541 #define vst1q_f32(__p0, __p1) __extension__ ({ \
26542   float32x4_t __s1 = __p1; \
26543   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 41); \
26544 })
26545 #else
26546 #define vst1q_f32(__p0, __p1) __extension__ ({ \
26547   float32x4_t __s1 = __p1; \
26548   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26549   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 41); \
26550 })
26551 #endif
26552 
26553 #ifdef __LITTLE_ENDIAN__
26554 #define vst1q_s32(__p0, __p1) __extension__ ({ \
26555   int32x4_t __s1 = __p1; \
26556   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 34); \
26557 })
26558 #else
26559 #define vst1q_s32(__p0, __p1) __extension__ ({ \
26560   int32x4_t __s1 = __p1; \
26561   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26562   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 34); \
26563 })
26564 #endif
26565 
26566 #ifdef __LITTLE_ENDIAN__
26567 #define vst1q_s64(__p0, __p1) __extension__ ({ \
26568   int64x2_t __s1 = __p1; \
26569   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 35); \
26570 })
26571 #else
26572 #define vst1q_s64(__p0, __p1) __extension__ ({ \
26573   int64x2_t __s1 = __p1; \
26574   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26575   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 35); \
26576 })
26577 #endif
26578 
26579 #ifdef __LITTLE_ENDIAN__
26580 #define vst1q_s16(__p0, __p1) __extension__ ({ \
26581   int16x8_t __s1 = __p1; \
26582   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 33); \
26583 })
26584 #else
26585 #define vst1q_s16(__p0, __p1) __extension__ ({ \
26586   int16x8_t __s1 = __p1; \
26587   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26588   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 33); \
26589 })
26590 #endif
26591 
26592 #ifdef __LITTLE_ENDIAN__
26593 #define vst1_u8(__p0, __p1) __extension__ ({ \
26594   uint8x8_t __s1 = __p1; \
26595   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 16); \
26596 })
26597 #else
26598 #define vst1_u8(__p0, __p1) __extension__ ({ \
26599   uint8x8_t __s1 = __p1; \
26600   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26601   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 16); \
26602 })
26603 #endif
26604 
26605 #ifdef __LITTLE_ENDIAN__
26606 #define vst1_u32(__p0, __p1) __extension__ ({ \
26607   uint32x2_t __s1 = __p1; \
26608   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 18); \
26609 })
26610 #else
26611 #define vst1_u32(__p0, __p1) __extension__ ({ \
26612   uint32x2_t __s1 = __p1; \
26613   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26614   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 18); \
26615 })
26616 #endif
26617 
26618 #define vst1_u64(__p0, __p1) __extension__ ({ \
26619   uint64x1_t __s1 = __p1; \
26620   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
26621 })
26622 #ifdef __LITTLE_ENDIAN__
26623 #define vst1_u16(__p0, __p1) __extension__ ({ \
26624   uint16x4_t __s1 = __p1; \
26625   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 17); \
26626 })
26627 #else
26628 #define vst1_u16(__p0, __p1) __extension__ ({ \
26629   uint16x4_t __s1 = __p1; \
26630   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26631   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 17); \
26632 })
26633 #endif
26634 
26635 #ifdef __LITTLE_ENDIAN__
26636 #define vst1_s8(__p0, __p1) __extension__ ({ \
26637   int8x8_t __s1 = __p1; \
26638   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 0); \
26639 })
26640 #else
26641 #define vst1_s8(__p0, __p1) __extension__ ({ \
26642   int8x8_t __s1 = __p1; \
26643   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26644   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 0); \
26645 })
26646 #endif
26647 
26648 #ifdef __LITTLE_ENDIAN__
26649 #define vst1_f32(__p0, __p1) __extension__ ({ \
26650   float32x2_t __s1 = __p1; \
26651   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 9); \
26652 })
26653 #else
26654 #define vst1_f32(__p0, __p1) __extension__ ({ \
26655   float32x2_t __s1 = __p1; \
26656   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26657   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 9); \
26658 })
26659 #endif
26660 
26661 #ifdef __LITTLE_ENDIAN__
26662 #define vst1_s32(__p0, __p1) __extension__ ({ \
26663   int32x2_t __s1 = __p1; \
26664   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 2); \
26665 })
26666 #else
26667 #define vst1_s32(__p0, __p1) __extension__ ({ \
26668   int32x2_t __s1 = __p1; \
26669   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26670   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 2); \
26671 })
26672 #endif
26673 
26674 #define vst1_s64(__p0, __p1) __extension__ ({ \
26675   int64x1_t __s1 = __p1; \
26676   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
26677 })
26678 #ifdef __LITTLE_ENDIAN__
26679 #define vst1_s16(__p0, __p1) __extension__ ({ \
26680   int16x4_t __s1 = __p1; \
26681   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 1); \
26682 })
26683 #else
26684 #define vst1_s16(__p0, __p1) __extension__ ({ \
26685   int16x4_t __s1 = __p1; \
26686   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26687   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 1); \
26688 })
26689 #endif
26690 
26691 #ifdef __LITTLE_ENDIAN__
26692 #define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26693   poly8x8_t __s1 = __p1; \
26694   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
26695 })
26696 #else
26697 #define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26698   poly8x8_t __s1 = __p1; \
26699   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26700   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
26701 })
26702 #endif
26703 
26704 #ifdef __LITTLE_ENDIAN__
26705 #define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26706   poly16x4_t __s1 = __p1; \
26707   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
26708 })
26709 #else
26710 #define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26711   poly16x4_t __s1 = __p1; \
26712   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26713   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
26714 })
26715 #endif
26716 
26717 #ifdef __LITTLE_ENDIAN__
26718 #define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26719   poly8x16_t __s1 = __p1; \
26720   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
26721 })
26722 #else
26723 #define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26724   poly8x16_t __s1 = __p1; \
26725   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26726   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
26727 })
26728 #endif
26729 
26730 #ifdef __LITTLE_ENDIAN__
26731 #define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26732   poly16x8_t __s1 = __p1; \
26733   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
26734 })
26735 #else
26736 #define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26737   poly16x8_t __s1 = __p1; \
26738   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26739   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
26740 })
26741 #endif
26742 
26743 #ifdef __LITTLE_ENDIAN__
26744 #define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26745   uint8x16_t __s1 = __p1; \
26746   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
26747 })
26748 #else
26749 #define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26750   uint8x16_t __s1 = __p1; \
26751   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26752   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
26753 })
26754 #endif
26755 
26756 #ifdef __LITTLE_ENDIAN__
26757 #define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26758   uint32x4_t __s1 = __p1; \
26759   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
26760 })
26761 #else
26762 #define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26763   uint32x4_t __s1 = __p1; \
26764   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26765   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
26766 })
26767 #endif
26768 
26769 #ifdef __LITTLE_ENDIAN__
26770 #define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
26771   uint64x2_t __s1 = __p1; \
26772   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
26773 })
26774 #else
26775 #define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
26776   uint64x2_t __s1 = __p1; \
26777   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26778   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
26779 })
26780 #endif
26781 
26782 #ifdef __LITTLE_ENDIAN__
26783 #define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26784   uint16x8_t __s1 = __p1; \
26785   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
26786 })
26787 #else
26788 #define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26789   uint16x8_t __s1 = __p1; \
26790   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26791   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
26792 })
26793 #endif
26794 
26795 #ifdef __LITTLE_ENDIAN__
26796 #define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26797   int8x16_t __s1 = __p1; \
26798   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
26799 })
26800 #else
26801 #define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26802   int8x16_t __s1 = __p1; \
26803   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26804   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
26805 })
26806 #endif
26807 
26808 #ifdef __LITTLE_ENDIAN__
26809 #define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26810   float32x4_t __s1 = __p1; \
26811   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
26812 })
26813 #else
26814 #define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26815   float32x4_t __s1 = __p1; \
26816   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26817   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
26818 })
26819 #endif
26820 
26821 #ifdef __LITTLE_ENDIAN__
26822 #define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26823   int32x4_t __s1 = __p1; \
26824   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
26825 })
26826 #else
26827 #define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26828   int32x4_t __s1 = __p1; \
26829   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26830   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
26831 })
26832 #endif
26833 
26834 #ifdef __LITTLE_ENDIAN__
26835 #define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
26836   int64x2_t __s1 = __p1; \
26837   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
26838 })
26839 #else
26840 #define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
26841   int64x2_t __s1 = __p1; \
26842   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26843   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
26844 })
26845 #endif
26846 
26847 #ifdef __LITTLE_ENDIAN__
26848 #define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26849   int16x8_t __s1 = __p1; \
26850   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
26851 })
26852 #else
26853 #define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26854   int16x8_t __s1 = __p1; \
26855   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26856   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
26857 })
26858 #endif
26859 
26860 #ifdef __LITTLE_ENDIAN__
26861 #define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26862   uint8x8_t __s1 = __p1; \
26863   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
26864 })
26865 #else
26866 #define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26867   uint8x8_t __s1 = __p1; \
26868   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26869   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
26870 })
26871 #endif
26872 
26873 #ifdef __LITTLE_ENDIAN__
26874 #define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26875   uint32x2_t __s1 = __p1; \
26876   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
26877 })
26878 #else
26879 #define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26880   uint32x2_t __s1 = __p1; \
26881   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26882   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
26883 })
26884 #endif
26885 
26886 #define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
26887   uint64x1_t __s1 = __p1; \
26888   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
26889 })
26890 #ifdef __LITTLE_ENDIAN__
26891 #define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26892   uint16x4_t __s1 = __p1; \
26893   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
26894 })
26895 #else
26896 #define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26897   uint16x4_t __s1 = __p1; \
26898   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26899   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
26900 })
26901 #endif
26902 
26903 #ifdef __LITTLE_ENDIAN__
26904 #define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26905   int8x8_t __s1 = __p1; \
26906   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
26907 })
26908 #else
26909 #define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26910   int8x8_t __s1 = __p1; \
26911   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26912   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
26913 })
26914 #endif
26915 
26916 #ifdef __LITTLE_ENDIAN__
26917 #define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26918   float32x2_t __s1 = __p1; \
26919   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
26920 })
26921 #else
26922 #define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26923   float32x2_t __s1 = __p1; \
26924   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26925   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
26926 })
26927 #endif
26928 
26929 #ifdef __LITTLE_ENDIAN__
26930 #define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26931   int32x2_t __s1 = __p1; \
26932   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
26933 })
26934 #else
26935 #define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26936   int32x2_t __s1 = __p1; \
26937   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26938   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
26939 })
26940 #endif
26941 
26942 #define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
26943   int64x1_t __s1 = __p1; \
26944   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
26945 })
26946 #ifdef __LITTLE_ENDIAN__
26947 #define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26948   int16x4_t __s1 = __p1; \
26949   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
26950 })
26951 #else
26952 #define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26953   int16x4_t __s1 = __p1; \
26954   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26955   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
26956 })
26957 #endif
26958 
26959 #ifdef __LITTLE_ENDIAN__
26960 #define vst1_p8_x2(__p0, __p1) __extension__ ({ \
26961   poly8x8x2_t __s1 = __p1; \
26962   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
26963 })
26964 #else
26965 #define vst1_p8_x2(__p0, __p1) __extension__ ({ \
26966   poly8x8x2_t __s1 = __p1; \
26967   poly8x8x2_t __rev1; \
26968   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26969   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26970   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
26971 })
26972 #endif
26973 
26974 #ifdef __LITTLE_ENDIAN__
26975 #define vst1_p16_x2(__p0, __p1) __extension__ ({ \
26976   poly16x4x2_t __s1 = __p1; \
26977   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
26978 })
26979 #else
26980 #define vst1_p16_x2(__p0, __p1) __extension__ ({ \
26981   poly16x4x2_t __s1 = __p1; \
26982   poly16x4x2_t __rev1; \
26983   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26984   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26985   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
26986 })
26987 #endif
26988 
26989 #ifdef __LITTLE_ENDIAN__
26990 #define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
26991   poly8x16x2_t __s1 = __p1; \
26992   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
26993 })
26994 #else
26995 #define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
26996   poly8x16x2_t __s1 = __p1; \
26997   poly8x16x2_t __rev1; \
26998   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26999   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27000   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
27001 })
27002 #endif
27003 
27004 #ifdef __LITTLE_ENDIAN__
27005 #define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
27006   poly16x8x2_t __s1 = __p1; \
27007   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
27008 })
27009 #else
27010 #define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
27011   poly16x8x2_t __s1 = __p1; \
27012   poly16x8x2_t __rev1; \
27013   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27014   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27015   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
27016 })
27017 #endif
27018 
27019 #ifdef __LITTLE_ENDIAN__
27020 #define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
27021   uint8x16x2_t __s1 = __p1; \
27022   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
27023 })
27024 #else
27025 #define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
27026   uint8x16x2_t __s1 = __p1; \
27027   uint8x16x2_t __rev1; \
27028   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27029   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27030   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
27031 })
27032 #endif
27033 
27034 #ifdef __LITTLE_ENDIAN__
27035 #define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
27036   uint32x4x2_t __s1 = __p1; \
27037   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
27038 })
27039 #else
27040 #define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
27041   uint32x4x2_t __s1 = __p1; \
27042   uint32x4x2_t __rev1; \
27043   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27044   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27045   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
27046 })
27047 #endif
27048 
27049 #ifdef __LITTLE_ENDIAN__
27050 #define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
27051   uint64x2x2_t __s1 = __p1; \
27052   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
27053 })
27054 #else
27055 #define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
27056   uint64x2x2_t __s1 = __p1; \
27057   uint64x2x2_t __rev1; \
27058   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27059   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27060   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
27061 })
27062 #endif
27063 
27064 #ifdef __LITTLE_ENDIAN__
27065 #define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
27066   uint16x8x2_t __s1 = __p1; \
27067   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
27068 })
27069 #else
27070 #define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
27071   uint16x8x2_t __s1 = __p1; \
27072   uint16x8x2_t __rev1; \
27073   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27074   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27075   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
27076 })
27077 #endif
27078 
27079 #ifdef __LITTLE_ENDIAN__
27080 #define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
27081   int8x16x2_t __s1 = __p1; \
27082   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
27083 })
27084 #else
27085 #define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
27086   int8x16x2_t __s1 = __p1; \
27087   int8x16x2_t __rev1; \
27088   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27089   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27090   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
27091 })
27092 #endif
27093 
27094 #ifdef __LITTLE_ENDIAN__
27095 #define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
27096   float32x4x2_t __s1 = __p1; \
27097   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 41); \
27098 })
27099 #else
27100 #define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
27101   float32x4x2_t __s1 = __p1; \
27102   float32x4x2_t __rev1; \
27103   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27104   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27105   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 41); \
27106 })
27107 #endif
27108 
27109 #ifdef __LITTLE_ENDIAN__
27110 #define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
27111   int32x4x2_t __s1 = __p1; \
27112   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 34); \
27113 })
27114 #else
27115 #define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
27116   int32x4x2_t __s1 = __p1; \
27117   int32x4x2_t __rev1; \
27118   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27119   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27120   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 34); \
27121 })
27122 #endif
27123 
27124 #ifdef __LITTLE_ENDIAN__
27125 #define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
27126   int64x2x2_t __s1 = __p1; \
27127   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 35); \
27128 })
27129 #else
27130 #define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
27131   int64x2x2_t __s1 = __p1; \
27132   int64x2x2_t __rev1; \
27133   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27134   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27135   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 35); \
27136 })
27137 #endif
27138 
27139 #ifdef __LITTLE_ENDIAN__
27140 #define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
27141   int16x8x2_t __s1 = __p1; \
27142   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 33); \
27143 })
27144 #else
27145 #define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
27146   int16x8x2_t __s1 = __p1; \
27147   int16x8x2_t __rev1; \
27148   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27149   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27150   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 33); \
27151 })
27152 #endif
27153 
27154 #ifdef __LITTLE_ENDIAN__
27155 #define vst1_u8_x2(__p0, __p1) __extension__ ({ \
27156   uint8x8x2_t __s1 = __p1; \
27157   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
27158 })
27159 #else
27160 #define vst1_u8_x2(__p0, __p1) __extension__ ({ \
27161   uint8x8x2_t __s1 = __p1; \
27162   uint8x8x2_t __rev1; \
27163   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27164   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27165   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
27166 })
27167 #endif
27168 
27169 #ifdef __LITTLE_ENDIAN__
27170 #define vst1_u32_x2(__p0, __p1) __extension__ ({ \
27171   uint32x2x2_t __s1 = __p1; \
27172   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
27173 })
27174 #else
27175 #define vst1_u32_x2(__p0, __p1) __extension__ ({ \
27176   uint32x2x2_t __s1 = __p1; \
27177   uint32x2x2_t __rev1; \
27178   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27179   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27180   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
27181 })
27182 #endif
27183 
27184 #define vst1_u64_x2(__p0, __p1) __extension__ ({ \
27185   uint64x1x2_t __s1 = __p1; \
27186   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
27187 })
27188 #ifdef __LITTLE_ENDIAN__
27189 #define vst1_u16_x2(__p0, __p1) __extension__ ({ \
27190   uint16x4x2_t __s1 = __p1; \
27191   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
27192 })
27193 #else
27194 #define vst1_u16_x2(__p0, __p1) __extension__ ({ \
27195   uint16x4x2_t __s1 = __p1; \
27196   uint16x4x2_t __rev1; \
27197   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27198   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27199   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
27200 })
27201 #endif
27202 
27203 #ifdef __LITTLE_ENDIAN__
27204 #define vst1_s8_x2(__p0, __p1) __extension__ ({ \
27205   int8x8x2_t __s1 = __p1; \
27206   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
27207 })
27208 #else
27209 #define vst1_s8_x2(__p0, __p1) __extension__ ({ \
27210   int8x8x2_t __s1 = __p1; \
27211   int8x8x2_t __rev1; \
27212   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27213   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27214   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
27215 })
27216 #endif
27217 
27218 #ifdef __LITTLE_ENDIAN__
27219 #define vst1_f32_x2(__p0, __p1) __extension__ ({ \
27220   float32x2x2_t __s1 = __p1; \
27221   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 9); \
27222 })
27223 #else
27224 #define vst1_f32_x2(__p0, __p1) __extension__ ({ \
27225   float32x2x2_t __s1 = __p1; \
27226   float32x2x2_t __rev1; \
27227   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27228   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27229   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 9); \
27230 })
27231 #endif
27232 
27233 #ifdef __LITTLE_ENDIAN__
27234 #define vst1_s32_x2(__p0, __p1) __extension__ ({ \
27235   int32x2x2_t __s1 = __p1; \
27236   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 2); \
27237 })
27238 #else
27239 #define vst1_s32_x2(__p0, __p1) __extension__ ({ \
27240   int32x2x2_t __s1 = __p1; \
27241   int32x2x2_t __rev1; \
27242   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27243   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27244   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 2); \
27245 })
27246 #endif
27247 
27248 #define vst1_s64_x2(__p0, __p1) __extension__ ({ \
27249   int64x1x2_t __s1 = __p1; \
27250   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 3); \
27251 })
27252 #ifdef __LITTLE_ENDIAN__
27253 #define vst1_s16_x2(__p0, __p1) __extension__ ({ \
27254   int16x4x2_t __s1 = __p1; \
27255   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 1); \
27256 })
27257 #else
27258 #define vst1_s16_x2(__p0, __p1) __extension__ ({ \
27259   int16x4x2_t __s1 = __p1; \
27260   int16x4x2_t __rev1; \
27261   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27262   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27263   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 1); \
27264 })
27265 #endif
27266 
27267 #ifdef __LITTLE_ENDIAN__
27268 #define vst1_p8_x3(__p0, __p1) __extension__ ({ \
27269   poly8x8x3_t __s1 = __p1; \
27270   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
27271 })
27272 #else
27273 #define vst1_p8_x3(__p0, __p1) __extension__ ({ \
27274   poly8x8x3_t __s1 = __p1; \
27275   poly8x8x3_t __rev1; \
27276   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27277   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27278   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27279   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
27280 })
27281 #endif
27282 
27283 #ifdef __LITTLE_ENDIAN__
27284 #define vst1_p16_x3(__p0, __p1) __extension__ ({ \
27285   poly16x4x3_t __s1 = __p1; \
27286   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
27287 })
27288 #else
27289 #define vst1_p16_x3(__p0, __p1) __extension__ ({ \
27290   poly16x4x3_t __s1 = __p1; \
27291   poly16x4x3_t __rev1; \
27292   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27293   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27294   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27295   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
27296 })
27297 #endif
27298 
27299 #ifdef __LITTLE_ENDIAN__
27300 #define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
27301   poly8x16x3_t __s1 = __p1; \
27302   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
27303 })
27304 #else
27305 #define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
27306   poly8x16x3_t __s1 = __p1; \
27307   poly8x16x3_t __rev1; \
27308   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27309   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27310   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27311   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
27312 })
27313 #endif
27314 
27315 #ifdef __LITTLE_ENDIAN__
27316 #define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
27317   poly16x8x3_t __s1 = __p1; \
27318   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
27319 })
27320 #else
27321 #define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
27322   poly16x8x3_t __s1 = __p1; \
27323   poly16x8x3_t __rev1; \
27324   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27325   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27326   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27327   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
27328 })
27329 #endif
27330 
27331 #ifdef __LITTLE_ENDIAN__
27332 #define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
27333   uint8x16x3_t __s1 = __p1; \
27334   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
27335 })
27336 #else
27337 #define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
27338   uint8x16x3_t __s1 = __p1; \
27339   uint8x16x3_t __rev1; \
27340   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27341   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27342   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27343   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
27344 })
27345 #endif
27346 
27347 #ifdef __LITTLE_ENDIAN__
27348 #define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
27349   uint32x4x3_t __s1 = __p1; \
27350   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
27351 })
27352 #else
27353 #define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
27354   uint32x4x3_t __s1 = __p1; \
27355   uint32x4x3_t __rev1; \
27356   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27357   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27358   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27359   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
27360 })
27361 #endif
27362 
27363 #ifdef __LITTLE_ENDIAN__
27364 #define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
27365   uint64x2x3_t __s1 = __p1; \
27366   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
27367 })
27368 #else
27369 #define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
27370   uint64x2x3_t __s1 = __p1; \
27371   uint64x2x3_t __rev1; \
27372   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27373   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27374   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27375   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
27376 })
27377 #endif
27378 
27379 #ifdef __LITTLE_ENDIAN__
27380 #define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
27381   uint16x8x3_t __s1 = __p1; \
27382   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
27383 })
27384 #else
27385 #define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
27386   uint16x8x3_t __s1 = __p1; \
27387   uint16x8x3_t __rev1; \
27388   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27389   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27390   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27391   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
27392 })
27393 #endif
27394 
27395 #ifdef __LITTLE_ENDIAN__
27396 #define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
27397   int8x16x3_t __s1 = __p1; \
27398   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
27399 })
27400 #else
27401 #define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
27402   int8x16x3_t __s1 = __p1; \
27403   int8x16x3_t __rev1; \
27404   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27405   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27406   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27407   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
27408 })
27409 #endif
27410 
27411 #ifdef __LITTLE_ENDIAN__
27412 #define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
27413   float32x4x3_t __s1 = __p1; \
27414   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 41); \
27415 })
27416 #else
27417 #define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
27418   float32x4x3_t __s1 = __p1; \
27419   float32x4x3_t __rev1; \
27420   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27421   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27422   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27423   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 41); \
27424 })
27425 #endif
27426 
27427 #ifdef __LITTLE_ENDIAN__
27428 #define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
27429   int32x4x3_t __s1 = __p1; \
27430   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 34); \
27431 })
27432 #else
27433 #define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
27434   int32x4x3_t __s1 = __p1; \
27435   int32x4x3_t __rev1; \
27436   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27437   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27438   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27439   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 34); \
27440 })
27441 #endif
27442 
27443 #ifdef __LITTLE_ENDIAN__
27444 #define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
27445   int64x2x3_t __s1 = __p1; \
27446   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 35); \
27447 })
27448 #else
27449 #define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
27450   int64x2x3_t __s1 = __p1; \
27451   int64x2x3_t __rev1; \
27452   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27453   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27454   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27455   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 35); \
27456 })
27457 #endif
27458 
27459 #ifdef __LITTLE_ENDIAN__
27460 #define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
27461   int16x8x3_t __s1 = __p1; \
27462   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 33); \
27463 })
27464 #else
27465 #define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
27466   int16x8x3_t __s1 = __p1; \
27467   int16x8x3_t __rev1; \
27468   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27469   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27470   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27471   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 33); \
27472 })
27473 #endif
27474 
27475 #ifdef __LITTLE_ENDIAN__
27476 #define vst1_u8_x3(__p0, __p1) __extension__ ({ \
27477   uint8x8x3_t __s1 = __p1; \
27478   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
27479 })
27480 #else
27481 #define vst1_u8_x3(__p0, __p1) __extension__ ({ \
27482   uint8x8x3_t __s1 = __p1; \
27483   uint8x8x3_t __rev1; \
27484   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27485   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27486   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27487   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
27488 })
27489 #endif
27490 
27491 #ifdef __LITTLE_ENDIAN__
27492 #define vst1_u32_x3(__p0, __p1) __extension__ ({ \
27493   uint32x2x3_t __s1 = __p1; \
27494   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
27495 })
27496 #else
27497 #define vst1_u32_x3(__p0, __p1) __extension__ ({ \
27498   uint32x2x3_t __s1 = __p1; \
27499   uint32x2x3_t __rev1; \
27500   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27501   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27502   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27503   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
27504 })
27505 #endif
27506 
27507 #define vst1_u64_x3(__p0, __p1) __extension__ ({ \
27508   uint64x1x3_t __s1 = __p1; \
27509   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
27510 })
27511 #ifdef __LITTLE_ENDIAN__
27512 #define vst1_u16_x3(__p0, __p1) __extension__ ({ \
27513   uint16x4x3_t __s1 = __p1; \
27514   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
27515 })
27516 #else
27517 #define vst1_u16_x3(__p0, __p1) __extension__ ({ \
27518   uint16x4x3_t __s1 = __p1; \
27519   uint16x4x3_t __rev1; \
27520   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27521   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27522   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27523   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
27524 })
27525 #endif
27526 
27527 #ifdef __LITTLE_ENDIAN__
27528 #define vst1_s8_x3(__p0, __p1) __extension__ ({ \
27529   int8x8x3_t __s1 = __p1; \
27530   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
27531 })
27532 #else
27533 #define vst1_s8_x3(__p0, __p1) __extension__ ({ \
27534   int8x8x3_t __s1 = __p1; \
27535   int8x8x3_t __rev1; \
27536   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27537   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27538   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27539   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
27540 })
27541 #endif
27542 
27543 #ifdef __LITTLE_ENDIAN__
27544 #define vst1_f32_x3(__p0, __p1) __extension__ ({ \
27545   float32x2x3_t __s1 = __p1; \
27546   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 9); \
27547 })
27548 #else
27549 #define vst1_f32_x3(__p0, __p1) __extension__ ({ \
27550   float32x2x3_t __s1 = __p1; \
27551   float32x2x3_t __rev1; \
27552   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27553   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27554   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27555   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 9); \
27556 })
27557 #endif
27558 
27559 #ifdef __LITTLE_ENDIAN__
27560 #define vst1_s32_x3(__p0, __p1) __extension__ ({ \
27561   int32x2x3_t __s1 = __p1; \
27562   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 2); \
27563 })
27564 #else
27565 #define vst1_s32_x3(__p0, __p1) __extension__ ({ \
27566   int32x2x3_t __s1 = __p1; \
27567   int32x2x3_t __rev1; \
27568   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27569   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27570   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27571   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 2); \
27572 })
27573 #endif
27574 
27575 #define vst1_s64_x3(__p0, __p1) __extension__ ({ \
27576   int64x1x3_t __s1 = __p1; \
27577   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 3); \
27578 })
27579 #ifdef __LITTLE_ENDIAN__
27580 #define vst1_s16_x3(__p0, __p1) __extension__ ({ \
27581   int16x4x3_t __s1 = __p1; \
27582   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 1); \
27583 })
27584 #else
27585 #define vst1_s16_x3(__p0, __p1) __extension__ ({ \
27586   int16x4x3_t __s1 = __p1; \
27587   int16x4x3_t __rev1; \
27588   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27589   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27590   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27591   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 1); \
27592 })
27593 #endif
27594 
27595 #ifdef __LITTLE_ENDIAN__
27596 #define vst1_p8_x4(__p0, __p1) __extension__ ({ \
27597   poly8x8x4_t __s1 = __p1; \
27598   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
27599 })
27600 #else
27601 #define vst1_p8_x4(__p0, __p1) __extension__ ({ \
27602   poly8x8x4_t __s1 = __p1; \
27603   poly8x8x4_t __rev1; \
27604   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27605   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27606   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27607   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27608   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
27609 })
27610 #endif
27611 
27612 #ifdef __LITTLE_ENDIAN__
27613 #define vst1_p16_x4(__p0, __p1) __extension__ ({ \
27614   poly16x4x4_t __s1 = __p1; \
27615   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
27616 })
27617 #else
27618 #define vst1_p16_x4(__p0, __p1) __extension__ ({ \
27619   poly16x4x4_t __s1 = __p1; \
27620   poly16x4x4_t __rev1; \
27621   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27622   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27623   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27624   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27625   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
27626 })
27627 #endif
27628 
27629 #ifdef __LITTLE_ENDIAN__
27630 #define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
27631   poly8x16x4_t __s1 = __p1; \
27632   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
27633 })
27634 #else
27635 #define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
27636   poly8x16x4_t __s1 = __p1; \
27637   poly8x16x4_t __rev1; \
27638   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27639   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27640   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27641   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27642   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
27643 })
27644 #endif
27645 
27646 #ifdef __LITTLE_ENDIAN__
27647 #define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
27648   poly16x8x4_t __s1 = __p1; \
27649   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
27650 })
27651 #else
27652 #define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
27653   poly16x8x4_t __s1 = __p1; \
27654   poly16x8x4_t __rev1; \
27655   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27656   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27657   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27658   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27659   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
27660 })
27661 #endif
27662 
27663 #ifdef __LITTLE_ENDIAN__
27664 #define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
27665   uint8x16x4_t __s1 = __p1; \
27666   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
27667 })
27668 #else
27669 #define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
27670   uint8x16x4_t __s1 = __p1; \
27671   uint8x16x4_t __rev1; \
27672   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27673   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27674   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27675   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27676   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
27677 })
27678 #endif
27679 
27680 #ifdef __LITTLE_ENDIAN__
27681 #define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
27682   uint32x4x4_t __s1 = __p1; \
27683   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
27684 })
27685 #else
27686 #define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
27687   uint32x4x4_t __s1 = __p1; \
27688   uint32x4x4_t __rev1; \
27689   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27690   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27691   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27692   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27693   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
27694 })
27695 #endif
27696 
27697 #ifdef __LITTLE_ENDIAN__
27698 #define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
27699   uint64x2x4_t __s1 = __p1; \
27700   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
27701 })
27702 #else
27703 #define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
27704   uint64x2x4_t __s1 = __p1; \
27705   uint64x2x4_t __rev1; \
27706   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27707   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27708   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27709   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27710   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
27711 })
27712 #endif
27713 
27714 #ifdef __LITTLE_ENDIAN__
27715 #define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
27716   uint16x8x4_t __s1 = __p1; \
27717   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
27718 })
27719 #else
27720 #define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
27721   uint16x8x4_t __s1 = __p1; \
27722   uint16x8x4_t __rev1; \
27723   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27724   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27725   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27726   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27727   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
27728 })
27729 #endif
27730 
27731 #ifdef __LITTLE_ENDIAN__
27732 #define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
27733   int8x16x4_t __s1 = __p1; \
27734   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
27735 })
27736 #else
27737 #define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
27738   int8x16x4_t __s1 = __p1; \
27739   int8x16x4_t __rev1; \
27740   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27741   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27742   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27743   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27744   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
27745 })
27746 #endif
27747 
27748 #ifdef __LITTLE_ENDIAN__
27749 #define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
27750   float32x4x4_t __s1 = __p1; \
27751   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 41); \
27752 })
27753 #else
27754 #define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
27755   float32x4x4_t __s1 = __p1; \
27756   float32x4x4_t __rev1; \
27757   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27758   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27759   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27760   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27761   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 41); \
27762 })
27763 #endif
27764 
27765 #ifdef __LITTLE_ENDIAN__
27766 #define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
27767   int32x4x4_t __s1 = __p1; \
27768   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 34); \
27769 })
27770 #else
27771 #define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
27772   int32x4x4_t __s1 = __p1; \
27773   int32x4x4_t __rev1; \
27774   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27775   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27776   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27777   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27778   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 34); \
27779 })
27780 #endif
27781 
27782 #ifdef __LITTLE_ENDIAN__
27783 #define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
27784   int64x2x4_t __s1 = __p1; \
27785   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 35); \
27786 })
27787 #else
27788 #define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
27789   int64x2x4_t __s1 = __p1; \
27790   int64x2x4_t __rev1; \
27791   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27792   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27793   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27794   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27795   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 35); \
27796 })
27797 #endif
27798 
27799 #ifdef __LITTLE_ENDIAN__
27800 #define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
27801   int16x8x4_t __s1 = __p1; \
27802   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 33); \
27803 })
27804 #else
27805 #define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
27806   int16x8x4_t __s1 = __p1; \
27807   int16x8x4_t __rev1; \
27808   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27809   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27810   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27811   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27812   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 33); \
27813 })
27814 #endif
27815 
27816 #ifdef __LITTLE_ENDIAN__
27817 #define vst1_u8_x4(__p0, __p1) __extension__ ({ \
27818   uint8x8x4_t __s1 = __p1; \
27819   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
27820 })
27821 #else
27822 #define vst1_u8_x4(__p0, __p1) __extension__ ({ \
27823   uint8x8x4_t __s1 = __p1; \
27824   uint8x8x4_t __rev1; \
27825   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27826   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27827   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27828   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27829   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
27830 })
27831 #endif
27832 
27833 #ifdef __LITTLE_ENDIAN__
27834 #define vst1_u32_x4(__p0, __p1) __extension__ ({ \
27835   uint32x2x4_t __s1 = __p1; \
27836   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
27837 })
27838 #else
27839 #define vst1_u32_x4(__p0, __p1) __extension__ ({ \
27840   uint32x2x4_t __s1 = __p1; \
27841   uint32x2x4_t __rev1; \
27842   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27843   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27844   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27845   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27846   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
27847 })
27848 #endif
27849 
27850 #define vst1_u64_x4(__p0, __p1) __extension__ ({ \
27851   uint64x1x4_t __s1 = __p1; \
27852   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
27853 })
27854 #ifdef __LITTLE_ENDIAN__
27855 #define vst1_u16_x4(__p0, __p1) __extension__ ({ \
27856   uint16x4x4_t __s1 = __p1; \
27857   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
27858 })
27859 #else
27860 #define vst1_u16_x4(__p0, __p1) __extension__ ({ \
27861   uint16x4x4_t __s1 = __p1; \
27862   uint16x4x4_t __rev1; \
27863   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27864   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27865   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27866   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27867   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
27868 })
27869 #endif
27870 
27871 #ifdef __LITTLE_ENDIAN__
27872 #define vst1_s8_x4(__p0, __p1) __extension__ ({ \
27873   int8x8x4_t __s1 = __p1; \
27874   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
27875 })
27876 #else
27877 #define vst1_s8_x4(__p0, __p1) __extension__ ({ \
27878   int8x8x4_t __s1 = __p1; \
27879   int8x8x4_t __rev1; \
27880   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27881   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27882   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27883   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27884   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
27885 })
27886 #endif
27887 
27888 #ifdef __LITTLE_ENDIAN__
27889 #define vst1_f32_x4(__p0, __p1) __extension__ ({ \
27890   float32x2x4_t __s1 = __p1; \
27891   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 9); \
27892 })
27893 #else
27894 #define vst1_f32_x4(__p0, __p1) __extension__ ({ \
27895   float32x2x4_t __s1 = __p1; \
27896   float32x2x4_t __rev1; \
27897   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27898   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27899   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27900   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27901   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 9); \
27902 })
27903 #endif
27904 
27905 #ifdef __LITTLE_ENDIAN__
27906 #define vst1_s32_x4(__p0, __p1) __extension__ ({ \
27907   int32x2x4_t __s1 = __p1; \
27908   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 2); \
27909 })
27910 #else
27911 #define vst1_s32_x4(__p0, __p1) __extension__ ({ \
27912   int32x2x4_t __s1 = __p1; \
27913   int32x2x4_t __rev1; \
27914   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27915   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27916   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27917   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27918   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 2); \
27919 })
27920 #endif
27921 
27922 #define vst1_s64_x4(__p0, __p1) __extension__ ({ \
27923   int64x1x4_t __s1 = __p1; \
27924   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 3); \
27925 })
27926 #ifdef __LITTLE_ENDIAN__
27927 #define vst1_s16_x4(__p0, __p1) __extension__ ({ \
27928   int16x4x4_t __s1 = __p1; \
27929   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 1); \
27930 })
27931 #else
27932 #define vst1_s16_x4(__p0, __p1) __extension__ ({ \
27933   int16x4x4_t __s1 = __p1; \
27934   int16x4x4_t __rev1; \
27935   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27936   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27937   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27938   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27939   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 1); \
27940 })
27941 #endif
27942 
27943 #ifdef __LITTLE_ENDIAN__
27944 #define vst2_p8(__p0, __p1) __extension__ ({ \
27945   poly8x8x2_t __s1 = __p1; \
27946   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
27947 })
27948 #else
27949 #define vst2_p8(__p0, __p1) __extension__ ({ \
27950   poly8x8x2_t __s1 = __p1; \
27951   poly8x8x2_t __rev1; \
27952   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27953   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27954   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
27955 })
27956 #endif
27957 
27958 #ifdef __LITTLE_ENDIAN__
27959 #define vst2_p16(__p0, __p1) __extension__ ({ \
27960   poly16x4x2_t __s1 = __p1; \
27961   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
27962 })
27963 #else
27964 #define vst2_p16(__p0, __p1) __extension__ ({ \
27965   poly16x4x2_t __s1 = __p1; \
27966   poly16x4x2_t __rev1; \
27967   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27968   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27969   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
27970 })
27971 #endif
27972 
27973 #ifdef __LITTLE_ENDIAN__
27974 #define vst2q_p8(__p0, __p1) __extension__ ({ \
27975   poly8x16x2_t __s1 = __p1; \
27976   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
27977 })
27978 #else
27979 #define vst2q_p8(__p0, __p1) __extension__ ({ \
27980   poly8x16x2_t __s1 = __p1; \
27981   poly8x16x2_t __rev1; \
27982   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27983   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27984   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
27985 })
27986 #endif
27987 
27988 #ifdef __LITTLE_ENDIAN__
27989 #define vst2q_p16(__p0, __p1) __extension__ ({ \
27990   poly16x8x2_t __s1 = __p1; \
27991   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
27992 })
27993 #else
27994 #define vst2q_p16(__p0, __p1) __extension__ ({ \
27995   poly16x8x2_t __s1 = __p1; \
27996   poly16x8x2_t __rev1; \
27997   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27998   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27999   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
28000 })
28001 #endif
28002 
28003 #ifdef __LITTLE_ENDIAN__
28004 #define vst2q_u8(__p0, __p1) __extension__ ({ \
28005   uint8x16x2_t __s1 = __p1; \
28006   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
28007 })
28008 #else
28009 #define vst2q_u8(__p0, __p1) __extension__ ({ \
28010   uint8x16x2_t __s1 = __p1; \
28011   uint8x16x2_t __rev1; \
28012   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28013   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28014   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
28015 })
28016 #endif
28017 
28018 #ifdef __LITTLE_ENDIAN__
28019 #define vst2q_u32(__p0, __p1) __extension__ ({ \
28020   uint32x4x2_t __s1 = __p1; \
28021   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
28022 })
28023 #else
28024 #define vst2q_u32(__p0, __p1) __extension__ ({ \
28025   uint32x4x2_t __s1 = __p1; \
28026   uint32x4x2_t __rev1; \
28027   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28028   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28029   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
28030 })
28031 #endif
28032 
28033 #ifdef __LITTLE_ENDIAN__
28034 #define vst2q_u16(__p0, __p1) __extension__ ({ \
28035   uint16x8x2_t __s1 = __p1; \
28036   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
28037 })
28038 #else
28039 #define vst2q_u16(__p0, __p1) __extension__ ({ \
28040   uint16x8x2_t __s1 = __p1; \
28041   uint16x8x2_t __rev1; \
28042   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28043   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28044   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
28045 })
28046 #endif
28047 
28048 #ifdef __LITTLE_ENDIAN__
28049 #define vst2q_s8(__p0, __p1) __extension__ ({ \
28050   int8x16x2_t __s1 = __p1; \
28051   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
28052 })
28053 #else
28054 #define vst2q_s8(__p0, __p1) __extension__ ({ \
28055   int8x16x2_t __s1 = __p1; \
28056   int8x16x2_t __rev1; \
28057   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28058   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28059   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
28060 })
28061 #endif
28062 
28063 #ifdef __LITTLE_ENDIAN__
28064 #define vst2q_f32(__p0, __p1) __extension__ ({ \
28065   float32x4x2_t __s1 = __p1; \
28066   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 41); \
28067 })
28068 #else
28069 #define vst2q_f32(__p0, __p1) __extension__ ({ \
28070   float32x4x2_t __s1 = __p1; \
28071   float32x4x2_t __rev1; \
28072   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28073   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28074   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 41); \
28075 })
28076 #endif
28077 
28078 #ifdef __LITTLE_ENDIAN__
28079 #define vst2q_s32(__p0, __p1) __extension__ ({ \
28080   int32x4x2_t __s1 = __p1; \
28081   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 34); \
28082 })
28083 #else
28084 #define vst2q_s32(__p0, __p1) __extension__ ({ \
28085   int32x4x2_t __s1 = __p1; \
28086   int32x4x2_t __rev1; \
28087   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28088   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28089   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 34); \
28090 })
28091 #endif
28092 
28093 #ifdef __LITTLE_ENDIAN__
28094 #define vst2q_s16(__p0, __p1) __extension__ ({ \
28095   int16x8x2_t __s1 = __p1; \
28096   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 33); \
28097 })
28098 #else
28099 #define vst2q_s16(__p0, __p1) __extension__ ({ \
28100   int16x8x2_t __s1 = __p1; \
28101   int16x8x2_t __rev1; \
28102   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28103   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28104   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 33); \
28105 })
28106 #endif
28107 
28108 #ifdef __LITTLE_ENDIAN__
28109 #define vst2_u8(__p0, __p1) __extension__ ({ \
28110   uint8x8x2_t __s1 = __p1; \
28111   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
28112 })
28113 #else
28114 #define vst2_u8(__p0, __p1) __extension__ ({ \
28115   uint8x8x2_t __s1 = __p1; \
28116   uint8x8x2_t __rev1; \
28117   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28118   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28119   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
28120 })
28121 #endif
28122 
28123 #ifdef __LITTLE_ENDIAN__
28124 #define vst2_u32(__p0, __p1) __extension__ ({ \
28125   uint32x2x2_t __s1 = __p1; \
28126   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
28127 })
28128 #else
28129 #define vst2_u32(__p0, __p1) __extension__ ({ \
28130   uint32x2x2_t __s1 = __p1; \
28131   uint32x2x2_t __rev1; \
28132   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28133   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28134   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
28135 })
28136 #endif
28137 
28138 #define vst2_u64(__p0, __p1) __extension__ ({ \
28139   uint64x1x2_t __s1 = __p1; \
28140   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
28141 })
28142 #ifdef __LITTLE_ENDIAN__
28143 #define vst2_u16(__p0, __p1) __extension__ ({ \
28144   uint16x4x2_t __s1 = __p1; \
28145   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
28146 })
28147 #else
28148 #define vst2_u16(__p0, __p1) __extension__ ({ \
28149   uint16x4x2_t __s1 = __p1; \
28150   uint16x4x2_t __rev1; \
28151   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28152   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28153   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
28154 })
28155 #endif
28156 
28157 #ifdef __LITTLE_ENDIAN__
28158 #define vst2_s8(__p0, __p1) __extension__ ({ \
28159   int8x8x2_t __s1 = __p1; \
28160   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
28161 })
28162 #else
28163 #define vst2_s8(__p0, __p1) __extension__ ({ \
28164   int8x8x2_t __s1 = __p1; \
28165   int8x8x2_t __rev1; \
28166   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28167   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28168   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
28169 })
28170 #endif
28171 
28172 #ifdef __LITTLE_ENDIAN__
28173 #define vst2_f32(__p0, __p1) __extension__ ({ \
28174   float32x2x2_t __s1 = __p1; \
28175   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 9); \
28176 })
28177 #else
28178 #define vst2_f32(__p0, __p1) __extension__ ({ \
28179   float32x2x2_t __s1 = __p1; \
28180   float32x2x2_t __rev1; \
28181   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28182   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28183   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 9); \
28184 })
28185 #endif
28186 
28187 #ifdef __LITTLE_ENDIAN__
28188 #define vst2_s32(__p0, __p1) __extension__ ({ \
28189   int32x2x2_t __s1 = __p1; \
28190   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 2); \
28191 })
28192 #else
28193 #define vst2_s32(__p0, __p1) __extension__ ({ \
28194   int32x2x2_t __s1 = __p1; \
28195   int32x2x2_t __rev1; \
28196   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28197   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28198   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 2); \
28199 })
28200 #endif
28201 
28202 #define vst2_s64(__p0, __p1) __extension__ ({ \
28203   int64x1x2_t __s1 = __p1; \
28204   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 3); \
28205 })
28206 #ifdef __LITTLE_ENDIAN__
28207 #define vst2_s16(__p0, __p1) __extension__ ({ \
28208   int16x4x2_t __s1 = __p1; \
28209   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 1); \
28210 })
28211 #else
28212 #define vst2_s16(__p0, __p1) __extension__ ({ \
28213   int16x4x2_t __s1 = __p1; \
28214   int16x4x2_t __rev1; \
28215   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28216   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28217   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 1); \
28218 })
28219 #endif
28220 
28221 #ifdef __LITTLE_ENDIAN__
28222 #define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
28223   poly8x8x2_t __s1 = __p1; \
28224   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
28225 })
28226 #else
28227 #define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
28228   poly8x8x2_t __s1 = __p1; \
28229   poly8x8x2_t __rev1; \
28230   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28231   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28232   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
28233 })
28234 #endif
28235 
28236 #ifdef __LITTLE_ENDIAN__
28237 #define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28238   poly16x4x2_t __s1 = __p1; \
28239   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
28240 })
28241 #else
28242 #define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28243   poly16x4x2_t __s1 = __p1; \
28244   poly16x4x2_t __rev1; \
28245   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28246   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28247   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
28248 })
28249 #endif
28250 
28251 #ifdef __LITTLE_ENDIAN__
28252 #define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28253   poly16x8x2_t __s1 = __p1; \
28254   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
28255 })
28256 #else
28257 #define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28258   poly16x8x2_t __s1 = __p1; \
28259   poly16x8x2_t __rev1; \
28260   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28261   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28262   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
28263 })
28264 #endif
28265 
28266 #ifdef __LITTLE_ENDIAN__
28267 #define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28268   uint32x4x2_t __s1 = __p1; \
28269   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
28270 })
28271 #else
28272 #define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28273   uint32x4x2_t __s1 = __p1; \
28274   uint32x4x2_t __rev1; \
28275   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28276   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28277   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
28278 })
28279 #endif
28280 
28281 #ifdef __LITTLE_ENDIAN__
28282 #define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28283   uint16x8x2_t __s1 = __p1; \
28284   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
28285 })
28286 #else
28287 #define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28288   uint16x8x2_t __s1 = __p1; \
28289   uint16x8x2_t __rev1; \
28290   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28291   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28292   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
28293 })
28294 #endif
28295 
28296 #ifdef __LITTLE_ENDIAN__
28297 #define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28298   float32x4x2_t __s1 = __p1; \
28299   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 41); \
28300 })
28301 #else
28302 #define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28303   float32x4x2_t __s1 = __p1; \
28304   float32x4x2_t __rev1; \
28305   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28306   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28307   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 41); \
28308 })
28309 #endif
28310 
28311 #ifdef __LITTLE_ENDIAN__
28312 #define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28313   int32x4x2_t __s1 = __p1; \
28314   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 34); \
28315 })
28316 #else
28317 #define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28318   int32x4x2_t __s1 = __p1; \
28319   int32x4x2_t __rev1; \
28320   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28321   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28322   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 34); \
28323 })
28324 #endif
28325 
28326 #ifdef __LITTLE_ENDIAN__
28327 #define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28328   int16x8x2_t __s1 = __p1; \
28329   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 33); \
28330 })
28331 #else
28332 #define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28333   int16x8x2_t __s1 = __p1; \
28334   int16x8x2_t __rev1; \
28335   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28336   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28337   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 33); \
28338 })
28339 #endif
28340 
28341 #ifdef __LITTLE_ENDIAN__
28342 #define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
28343   uint8x8x2_t __s1 = __p1; \
28344   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
28345 })
28346 #else
28347 #define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
28348   uint8x8x2_t __s1 = __p1; \
28349   uint8x8x2_t __rev1; \
28350   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28351   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28352   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
28353 })
28354 #endif
28355 
28356 #ifdef __LITTLE_ENDIAN__
28357 #define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28358   uint32x2x2_t __s1 = __p1; \
28359   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
28360 })
28361 #else
28362 #define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28363   uint32x2x2_t __s1 = __p1; \
28364   uint32x2x2_t __rev1; \
28365   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28366   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28367   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
28368 })
28369 #endif
28370 
28371 #ifdef __LITTLE_ENDIAN__
28372 #define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28373   uint16x4x2_t __s1 = __p1; \
28374   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
28375 })
28376 #else
28377 #define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28378   uint16x4x2_t __s1 = __p1; \
28379   uint16x4x2_t __rev1; \
28380   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28381   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28382   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
28383 })
28384 #endif
28385 
28386 #ifdef __LITTLE_ENDIAN__
28387 #define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
28388   int8x8x2_t __s1 = __p1; \
28389   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
28390 })
28391 #else
28392 #define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
28393   int8x8x2_t __s1 = __p1; \
28394   int8x8x2_t __rev1; \
28395   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28396   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28397   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
28398 })
28399 #endif
28400 
28401 #ifdef __LITTLE_ENDIAN__
28402 #define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28403   float32x2x2_t __s1 = __p1; \
28404   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 9); \
28405 })
28406 #else
28407 #define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28408   float32x2x2_t __s1 = __p1; \
28409   float32x2x2_t __rev1; \
28410   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28411   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28412   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 9); \
28413 })
28414 #endif
28415 
28416 #ifdef __LITTLE_ENDIAN__
28417 #define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28418   int32x2x2_t __s1 = __p1; \
28419   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 2); \
28420 })
28421 #else
28422 #define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28423   int32x2x2_t __s1 = __p1; \
28424   int32x2x2_t __rev1; \
28425   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28426   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28427   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 2); \
28428 })
28429 #endif
28430 
28431 #ifdef __LITTLE_ENDIAN__
28432 #define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28433   int16x4x2_t __s1 = __p1; \
28434   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 1); \
28435 })
28436 #else
28437 #define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28438   int16x4x2_t __s1 = __p1; \
28439   int16x4x2_t __rev1; \
28440   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28441   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28442   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 1); \
28443 })
28444 #endif
28445 
28446 #ifdef __LITTLE_ENDIAN__
28447 #define vst3_p8(__p0, __p1) __extension__ ({ \
28448   poly8x8x3_t __s1 = __p1; \
28449   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
28450 })
28451 #else
28452 #define vst3_p8(__p0, __p1) __extension__ ({ \
28453   poly8x8x3_t __s1 = __p1; \
28454   poly8x8x3_t __rev1; \
28455   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28456   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28457   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28458   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
28459 })
28460 #endif
28461 
28462 #ifdef __LITTLE_ENDIAN__
28463 #define vst3_p16(__p0, __p1) __extension__ ({ \
28464   poly16x4x3_t __s1 = __p1; \
28465   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
28466 })
28467 #else
28468 #define vst3_p16(__p0, __p1) __extension__ ({ \
28469   poly16x4x3_t __s1 = __p1; \
28470   poly16x4x3_t __rev1; \
28471   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28472   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28473   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28474   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
28475 })
28476 #endif
28477 
28478 #ifdef __LITTLE_ENDIAN__
28479 #define vst3q_p8(__p0, __p1) __extension__ ({ \
28480   poly8x16x3_t __s1 = __p1; \
28481   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
28482 })
28483 #else
28484 #define vst3q_p8(__p0, __p1) __extension__ ({ \
28485   poly8x16x3_t __s1 = __p1; \
28486   poly8x16x3_t __rev1; \
28487   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28488   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28489   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28490   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
28491 })
28492 #endif
28493 
28494 #ifdef __LITTLE_ENDIAN__
28495 #define vst3q_p16(__p0, __p1) __extension__ ({ \
28496   poly16x8x3_t __s1 = __p1; \
28497   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
28498 })
28499 #else
28500 #define vst3q_p16(__p0, __p1) __extension__ ({ \
28501   poly16x8x3_t __s1 = __p1; \
28502   poly16x8x3_t __rev1; \
28503   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28504   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28505   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28506   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
28507 })
28508 #endif
28509 
28510 #ifdef __LITTLE_ENDIAN__
28511 #define vst3q_u8(__p0, __p1) __extension__ ({ \
28512   uint8x16x3_t __s1 = __p1; \
28513   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
28514 })
28515 #else
28516 #define vst3q_u8(__p0, __p1) __extension__ ({ \
28517   uint8x16x3_t __s1 = __p1; \
28518   uint8x16x3_t __rev1; \
28519   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28520   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28521   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28522   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
28523 })
28524 #endif
28525 
28526 #ifdef __LITTLE_ENDIAN__
28527 #define vst3q_u32(__p0, __p1) __extension__ ({ \
28528   uint32x4x3_t __s1 = __p1; \
28529   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
28530 })
28531 #else
28532 #define vst3q_u32(__p0, __p1) __extension__ ({ \
28533   uint32x4x3_t __s1 = __p1; \
28534   uint32x4x3_t __rev1; \
28535   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28536   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28537   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28538   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
28539 })
28540 #endif
28541 
28542 #ifdef __LITTLE_ENDIAN__
28543 #define vst3q_u16(__p0, __p1) __extension__ ({ \
28544   uint16x8x3_t __s1 = __p1; \
28545   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
28546 })
28547 #else
28548 #define vst3q_u16(__p0, __p1) __extension__ ({ \
28549   uint16x8x3_t __s1 = __p1; \
28550   uint16x8x3_t __rev1; \
28551   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28552   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28553   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28554   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
28555 })
28556 #endif
28557 
28558 #ifdef __LITTLE_ENDIAN__
28559 #define vst3q_s8(__p0, __p1) __extension__ ({ \
28560   int8x16x3_t __s1 = __p1; \
28561   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
28562 })
28563 #else
28564 #define vst3q_s8(__p0, __p1) __extension__ ({ \
28565   int8x16x3_t __s1 = __p1; \
28566   int8x16x3_t __rev1; \
28567   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28568   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28569   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28570   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
28571 })
28572 #endif
28573 
28574 #ifdef __LITTLE_ENDIAN__
28575 #define vst3q_f32(__p0, __p1) __extension__ ({ \
28576   float32x4x3_t __s1 = __p1; \
28577   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 41); \
28578 })
28579 #else
28580 #define vst3q_f32(__p0, __p1) __extension__ ({ \
28581   float32x4x3_t __s1 = __p1; \
28582   float32x4x3_t __rev1; \
28583   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28584   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28585   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28586   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 41); \
28587 })
28588 #endif
28589 
28590 #ifdef __LITTLE_ENDIAN__
28591 #define vst3q_s32(__p0, __p1) __extension__ ({ \
28592   int32x4x3_t __s1 = __p1; \
28593   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 34); \
28594 })
28595 #else
28596 #define vst3q_s32(__p0, __p1) __extension__ ({ \
28597   int32x4x3_t __s1 = __p1; \
28598   int32x4x3_t __rev1; \
28599   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28600   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28601   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28602   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 34); \
28603 })
28604 #endif
28605 
28606 #ifdef __LITTLE_ENDIAN__
28607 #define vst3q_s16(__p0, __p1) __extension__ ({ \
28608   int16x8x3_t __s1 = __p1; \
28609   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 33); \
28610 })
28611 #else
28612 #define vst3q_s16(__p0, __p1) __extension__ ({ \
28613   int16x8x3_t __s1 = __p1; \
28614   int16x8x3_t __rev1; \
28615   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28616   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28617   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28618   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 33); \
28619 })
28620 #endif
28621 
28622 #ifdef __LITTLE_ENDIAN__
28623 #define vst3_u8(__p0, __p1) __extension__ ({ \
28624   uint8x8x3_t __s1 = __p1; \
28625   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
28626 })
28627 #else
28628 #define vst3_u8(__p0, __p1) __extension__ ({ \
28629   uint8x8x3_t __s1 = __p1; \
28630   uint8x8x3_t __rev1; \
28631   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28632   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28633   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28634   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
28635 })
28636 #endif
28637 
28638 #ifdef __LITTLE_ENDIAN__
28639 #define vst3_u32(__p0, __p1) __extension__ ({ \
28640   uint32x2x3_t __s1 = __p1; \
28641   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
28642 })
28643 #else
28644 #define vst3_u32(__p0, __p1) __extension__ ({ \
28645   uint32x2x3_t __s1 = __p1; \
28646   uint32x2x3_t __rev1; \
28647   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28648   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28649   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28650   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
28651 })
28652 #endif
28653 
28654 #define vst3_u64(__p0, __p1) __extension__ ({ \
28655   uint64x1x3_t __s1 = __p1; \
28656   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
28657 })
28658 #ifdef __LITTLE_ENDIAN__
28659 #define vst3_u16(__p0, __p1) __extension__ ({ \
28660   uint16x4x3_t __s1 = __p1; \
28661   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
28662 })
28663 #else
28664 #define vst3_u16(__p0, __p1) __extension__ ({ \
28665   uint16x4x3_t __s1 = __p1; \
28666   uint16x4x3_t __rev1; \
28667   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28668   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28669   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28670   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
28671 })
28672 #endif
28673 
28674 #ifdef __LITTLE_ENDIAN__
28675 #define vst3_s8(__p0, __p1) __extension__ ({ \
28676   int8x8x3_t __s1 = __p1; \
28677   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
28678 })
28679 #else
28680 #define vst3_s8(__p0, __p1) __extension__ ({ \
28681   int8x8x3_t __s1 = __p1; \
28682   int8x8x3_t __rev1; \
28683   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28684   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28685   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28686   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
28687 })
28688 #endif
28689 
28690 #ifdef __LITTLE_ENDIAN__
28691 #define vst3_f32(__p0, __p1) __extension__ ({ \
28692   float32x2x3_t __s1 = __p1; \
28693   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 9); \
28694 })
28695 #else
28696 #define vst3_f32(__p0, __p1) __extension__ ({ \
28697   float32x2x3_t __s1 = __p1; \
28698   float32x2x3_t __rev1; \
28699   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28700   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28701   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28702   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 9); \
28703 })
28704 #endif
28705 
28706 #ifdef __LITTLE_ENDIAN__
28707 #define vst3_s32(__p0, __p1) __extension__ ({ \
28708   int32x2x3_t __s1 = __p1; \
28709   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 2); \
28710 })
28711 #else
28712 #define vst3_s32(__p0, __p1) __extension__ ({ \
28713   int32x2x3_t __s1 = __p1; \
28714   int32x2x3_t __rev1; \
28715   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28716   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28717   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28718   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 2); \
28719 })
28720 #endif
28721 
28722 #define vst3_s64(__p0, __p1) __extension__ ({ \
28723   int64x1x3_t __s1 = __p1; \
28724   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 3); \
28725 })
28726 #ifdef __LITTLE_ENDIAN__
28727 #define vst3_s16(__p0, __p1) __extension__ ({ \
28728   int16x4x3_t __s1 = __p1; \
28729   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 1); \
28730 })
28731 #else
28732 #define vst3_s16(__p0, __p1) __extension__ ({ \
28733   int16x4x3_t __s1 = __p1; \
28734   int16x4x3_t __rev1; \
28735   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28736   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28737   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28738   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 1); \
28739 })
28740 #endif
28741 
28742 #ifdef __LITTLE_ENDIAN__
28743 #define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
28744   poly8x8x3_t __s1 = __p1; \
28745   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
28746 })
28747 #else
28748 #define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
28749   poly8x8x3_t __s1 = __p1; \
28750   poly8x8x3_t __rev1; \
28751   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28752   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28753   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28754   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
28755 })
28756 #endif
28757 
28758 #ifdef __LITTLE_ENDIAN__
28759 #define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28760   poly16x4x3_t __s1 = __p1; \
28761   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
28762 })
28763 #else
28764 #define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28765   poly16x4x3_t __s1 = __p1; \
28766   poly16x4x3_t __rev1; \
28767   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28768   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28769   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28770   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
28771 })
28772 #endif
28773 
28774 #ifdef __LITTLE_ENDIAN__
28775 #define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28776   poly16x8x3_t __s1 = __p1; \
28777   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
28778 })
28779 #else
28780 #define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28781   poly16x8x3_t __s1 = __p1; \
28782   poly16x8x3_t __rev1; \
28783   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28784   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28785   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28786   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
28787 })
28788 #endif
28789 
28790 #ifdef __LITTLE_ENDIAN__
28791 #define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28792   uint32x4x3_t __s1 = __p1; \
28793   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
28794 })
28795 #else
28796 #define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28797   uint32x4x3_t __s1 = __p1; \
28798   uint32x4x3_t __rev1; \
28799   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28800   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28801   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28802   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
28803 })
28804 #endif
28805 
28806 #ifdef __LITTLE_ENDIAN__
28807 #define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28808   uint16x8x3_t __s1 = __p1; \
28809   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
28810 })
28811 #else
28812 #define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28813   uint16x8x3_t __s1 = __p1; \
28814   uint16x8x3_t __rev1; \
28815   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28816   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28817   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28818   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
28819 })
28820 #endif
28821 
28822 #ifdef __LITTLE_ENDIAN__
28823 #define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28824   float32x4x3_t __s1 = __p1; \
28825   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 41); \
28826 })
28827 #else
28828 #define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28829   float32x4x3_t __s1 = __p1; \
28830   float32x4x3_t __rev1; \
28831   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28832   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28833   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28834   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 41); \
28835 })
28836 #endif
28837 
28838 #ifdef __LITTLE_ENDIAN__
28839 #define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28840   int32x4x3_t __s1 = __p1; \
28841   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 34); \
28842 })
28843 #else
28844 #define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28845   int32x4x3_t __s1 = __p1; \
28846   int32x4x3_t __rev1; \
28847   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28848   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28849   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28850   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 34); \
28851 })
28852 #endif
28853 
28854 #ifdef __LITTLE_ENDIAN__
28855 #define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28856   int16x8x3_t __s1 = __p1; \
28857   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 33); \
28858 })
28859 #else
28860 #define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28861   int16x8x3_t __s1 = __p1; \
28862   int16x8x3_t __rev1; \
28863   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28864   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28865   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28866   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 33); \
28867 })
28868 #endif
28869 
28870 #ifdef __LITTLE_ENDIAN__
28871 #define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
28872   uint8x8x3_t __s1 = __p1; \
28873   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
28874 })
28875 #else
28876 #define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
28877   uint8x8x3_t __s1 = __p1; \
28878   uint8x8x3_t __rev1; \
28879   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28880   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28881   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28882   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
28883 })
28884 #endif
28885 
28886 #ifdef __LITTLE_ENDIAN__
28887 #define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28888   uint32x2x3_t __s1 = __p1; \
28889   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
28890 })
28891 #else
28892 #define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28893   uint32x2x3_t __s1 = __p1; \
28894   uint32x2x3_t __rev1; \
28895   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28896   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28897   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28898   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
28899 })
28900 #endif
28901 
28902 #ifdef __LITTLE_ENDIAN__
28903 #define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28904   uint16x4x3_t __s1 = __p1; \
28905   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
28906 })
28907 #else
28908 #define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28909   uint16x4x3_t __s1 = __p1; \
28910   uint16x4x3_t __rev1; \
28911   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28912   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28913   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28914   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
28915 })
28916 #endif
28917 
28918 #ifdef __LITTLE_ENDIAN__
28919 #define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
28920   int8x8x3_t __s1 = __p1; \
28921   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
28922 })
28923 #else
28924 #define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
28925   int8x8x3_t __s1 = __p1; \
28926   int8x8x3_t __rev1; \
28927   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28928   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28929   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28930   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
28931 })
28932 #endif
28933 
28934 #ifdef __LITTLE_ENDIAN__
28935 #define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28936   float32x2x3_t __s1 = __p1; \
28937   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 9); \
28938 })
28939 #else
28940 #define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28941   float32x2x3_t __s1 = __p1; \
28942   float32x2x3_t __rev1; \
28943   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28944   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28945   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28946   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 9); \
28947 })
28948 #endif
28949 
28950 #ifdef __LITTLE_ENDIAN__
28951 #define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28952   int32x2x3_t __s1 = __p1; \
28953   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 2); \
28954 })
28955 #else
28956 #define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28957   int32x2x3_t __s1 = __p1; \
28958   int32x2x3_t __rev1; \
28959   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28960   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28961   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28962   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 2); \
28963 })
28964 #endif
28965 
28966 #ifdef __LITTLE_ENDIAN__
28967 #define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28968   int16x4x3_t __s1 = __p1; \
28969   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 1); \
28970 })
28971 #else
28972 #define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28973   int16x4x3_t __s1 = __p1; \
28974   int16x4x3_t __rev1; \
28975   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28976   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28977   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28978   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 1); \
28979 })
28980 #endif
28981 
28982 #ifdef __LITTLE_ENDIAN__
28983 #define vst4_p8(__p0, __p1) __extension__ ({ \
28984   poly8x8x4_t __s1 = __p1; \
28985   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
28986 })
28987 #else
28988 #define vst4_p8(__p0, __p1) __extension__ ({ \
28989   poly8x8x4_t __s1 = __p1; \
28990   poly8x8x4_t __rev1; \
28991   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28992   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28993   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28994   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
28995   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
28996 })
28997 #endif
28998 
28999 #ifdef __LITTLE_ENDIAN__
29000 #define vst4_p16(__p0, __p1) __extension__ ({ \
29001   poly16x4x4_t __s1 = __p1; \
29002   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
29003 })
29004 #else
29005 #define vst4_p16(__p0, __p1) __extension__ ({ \
29006   poly16x4x4_t __s1 = __p1; \
29007   poly16x4x4_t __rev1; \
29008   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29009   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29010   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29011   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29012   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
29013 })
29014 #endif
29015 
29016 #ifdef __LITTLE_ENDIAN__
29017 #define vst4q_p8(__p0, __p1) __extension__ ({ \
29018   poly8x16x4_t __s1 = __p1; \
29019   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
29020 })
29021 #else
29022 #define vst4q_p8(__p0, __p1) __extension__ ({ \
29023   poly8x16x4_t __s1 = __p1; \
29024   poly8x16x4_t __rev1; \
29025   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29026   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29027   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29028   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29029   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
29030 })
29031 #endif
29032 
29033 #ifdef __LITTLE_ENDIAN__
29034 #define vst4q_p16(__p0, __p1) __extension__ ({ \
29035   poly16x8x4_t __s1 = __p1; \
29036   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
29037 })
29038 #else
29039 #define vst4q_p16(__p0, __p1) __extension__ ({ \
29040   poly16x8x4_t __s1 = __p1; \
29041   poly16x8x4_t __rev1; \
29042   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29043   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29044   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29045   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29046   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
29047 })
29048 #endif
29049 
29050 #ifdef __LITTLE_ENDIAN__
29051 #define vst4q_u8(__p0, __p1) __extension__ ({ \
29052   uint8x16x4_t __s1 = __p1; \
29053   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
29054 })
29055 #else
29056 #define vst4q_u8(__p0, __p1) __extension__ ({ \
29057   uint8x16x4_t __s1 = __p1; \
29058   uint8x16x4_t __rev1; \
29059   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29060   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29061   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29062   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29063   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
29064 })
29065 #endif
29066 
29067 #ifdef __LITTLE_ENDIAN__
29068 #define vst4q_u32(__p0, __p1) __extension__ ({ \
29069   uint32x4x4_t __s1 = __p1; \
29070   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
29071 })
29072 #else
29073 #define vst4q_u32(__p0, __p1) __extension__ ({ \
29074   uint32x4x4_t __s1 = __p1; \
29075   uint32x4x4_t __rev1; \
29076   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29077   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29078   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29079   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29080   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
29081 })
29082 #endif
29083 
29084 #ifdef __LITTLE_ENDIAN__
29085 #define vst4q_u16(__p0, __p1) __extension__ ({ \
29086   uint16x8x4_t __s1 = __p1; \
29087   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
29088 })
29089 #else
29090 #define vst4q_u16(__p0, __p1) __extension__ ({ \
29091   uint16x8x4_t __s1 = __p1; \
29092   uint16x8x4_t __rev1; \
29093   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29094   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29095   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29096   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29097   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
29098 })
29099 #endif
29100 
29101 #ifdef __LITTLE_ENDIAN__
29102 #define vst4q_s8(__p0, __p1) __extension__ ({ \
29103   int8x16x4_t __s1 = __p1; \
29104   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
29105 })
29106 #else
29107 #define vst4q_s8(__p0, __p1) __extension__ ({ \
29108   int8x16x4_t __s1 = __p1; \
29109   int8x16x4_t __rev1; \
29110   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29111   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29112   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29113   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29114   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
29115 })
29116 #endif
29117 
29118 #ifdef __LITTLE_ENDIAN__
29119 #define vst4q_f32(__p0, __p1) __extension__ ({ \
29120   float32x4x4_t __s1 = __p1; \
29121   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 41); \
29122 })
29123 #else
29124 #define vst4q_f32(__p0, __p1) __extension__ ({ \
29125   float32x4x4_t __s1 = __p1; \
29126   float32x4x4_t __rev1; \
29127   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29128   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29129   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29130   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29131   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 41); \
29132 })
29133 #endif
29134 
29135 #ifdef __LITTLE_ENDIAN__
29136 #define vst4q_s32(__p0, __p1) __extension__ ({ \
29137   int32x4x4_t __s1 = __p1; \
29138   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 34); \
29139 })
29140 #else
29141 #define vst4q_s32(__p0, __p1) __extension__ ({ \
29142   int32x4x4_t __s1 = __p1; \
29143   int32x4x4_t __rev1; \
29144   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29145   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29146   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29147   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29148   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 34); \
29149 })
29150 #endif
29151 
29152 #ifdef __LITTLE_ENDIAN__
29153 #define vst4q_s16(__p0, __p1) __extension__ ({ \
29154   int16x8x4_t __s1 = __p1; \
29155   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 33); \
29156 })
29157 #else
29158 #define vst4q_s16(__p0, __p1) __extension__ ({ \
29159   int16x8x4_t __s1 = __p1; \
29160   int16x8x4_t __rev1; \
29161   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29162   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29163   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29164   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29165   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 33); \
29166 })
29167 #endif
29168 
29169 #ifdef __LITTLE_ENDIAN__
29170 #define vst4_u8(__p0, __p1) __extension__ ({ \
29171   uint8x8x4_t __s1 = __p1; \
29172   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
29173 })
29174 #else
29175 #define vst4_u8(__p0, __p1) __extension__ ({ \
29176   uint8x8x4_t __s1 = __p1; \
29177   uint8x8x4_t __rev1; \
29178   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29179   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29180   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29181   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29182   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
29183 })
29184 #endif
29185 
29186 #ifdef __LITTLE_ENDIAN__
29187 #define vst4_u32(__p0, __p1) __extension__ ({ \
29188   uint32x2x4_t __s1 = __p1; \
29189   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
29190 })
29191 #else
29192 #define vst4_u32(__p0, __p1) __extension__ ({ \
29193   uint32x2x4_t __s1 = __p1; \
29194   uint32x2x4_t __rev1; \
29195   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29196   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29197   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29198   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29199   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
29200 })
29201 #endif
29202 
29203 #define vst4_u64(__p0, __p1) __extension__ ({ \
29204   uint64x1x4_t __s1 = __p1; \
29205   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
29206 })
29207 #ifdef __LITTLE_ENDIAN__
29208 #define vst4_u16(__p0, __p1) __extension__ ({ \
29209   uint16x4x4_t __s1 = __p1; \
29210   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
29211 })
29212 #else
29213 #define vst4_u16(__p0, __p1) __extension__ ({ \
29214   uint16x4x4_t __s1 = __p1; \
29215   uint16x4x4_t __rev1; \
29216   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29217   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29218   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29219   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29220   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
29221 })
29222 #endif
29223 
29224 #ifdef __LITTLE_ENDIAN__
29225 #define vst4_s8(__p0, __p1) __extension__ ({ \
29226   int8x8x4_t __s1 = __p1; \
29227   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
29228 })
29229 #else
29230 #define vst4_s8(__p0, __p1) __extension__ ({ \
29231   int8x8x4_t __s1 = __p1; \
29232   int8x8x4_t __rev1; \
29233   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29234   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29235   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29236   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29237   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
29238 })
29239 #endif
29240 
29241 #ifdef __LITTLE_ENDIAN__
29242 #define vst4_f32(__p0, __p1) __extension__ ({ \
29243   float32x2x4_t __s1 = __p1; \
29244   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 9); \
29245 })
29246 #else
29247 #define vst4_f32(__p0, __p1) __extension__ ({ \
29248   float32x2x4_t __s1 = __p1; \
29249   float32x2x4_t __rev1; \
29250   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29251   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29252   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29253   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29254   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 9); \
29255 })
29256 #endif
29257 
29258 #ifdef __LITTLE_ENDIAN__
29259 #define vst4_s32(__p0, __p1) __extension__ ({ \
29260   int32x2x4_t __s1 = __p1; \
29261   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 2); \
29262 })
29263 #else
29264 #define vst4_s32(__p0, __p1) __extension__ ({ \
29265   int32x2x4_t __s1 = __p1; \
29266   int32x2x4_t __rev1; \
29267   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29268   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29269   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29270   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29271   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 2); \
29272 })
29273 #endif
29274 
29275 #define vst4_s64(__p0, __p1) __extension__ ({ \
29276   int64x1x4_t __s1 = __p1; \
29277   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 3); \
29278 })
29279 #ifdef __LITTLE_ENDIAN__
29280 #define vst4_s16(__p0, __p1) __extension__ ({ \
29281   int16x4x4_t __s1 = __p1; \
29282   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 1); \
29283 })
29284 #else
29285 #define vst4_s16(__p0, __p1) __extension__ ({ \
29286   int16x4x4_t __s1 = __p1; \
29287   int16x4x4_t __rev1; \
29288   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29289   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29290   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29291   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29292   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 1); \
29293 })
29294 #endif
29295 
29296 #ifdef __LITTLE_ENDIAN__
29297 #define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
29298   poly8x8x4_t __s1 = __p1; \
29299   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
29300 })
29301 #else
29302 #define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
29303   poly8x8x4_t __s1 = __p1; \
29304   poly8x8x4_t __rev1; \
29305   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29306   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29307   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29308   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29309   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
29310 })
29311 #endif
29312 
29313 #ifdef __LITTLE_ENDIAN__
29314 #define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
29315   poly16x4x4_t __s1 = __p1; \
29316   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
29317 })
29318 #else
29319 #define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
29320   poly16x4x4_t __s1 = __p1; \
29321   poly16x4x4_t __rev1; \
29322   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29323   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29324   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29325   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29326   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
29327 })
29328 #endif
29329 
29330 #ifdef __LITTLE_ENDIAN__
29331 #define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
29332   poly16x8x4_t __s1 = __p1; \
29333   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
29334 })
29335 #else
29336 #define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
29337   poly16x8x4_t __s1 = __p1; \
29338   poly16x8x4_t __rev1; \
29339   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29340   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29341   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29342   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29343   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
29344 })
29345 #endif
29346 
29347 #ifdef __LITTLE_ENDIAN__
29348 #define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29349   uint32x4x4_t __s1 = __p1; \
29350   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
29351 })
29352 #else
29353 #define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29354   uint32x4x4_t __s1 = __p1; \
29355   uint32x4x4_t __rev1; \
29356   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29357   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29358   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29359   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29360   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
29361 })
29362 #endif
29363 
29364 #ifdef __LITTLE_ENDIAN__
29365 #define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29366   uint16x8x4_t __s1 = __p1; \
29367   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
29368 })
29369 #else
29370 #define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29371   uint16x8x4_t __s1 = __p1; \
29372   uint16x8x4_t __rev1; \
29373   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29374   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29375   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29376   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29377   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
29378 })
29379 #endif
29380 
29381 #ifdef __LITTLE_ENDIAN__
29382 #define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29383   float32x4x4_t __s1 = __p1; \
29384   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 41); \
29385 })
29386 #else
29387 #define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29388   float32x4x4_t __s1 = __p1; \
29389   float32x4x4_t __rev1; \
29390   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29391   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29392   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29393   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29394   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 41); \
29395 })
29396 #endif
29397 
29398 #ifdef __LITTLE_ENDIAN__
29399 #define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29400   int32x4x4_t __s1 = __p1; \
29401   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 34); \
29402 })
29403 #else
29404 #define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29405   int32x4x4_t __s1 = __p1; \
29406   int32x4x4_t __rev1; \
29407   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29408   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29409   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29410   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29411   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 34); \
29412 })
29413 #endif
29414 
29415 #ifdef __LITTLE_ENDIAN__
29416 #define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29417   int16x8x4_t __s1 = __p1; \
29418   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 33); \
29419 })
29420 #else
29421 #define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29422   int16x8x4_t __s1 = __p1; \
29423   int16x8x4_t __rev1; \
29424   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29425   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29426   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29427   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29428   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 33); \
29429 })
29430 #endif
29431 
29432 #ifdef __LITTLE_ENDIAN__
29433 #define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
29434   uint8x8x4_t __s1 = __p1; \
29435   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
29436 })
29437 #else
29438 #define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
29439   uint8x8x4_t __s1 = __p1; \
29440   uint8x8x4_t __rev1; \
29441   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29442   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29443   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29444   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29445   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
29446 })
29447 #endif
29448 
29449 #ifdef __LITTLE_ENDIAN__
29450 #define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29451   uint32x2x4_t __s1 = __p1; \
29452   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
29453 })
29454 #else
29455 #define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29456   uint32x2x4_t __s1 = __p1; \
29457   uint32x2x4_t __rev1; \
29458   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29459   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29460   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29461   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29462   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
29463 })
29464 #endif
29465 
29466 #ifdef __LITTLE_ENDIAN__
29467 #define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29468   uint16x4x4_t __s1 = __p1; \
29469   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
29470 })
29471 #else
29472 #define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29473   uint16x4x4_t __s1 = __p1; \
29474   uint16x4x4_t __rev1; \
29475   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29476   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29477   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29478   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29479   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
29480 })
29481 #endif
29482 
29483 #ifdef __LITTLE_ENDIAN__
29484 #define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
29485   int8x8x4_t __s1 = __p1; \
29486   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
29487 })
29488 #else
29489 #define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
29490   int8x8x4_t __s1 = __p1; \
29491   int8x8x4_t __rev1; \
29492   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29493   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29494   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29495   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29496   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
29497 })
29498 #endif
29499 
29500 #ifdef __LITTLE_ENDIAN__
29501 #define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29502   float32x2x4_t __s1 = __p1; \
29503   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 9); \
29504 })
29505 #else
29506 #define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29507   float32x2x4_t __s1 = __p1; \
29508   float32x2x4_t __rev1; \
29509   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29510   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29511   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29512   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29513   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 9); \
29514 })
29515 #endif
29516 
29517 #ifdef __LITTLE_ENDIAN__
29518 #define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29519   int32x2x4_t __s1 = __p1; \
29520   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 2); \
29521 })
29522 #else
29523 #define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29524   int32x2x4_t __s1 = __p1; \
29525   int32x2x4_t __rev1; \
29526   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29527   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29528   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29529   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29530   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 2); \
29531 })
29532 #endif
29533 
29534 #ifdef __LITTLE_ENDIAN__
29535 #define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29536   int16x4x4_t __s1 = __p1; \
29537   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 1); \
29538 })
29539 #else
29540 #define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29541   int16x4x4_t __s1 = __p1; \
29542   int16x4x4_t __rev1; \
29543   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29544   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29545   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29546   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29547   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 1); \
29548 })
29549 #endif
29550 
29551 #ifdef __LITTLE_ENDIAN__
vsubq_u8(uint8x16_t __p0,uint8x16_t __p1)29552 __ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29553   uint8x16_t __ret;
29554   __ret = __p0 - __p1;
29555   return __ret;
29556 }
29557 #else
vsubq_u8(uint8x16_t __p0,uint8x16_t __p1)29558 __ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29559   uint8x16_t __ret;
29560   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29561   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29562   __ret = __rev0 - __rev1;
29563   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29564   return __ret;
29565 }
29566 #endif
29567 
29568 #ifdef __LITTLE_ENDIAN__
vsubq_u32(uint32x4_t __p0,uint32x4_t __p1)29569 __ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29570   uint32x4_t __ret;
29571   __ret = __p0 - __p1;
29572   return __ret;
29573 }
29574 #else
vsubq_u32(uint32x4_t __p0,uint32x4_t __p1)29575 __ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29576   uint32x4_t __ret;
29577   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29578   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29579   __ret = __rev0 - __rev1;
29580   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29581   return __ret;
29582 }
29583 #endif
29584 
29585 #ifdef __LITTLE_ENDIAN__
vsubq_u64(uint64x2_t __p0,uint64x2_t __p1)29586 __ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
29587   uint64x2_t __ret;
29588   __ret = __p0 - __p1;
29589   return __ret;
29590 }
29591 #else
vsubq_u64(uint64x2_t __p0,uint64x2_t __p1)29592 __ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
29593   uint64x2_t __ret;
29594   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29595   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29596   __ret = __rev0 - __rev1;
29597   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29598   return __ret;
29599 }
29600 #endif
29601 
29602 #ifdef __LITTLE_ENDIAN__
vsubq_u16(uint16x8_t __p0,uint16x8_t __p1)29603 __ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29604   uint16x8_t __ret;
29605   __ret = __p0 - __p1;
29606   return __ret;
29607 }
29608 #else
vsubq_u16(uint16x8_t __p0,uint16x8_t __p1)29609 __ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29610   uint16x8_t __ret;
29611   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29612   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29613   __ret = __rev0 - __rev1;
29614   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29615   return __ret;
29616 }
29617 #endif
29618 
29619 #ifdef __LITTLE_ENDIAN__
vsubq_s8(int8x16_t __p0,int8x16_t __p1)29620 __ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
29621   int8x16_t __ret;
29622   __ret = __p0 - __p1;
29623   return __ret;
29624 }
29625 #else
vsubq_s8(int8x16_t __p0,int8x16_t __p1)29626 __ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
29627   int8x16_t __ret;
29628   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29629   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29630   __ret = __rev0 - __rev1;
29631   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29632   return __ret;
29633 }
29634 #endif
29635 
29636 #ifdef __LITTLE_ENDIAN__
vsubq_f32(float32x4_t __p0,float32x4_t __p1)29637 __ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
29638   float32x4_t __ret;
29639   __ret = __p0 - __p1;
29640   return __ret;
29641 }
29642 #else
vsubq_f32(float32x4_t __p0,float32x4_t __p1)29643 __ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
29644   float32x4_t __ret;
29645   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29646   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29647   __ret = __rev0 - __rev1;
29648   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29649   return __ret;
29650 }
29651 #endif
29652 
29653 #ifdef __LITTLE_ENDIAN__
vsubq_s32(int32x4_t __p0,int32x4_t __p1)29654 __ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
29655   int32x4_t __ret;
29656   __ret = __p0 - __p1;
29657   return __ret;
29658 }
29659 #else
vsubq_s32(int32x4_t __p0,int32x4_t __p1)29660 __ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
29661   int32x4_t __ret;
29662   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29663   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29664   __ret = __rev0 - __rev1;
29665   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29666   return __ret;
29667 }
29668 #endif
29669 
29670 #ifdef __LITTLE_ENDIAN__
vsubq_s64(int64x2_t __p0,int64x2_t __p1)29671 __ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
29672   int64x2_t __ret;
29673   __ret = __p0 - __p1;
29674   return __ret;
29675 }
29676 #else
vsubq_s64(int64x2_t __p0,int64x2_t __p1)29677 __ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
29678   int64x2_t __ret;
29679   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29680   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29681   __ret = __rev0 - __rev1;
29682   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29683   return __ret;
29684 }
29685 #endif
29686 
29687 #ifdef __LITTLE_ENDIAN__
vsubq_s16(int16x8_t __p0,int16x8_t __p1)29688 __ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
29689   int16x8_t __ret;
29690   __ret = __p0 - __p1;
29691   return __ret;
29692 }
29693 #else
vsubq_s16(int16x8_t __p0,int16x8_t __p1)29694 __ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
29695   int16x8_t __ret;
29696   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29697   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29698   __ret = __rev0 - __rev1;
29699   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29700   return __ret;
29701 }
29702 #endif
29703 
29704 #ifdef __LITTLE_ENDIAN__
vsub_u8(uint8x8_t __p0,uint8x8_t __p1)29705 __ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
29706   uint8x8_t __ret;
29707   __ret = __p0 - __p1;
29708   return __ret;
29709 }
29710 #else
vsub_u8(uint8x8_t __p0,uint8x8_t __p1)29711 __ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
29712   uint8x8_t __ret;
29713   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29714   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29715   __ret = __rev0 - __rev1;
29716   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29717   return __ret;
29718 }
29719 #endif
29720 
29721 #ifdef __LITTLE_ENDIAN__
vsub_u32(uint32x2_t __p0,uint32x2_t __p1)29722 __ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
29723   uint32x2_t __ret;
29724   __ret = __p0 - __p1;
29725   return __ret;
29726 }
29727 #else
vsub_u32(uint32x2_t __p0,uint32x2_t __p1)29728 __ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
29729   uint32x2_t __ret;
29730   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29731   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29732   __ret = __rev0 - __rev1;
29733   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29734   return __ret;
29735 }
29736 #endif
29737 
vsub_u64(uint64x1_t __p0,uint64x1_t __p1)29738 __ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
29739   uint64x1_t __ret;
29740   __ret = __p0 - __p1;
29741   return __ret;
29742 }
29743 #ifdef __LITTLE_ENDIAN__
vsub_u16(uint16x4_t __p0,uint16x4_t __p1)29744 __ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
29745   uint16x4_t __ret;
29746   __ret = __p0 - __p1;
29747   return __ret;
29748 }
29749 #else
vsub_u16(uint16x4_t __p0,uint16x4_t __p1)29750 __ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
29751   uint16x4_t __ret;
29752   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29753   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29754   __ret = __rev0 - __rev1;
29755   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29756   return __ret;
29757 }
29758 #endif
29759 
29760 #ifdef __LITTLE_ENDIAN__
vsub_s8(int8x8_t __p0,int8x8_t __p1)29761 __ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
29762   int8x8_t __ret;
29763   __ret = __p0 - __p1;
29764   return __ret;
29765 }
29766 #else
vsub_s8(int8x8_t __p0,int8x8_t __p1)29767 __ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
29768   int8x8_t __ret;
29769   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29770   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29771   __ret = __rev0 - __rev1;
29772   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29773   return __ret;
29774 }
29775 #endif
29776 
29777 #ifdef __LITTLE_ENDIAN__
vsub_f32(float32x2_t __p0,float32x2_t __p1)29778 __ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
29779   float32x2_t __ret;
29780   __ret = __p0 - __p1;
29781   return __ret;
29782 }
29783 #else
vsub_f32(float32x2_t __p0,float32x2_t __p1)29784 __ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
29785   float32x2_t __ret;
29786   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29787   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29788   __ret = __rev0 - __rev1;
29789   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29790   return __ret;
29791 }
29792 #endif
29793 
29794 #ifdef __LITTLE_ENDIAN__
vsub_s32(int32x2_t __p0,int32x2_t __p1)29795 __ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
29796   int32x2_t __ret;
29797   __ret = __p0 - __p1;
29798   return __ret;
29799 }
29800 #else
vsub_s32(int32x2_t __p0,int32x2_t __p1)29801 __ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
29802   int32x2_t __ret;
29803   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29804   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29805   __ret = __rev0 - __rev1;
29806   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29807   return __ret;
29808 }
29809 #endif
29810 
vsub_s64(int64x1_t __p0,int64x1_t __p1)29811 __ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
29812   int64x1_t __ret;
29813   __ret = __p0 - __p1;
29814   return __ret;
29815 }
29816 #ifdef __LITTLE_ENDIAN__
vsub_s16(int16x4_t __p0,int16x4_t __p1)29817 __ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
29818   int16x4_t __ret;
29819   __ret = __p0 - __p1;
29820   return __ret;
29821 }
29822 #else
vsub_s16(int16x4_t __p0,int16x4_t __p1)29823 __ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
29824   int16x4_t __ret;
29825   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29826   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29827   __ret = __rev0 - __rev1;
29828   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29829   return __ret;
29830 }
29831 #endif
29832 
29833 #ifdef __LITTLE_ENDIAN__
vsubhn_u32(uint32x4_t __p0,uint32x4_t __p1)29834 __ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
29835   uint16x4_t __ret;
29836   __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
29837   return __ret;
29838 }
29839 #else
vsubhn_u32(uint32x4_t __p0,uint32x4_t __p1)29840 __ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
29841   uint16x4_t __ret;
29842   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29843   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29844   __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
29845   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29846   return __ret;
29847 }
__noswap_vsubhn_u32(uint32x4_t __p0,uint32x4_t __p1)29848 __ai uint16x4_t __noswap_vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
29849   uint16x4_t __ret;
29850   __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
29851   return __ret;
29852 }
29853 #endif
29854 
29855 #ifdef __LITTLE_ENDIAN__
vsubhn_u64(uint64x2_t __p0,uint64x2_t __p1)29856 __ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
29857   uint32x2_t __ret;
29858   __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
29859   return __ret;
29860 }
29861 #else
vsubhn_u64(uint64x2_t __p0,uint64x2_t __p1)29862 __ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
29863   uint32x2_t __ret;
29864   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29865   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29866   __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
29867   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29868   return __ret;
29869 }
__noswap_vsubhn_u64(uint64x2_t __p0,uint64x2_t __p1)29870 __ai uint32x2_t __noswap_vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
29871   uint32x2_t __ret;
29872   __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
29873   return __ret;
29874 }
29875 #endif
29876 
29877 #ifdef __LITTLE_ENDIAN__
vsubhn_u16(uint16x8_t __p0,uint16x8_t __p1)29878 __ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
29879   uint8x8_t __ret;
29880   __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
29881   return __ret;
29882 }
29883 #else
vsubhn_u16(uint16x8_t __p0,uint16x8_t __p1)29884 __ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
29885   uint8x8_t __ret;
29886   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29887   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29888   __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
29889   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29890   return __ret;
29891 }
__noswap_vsubhn_u16(uint16x8_t __p0,uint16x8_t __p1)29892 __ai uint8x8_t __noswap_vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
29893   uint8x8_t __ret;
29894   __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
29895   return __ret;
29896 }
29897 #endif
29898 
29899 #ifdef __LITTLE_ENDIAN__
vsubhn_s32(int32x4_t __p0,int32x4_t __p1)29900 __ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
29901   int16x4_t __ret;
29902   __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
29903   return __ret;
29904 }
29905 #else
vsubhn_s32(int32x4_t __p0,int32x4_t __p1)29906 __ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
29907   int16x4_t __ret;
29908   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29909   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29910   __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
29911   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29912   return __ret;
29913 }
__noswap_vsubhn_s32(int32x4_t __p0,int32x4_t __p1)29914 __ai int16x4_t __noswap_vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
29915   int16x4_t __ret;
29916   __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
29917   return __ret;
29918 }
29919 #endif
29920 
29921 #ifdef __LITTLE_ENDIAN__
vsubhn_s64(int64x2_t __p0,int64x2_t __p1)29922 __ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
29923   int32x2_t __ret;
29924   __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
29925   return __ret;
29926 }
29927 #else
vsubhn_s64(int64x2_t __p0,int64x2_t __p1)29928 __ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
29929   int32x2_t __ret;
29930   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29931   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29932   __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
29933   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29934   return __ret;
29935 }
__noswap_vsubhn_s64(int64x2_t __p0,int64x2_t __p1)29936 __ai int32x2_t __noswap_vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
29937   int32x2_t __ret;
29938   __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
29939   return __ret;
29940 }
29941 #endif
29942 
29943 #ifdef __LITTLE_ENDIAN__
vsubhn_s16(int16x8_t __p0,int16x8_t __p1)29944 __ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
29945   int8x8_t __ret;
29946   __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
29947   return __ret;
29948 }
29949 #else
vsubhn_s16(int16x8_t __p0,int16x8_t __p1)29950 __ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
29951   int8x8_t __ret;
29952   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29953   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29954   __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
29955   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29956   return __ret;
29957 }
__noswap_vsubhn_s16(int16x8_t __p0,int16x8_t __p1)29958 __ai int8x8_t __noswap_vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
29959   int8x8_t __ret;
29960   __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
29961   return __ret;
29962 }
29963 #endif
29964 
29965 #ifdef __LITTLE_ENDIAN__
vsubl_u8(uint8x8_t __p0,uint8x8_t __p1)29966 __ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
29967   uint16x8_t __ret;
29968   __ret = vmovl_u8(__p0) - vmovl_u8(__p1);
29969   return __ret;
29970 }
29971 #else
vsubl_u8(uint8x8_t __p0,uint8x8_t __p1)29972 __ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
29973   uint16x8_t __ret;
29974   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29975   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29976   __ret = __noswap_vmovl_u8(__rev0) - __noswap_vmovl_u8(__rev1);
29977   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29978   return __ret;
29979 }
29980 #endif
29981 
29982 #ifdef __LITTLE_ENDIAN__
vsubl_u32(uint32x2_t __p0,uint32x2_t __p1)29983 __ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
29984   uint64x2_t __ret;
29985   __ret = vmovl_u32(__p0) - vmovl_u32(__p1);
29986   return __ret;
29987 }
29988 #else
vsubl_u32(uint32x2_t __p0,uint32x2_t __p1)29989 __ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
29990   uint64x2_t __ret;
29991   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29992   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29993   __ret = __noswap_vmovl_u32(__rev0) - __noswap_vmovl_u32(__rev1);
29994   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29995   return __ret;
29996 }
29997 #endif
29998 
29999 #ifdef __LITTLE_ENDIAN__
vsubl_u16(uint16x4_t __p0,uint16x4_t __p1)30000 __ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
30001   uint32x4_t __ret;
30002   __ret = vmovl_u16(__p0) - vmovl_u16(__p1);
30003   return __ret;
30004 }
30005 #else
vsubl_u16(uint16x4_t __p0,uint16x4_t __p1)30006 __ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
30007   uint32x4_t __ret;
30008   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30009   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30010   __ret = __noswap_vmovl_u16(__rev0) - __noswap_vmovl_u16(__rev1);
30011   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
30012   return __ret;
30013 }
30014 #endif
30015 
30016 #ifdef __LITTLE_ENDIAN__
vsubl_s8(int8x8_t __p0,int8x8_t __p1)30017 __ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
30018   int16x8_t __ret;
30019   __ret = vmovl_s8(__p0) - vmovl_s8(__p1);
30020   return __ret;
30021 }
30022 #else
vsubl_s8(int8x8_t __p0,int8x8_t __p1)30023 __ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
30024   int16x8_t __ret;
30025   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30026   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30027   __ret = __noswap_vmovl_s8(__rev0) - __noswap_vmovl_s8(__rev1);
30028   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30029   return __ret;
30030 }
30031 #endif
30032 
30033 #ifdef __LITTLE_ENDIAN__
vsubl_s32(int32x2_t __p0,int32x2_t __p1)30034 __ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
30035   int64x2_t __ret;
30036   __ret = vmovl_s32(__p0) - vmovl_s32(__p1);
30037   return __ret;
30038 }
30039 #else
vsubl_s32(int32x2_t __p0,int32x2_t __p1)30040 __ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
30041   int64x2_t __ret;
30042   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30043   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30044   __ret = __noswap_vmovl_s32(__rev0) - __noswap_vmovl_s32(__rev1);
30045   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
30046   return __ret;
30047 }
30048 #endif
30049 
30050 #ifdef __LITTLE_ENDIAN__
vsubl_s16(int16x4_t __p0,int16x4_t __p1)30051 __ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
30052   int32x4_t __ret;
30053   __ret = vmovl_s16(__p0) - vmovl_s16(__p1);
30054   return __ret;
30055 }
30056 #else
vsubl_s16(int16x4_t __p0,int16x4_t __p1)30057 __ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
30058   int32x4_t __ret;
30059   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30060   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30061   __ret = __noswap_vmovl_s16(__rev0) - __noswap_vmovl_s16(__rev1);
30062   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
30063   return __ret;
30064 }
30065 #endif
30066 
30067 #ifdef __LITTLE_ENDIAN__
vsubw_u8(uint16x8_t __p0,uint8x8_t __p1)30068 __ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
30069   uint16x8_t __ret;
30070   __ret = __p0 - vmovl_u8(__p1);
30071   return __ret;
30072 }
30073 #else
vsubw_u8(uint16x8_t __p0,uint8x8_t __p1)30074 __ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
30075   uint16x8_t __ret;
30076   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30077   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30078   __ret = __rev0 - __noswap_vmovl_u8(__rev1);
30079   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30080   return __ret;
30081 }
30082 #endif
30083 
30084 #ifdef __LITTLE_ENDIAN__
vsubw_u32(uint64x2_t __p0,uint32x2_t __p1)30085 __ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
30086   uint64x2_t __ret;
30087   __ret = __p0 - vmovl_u32(__p1);
30088   return __ret;
30089 }
30090 #else
vsubw_u32(uint64x2_t __p0,uint32x2_t __p1)30091 __ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
30092   uint64x2_t __ret;
30093   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30094   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30095   __ret = __rev0 - __noswap_vmovl_u32(__rev1);
30096   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
30097   return __ret;
30098 }
30099 #endif
30100 
30101 #ifdef __LITTLE_ENDIAN__
vsubw_u16(uint32x4_t __p0,uint16x4_t __p1)30102 __ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
30103   uint32x4_t __ret;
30104   __ret = __p0 - vmovl_u16(__p1);
30105   return __ret;
30106 }
30107 #else
vsubw_u16(uint32x4_t __p0,uint16x4_t __p1)30108 __ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
30109   uint32x4_t __ret;
30110   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30111   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30112   __ret = __rev0 - __noswap_vmovl_u16(__rev1);
30113   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
30114   return __ret;
30115 }
30116 #endif
30117 
30118 #ifdef __LITTLE_ENDIAN__
vsubw_s8(int16x8_t __p0,int8x8_t __p1)30119 __ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
30120   int16x8_t __ret;
30121   __ret = __p0 - vmovl_s8(__p1);
30122   return __ret;
30123 }
30124 #else
vsubw_s8(int16x8_t __p0,int8x8_t __p1)30125 __ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
30126   int16x8_t __ret;
30127   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30128   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30129   __ret = __rev0 - __noswap_vmovl_s8(__rev1);
30130   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30131   return __ret;
30132 }
30133 #endif
30134 
30135 #ifdef __LITTLE_ENDIAN__
vsubw_s32(int64x2_t __p0,int32x2_t __p1)30136 __ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
30137   int64x2_t __ret;
30138   __ret = __p0 - vmovl_s32(__p1);
30139   return __ret;
30140 }
30141 #else
vsubw_s32(int64x2_t __p0,int32x2_t __p1)30142 __ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
30143   int64x2_t __ret;
30144   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30145   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30146   __ret = __rev0 - __noswap_vmovl_s32(__rev1);
30147   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
30148   return __ret;
30149 }
30150 #endif
30151 
30152 #ifdef __LITTLE_ENDIAN__
vsubw_s16(int32x4_t __p0,int16x4_t __p1)30153 __ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
30154   int32x4_t __ret;
30155   __ret = __p0 - vmovl_s16(__p1);
30156   return __ret;
30157 }
30158 #else
vsubw_s16(int32x4_t __p0,int16x4_t __p1)30159 __ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
30160   int32x4_t __ret;
30161   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30162   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30163   __ret = __rev0 - __noswap_vmovl_s16(__rev1);
30164   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
30165   return __ret;
30166 }
30167 #endif
30168 
30169 #ifdef __LITTLE_ENDIAN__
vtbl1_p8(poly8x8_t __p0,uint8x8_t __p1)30170 __ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
30171   poly8x8_t __ret;
30172   __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
30173   return __ret;
30174 }
30175 #else
vtbl1_p8(poly8x8_t __p0,uint8x8_t __p1)30176 __ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
30177   poly8x8_t __ret;
30178   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30179   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30180   __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
30181   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30182   return __ret;
30183 }
30184 #endif
30185 
30186 #ifdef __LITTLE_ENDIAN__
vtbl1_u8(uint8x8_t __p0,uint8x8_t __p1)30187 __ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
30188   uint8x8_t __ret;
30189   __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
30190   return __ret;
30191 }
30192 #else
vtbl1_u8(uint8x8_t __p0,uint8x8_t __p1)30193 __ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
30194   uint8x8_t __ret;
30195   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30196   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30197   __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
30198   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30199   return __ret;
30200 }
30201 #endif
30202 
30203 #ifdef __LITTLE_ENDIAN__
vtbl1_s8(int8x8_t __p0,int8x8_t __p1)30204 __ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
30205   int8x8_t __ret;
30206   __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
30207   return __ret;
30208 }
30209 #else
vtbl1_s8(int8x8_t __p0,int8x8_t __p1)30210 __ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
30211   int8x8_t __ret;
30212   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30213   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30214   __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
30215   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30216   return __ret;
30217 }
30218 #endif
30219 
30220 #ifdef __LITTLE_ENDIAN__
vtbl2_p8(poly8x8x2_t __p0,uint8x8_t __p1)30221 __ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
30222   poly8x8_t __ret;
30223   __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 4);
30224   return __ret;
30225 }
30226 #else
vtbl2_p8(poly8x8x2_t __p0,uint8x8_t __p1)30227 __ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
30228   poly8x8_t __ret;
30229   poly8x8x2_t __rev0;
30230   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30231   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30232   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30233   __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 4);
30234   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30235   return __ret;
30236 }
30237 #endif
30238 
30239 #ifdef __LITTLE_ENDIAN__
vtbl2_u8(uint8x8x2_t __p0,uint8x8_t __p1)30240 __ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
30241   uint8x8_t __ret;
30242   __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 16);
30243   return __ret;
30244 }
30245 #else
vtbl2_u8(uint8x8x2_t __p0,uint8x8_t __p1)30246 __ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
30247   uint8x8_t __ret;
30248   uint8x8x2_t __rev0;
30249   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30250   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30251   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30252   __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 16);
30253   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30254   return __ret;
30255 }
30256 #endif
30257 
30258 #ifdef __LITTLE_ENDIAN__
vtbl2_s8(int8x8x2_t __p0,int8x8_t __p1)30259 __ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
30260   int8x8_t __ret;
30261   __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 0);
30262   return __ret;
30263 }
30264 #else
vtbl2_s8(int8x8x2_t __p0,int8x8_t __p1)30265 __ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
30266   int8x8_t __ret;
30267   int8x8x2_t __rev0;
30268   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30269   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30270   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30271   __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 0);
30272   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30273   return __ret;
30274 }
30275 #endif
30276 
30277 #ifdef __LITTLE_ENDIAN__
vtbl3_p8(poly8x8x3_t __p0,uint8x8_t __p1)30278 __ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
30279   poly8x8_t __ret;
30280   __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 4);
30281   return __ret;
30282 }
30283 #else
vtbl3_p8(poly8x8x3_t __p0,uint8x8_t __p1)30284 __ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
30285   poly8x8_t __ret;
30286   poly8x8x3_t __rev0;
30287   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30288   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30289   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30290   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30291   __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 4);
30292   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30293   return __ret;
30294 }
30295 #endif
30296 
30297 #ifdef __LITTLE_ENDIAN__
vtbl3_u8(uint8x8x3_t __p0,uint8x8_t __p1)30298 __ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
30299   uint8x8_t __ret;
30300   __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 16);
30301   return __ret;
30302 }
30303 #else
vtbl3_u8(uint8x8x3_t __p0,uint8x8_t __p1)30304 __ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
30305   uint8x8_t __ret;
30306   uint8x8x3_t __rev0;
30307   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30308   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30309   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30310   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30311   __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 16);
30312   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30313   return __ret;
30314 }
30315 #endif
30316 
30317 #ifdef __LITTLE_ENDIAN__
vtbl3_s8(int8x8x3_t __p0,int8x8_t __p1)30318 __ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
30319   int8x8_t __ret;
30320   __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 0);
30321   return __ret;
30322 }
30323 #else
vtbl3_s8(int8x8x3_t __p0,int8x8_t __p1)30324 __ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
30325   int8x8_t __ret;
30326   int8x8x3_t __rev0;
30327   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30328   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30329   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30330   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30331   __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 0);
30332   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30333   return __ret;
30334 }
30335 #endif
30336 
30337 #ifdef __LITTLE_ENDIAN__
vtbl4_p8(poly8x8x4_t __p0,uint8x8_t __p1)30338 __ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
30339   poly8x8_t __ret;
30340   __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 4);
30341   return __ret;
30342 }
30343 #else
vtbl4_p8(poly8x8x4_t __p0,uint8x8_t __p1)30344 __ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
30345   poly8x8_t __ret;
30346   poly8x8x4_t __rev0;
30347   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30348   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30349   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30350   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30351   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30352   __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 4);
30353   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30354   return __ret;
30355 }
30356 #endif
30357 
30358 #ifdef __LITTLE_ENDIAN__
vtbl4_u8(uint8x8x4_t __p0,uint8x8_t __p1)30359 __ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
30360   uint8x8_t __ret;
30361   __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 16);
30362   return __ret;
30363 }
30364 #else
vtbl4_u8(uint8x8x4_t __p0,uint8x8_t __p1)30365 __ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
30366   uint8x8_t __ret;
30367   uint8x8x4_t __rev0;
30368   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30369   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30370   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30371   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30372   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30373   __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 16);
30374   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30375   return __ret;
30376 }
30377 #endif
30378 
30379 #ifdef __LITTLE_ENDIAN__
vtbl4_s8(int8x8x4_t __p0,int8x8_t __p1)30380 __ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
30381   int8x8_t __ret;
30382   __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 0);
30383   return __ret;
30384 }
30385 #else
vtbl4_s8(int8x8x4_t __p0,int8x8_t __p1)30386 __ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
30387   int8x8_t __ret;
30388   int8x8x4_t __rev0;
30389   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30390   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30391   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30392   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30393   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30394   __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 0);
30395   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30396   return __ret;
30397 }
30398 #endif
30399 
30400 #ifdef __LITTLE_ENDIAN__
vtbx1_p8(poly8x8_t __p0,poly8x8_t __p1,uint8x8_t __p2)30401 __ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
30402   poly8x8_t __ret;
30403   __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
30404   return __ret;
30405 }
30406 #else
vtbx1_p8(poly8x8_t __p0,poly8x8_t __p1,uint8x8_t __p2)30407 __ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
30408   poly8x8_t __ret;
30409   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30410   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30411   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30412   __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
30413   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30414   return __ret;
30415 }
30416 #endif
30417 
30418 #ifdef __LITTLE_ENDIAN__
vtbx1_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)30419 __ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
30420   uint8x8_t __ret;
30421   __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
30422   return __ret;
30423 }
30424 #else
vtbx1_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)30425 __ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
30426   uint8x8_t __ret;
30427   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30428   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30429   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30430   __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
30431   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30432   return __ret;
30433 }
30434 #endif
30435 
30436 #ifdef __LITTLE_ENDIAN__
vtbx1_s8(int8x8_t __p0,int8x8_t __p1,int8x8_t __p2)30437 __ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
30438   int8x8_t __ret;
30439   __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
30440   return __ret;
30441 }
30442 #else
vtbx1_s8(int8x8_t __p0,int8x8_t __p1,int8x8_t __p2)30443 __ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
30444   int8x8_t __ret;
30445   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30446   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30447   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30448   __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
30449   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30450   return __ret;
30451 }
30452 #endif
30453 
30454 #ifdef __LITTLE_ENDIAN__
vtbx2_p8(poly8x8_t __p0,poly8x8x2_t __p1,uint8x8_t __p2)30455 __ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
30456   poly8x8_t __ret;
30457   __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 4);
30458   return __ret;
30459 }
30460 #else
vtbx2_p8(poly8x8_t __p0,poly8x8x2_t __p1,uint8x8_t __p2)30461 __ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
30462   poly8x8_t __ret;
30463   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30464   poly8x8x2_t __rev1;
30465   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30466   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30467   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30468   __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 4);
30469   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30470   return __ret;
30471 }
30472 #endif
30473 
30474 #ifdef __LITTLE_ENDIAN__
vtbx2_u8(uint8x8_t __p0,uint8x8x2_t __p1,uint8x8_t __p2)30475 __ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
30476   uint8x8_t __ret;
30477   __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 16);
30478   return __ret;
30479 }
30480 #else
vtbx2_u8(uint8x8_t __p0,uint8x8x2_t __p1,uint8x8_t __p2)30481 __ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
30482   uint8x8_t __ret;
30483   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30484   uint8x8x2_t __rev1;
30485   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30486   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30487   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30488   __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 16);
30489   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30490   return __ret;
30491 }
30492 #endif
30493 
30494 #ifdef __LITTLE_ENDIAN__
vtbx2_s8(int8x8_t __p0,int8x8x2_t __p1,int8x8_t __p2)30495 __ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
30496   int8x8_t __ret;
30497   __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 0);
30498   return __ret;
30499 }
30500 #else
vtbx2_s8(int8x8_t __p0,int8x8x2_t __p1,int8x8_t __p2)30501 __ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
30502   int8x8_t __ret;
30503   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30504   int8x8x2_t __rev1;
30505   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30506   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30507   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30508   __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 0);
30509   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30510   return __ret;
30511 }
30512 #endif
30513 
30514 #ifdef __LITTLE_ENDIAN__
vtbx3_p8(poly8x8_t __p0,poly8x8x3_t __p1,uint8x8_t __p2)30515 __ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
30516   poly8x8_t __ret;
30517   __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 4);
30518   return __ret;
30519 }
30520 #else
vtbx3_p8(poly8x8_t __p0,poly8x8x3_t __p1,uint8x8_t __p2)30521 __ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
30522   poly8x8_t __ret;
30523   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30524   poly8x8x3_t __rev1;
30525   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30526   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30527   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30528   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30529   __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 4);
30530   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30531   return __ret;
30532 }
30533 #endif
30534 
30535 #ifdef __LITTLE_ENDIAN__
vtbx3_u8(uint8x8_t __p0,uint8x8x3_t __p1,uint8x8_t __p2)30536 __ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
30537   uint8x8_t __ret;
30538   __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 16);
30539   return __ret;
30540 }
30541 #else
vtbx3_u8(uint8x8_t __p0,uint8x8x3_t __p1,uint8x8_t __p2)30542 __ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
30543   uint8x8_t __ret;
30544   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30545   uint8x8x3_t __rev1;
30546   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30547   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30548   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30549   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30550   __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 16);
30551   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30552   return __ret;
30553 }
30554 #endif
30555 
30556 #ifdef __LITTLE_ENDIAN__
vtbx3_s8(int8x8_t __p0,int8x8x3_t __p1,int8x8_t __p2)30557 __ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
30558   int8x8_t __ret;
30559   __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 0);
30560   return __ret;
30561 }
30562 #else
vtbx3_s8(int8x8_t __p0,int8x8x3_t __p1,int8x8_t __p2)30563 __ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
30564   int8x8_t __ret;
30565   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30566   int8x8x3_t __rev1;
30567   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30568   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30569   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30570   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30571   __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 0);
30572   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30573   return __ret;
30574 }
30575 #endif
30576 
30577 #ifdef __LITTLE_ENDIAN__
vtbx4_p8(poly8x8_t __p0,poly8x8x4_t __p1,uint8x8_t __p2)30578 __ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
30579   poly8x8_t __ret;
30580   __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 4);
30581   return __ret;
30582 }
30583 #else
vtbx4_p8(poly8x8_t __p0,poly8x8x4_t __p1,uint8x8_t __p2)30584 __ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
30585   poly8x8_t __ret;
30586   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30587   poly8x8x4_t __rev1;
30588   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30589   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30590   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30591   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30592   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30593   __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 4);
30594   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30595   return __ret;
30596 }
30597 #endif
30598 
30599 #ifdef __LITTLE_ENDIAN__
vtbx4_u8(uint8x8_t __p0,uint8x8x4_t __p1,uint8x8_t __p2)30600 __ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
30601   uint8x8_t __ret;
30602   __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 16);
30603   return __ret;
30604 }
30605 #else
vtbx4_u8(uint8x8_t __p0,uint8x8x4_t __p1,uint8x8_t __p2)30606 __ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
30607   uint8x8_t __ret;
30608   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30609   uint8x8x4_t __rev1;
30610   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30611   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30612   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30613   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30614   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30615   __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 16);
30616   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30617   return __ret;
30618 }
30619 #endif
30620 
30621 #ifdef __LITTLE_ENDIAN__
vtbx4_s8(int8x8_t __p0,int8x8x4_t __p1,int8x8_t __p2)30622 __ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
30623   int8x8_t __ret;
30624   __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 0);
30625   return __ret;
30626 }
30627 #else
vtbx4_s8(int8x8_t __p0,int8x8x4_t __p1,int8x8_t __p2)30628 __ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
30629   int8x8_t __ret;
30630   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30631   int8x8x4_t __rev1;
30632   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30633   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30634   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30635   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30636   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30637   __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 0);
30638   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30639   return __ret;
30640 }
30641 #endif
30642 
30643 #ifdef __LITTLE_ENDIAN__
vtrn_p8(poly8x8_t __p0,poly8x8_t __p1)30644 __ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
30645   poly8x8x2_t __ret;
30646   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
30647   return __ret;
30648 }
30649 #else
vtrn_p8(poly8x8_t __p0,poly8x8_t __p1)30650 __ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
30651   poly8x8x2_t __ret;
30652   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30653   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30654   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
30655 
30656   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30657   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30658   return __ret;
30659 }
30660 #endif
30661 
30662 #ifdef __LITTLE_ENDIAN__
vtrn_p16(poly16x4_t __p0,poly16x4_t __p1)30663 __ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
30664   poly16x4x2_t __ret;
30665   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
30666   return __ret;
30667 }
30668 #else
vtrn_p16(poly16x4_t __p0,poly16x4_t __p1)30669 __ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
30670   poly16x4x2_t __ret;
30671   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30672   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30673   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
30674 
30675   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30676   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30677   return __ret;
30678 }
30679 #endif
30680 
30681 #ifdef __LITTLE_ENDIAN__
vtrnq_p8(poly8x16_t __p0,poly8x16_t __p1)30682 __ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
30683   poly8x16x2_t __ret;
30684   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
30685   return __ret;
30686 }
30687 #else
vtrnq_p8(poly8x16_t __p0,poly8x16_t __p1)30688 __ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
30689   poly8x16x2_t __ret;
30690   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30691   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30692   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
30693 
30694   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30695   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30696   return __ret;
30697 }
30698 #endif
30699 
30700 #ifdef __LITTLE_ENDIAN__
vtrnq_p16(poly16x8_t __p0,poly16x8_t __p1)30701 __ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
30702   poly16x8x2_t __ret;
30703   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
30704   return __ret;
30705 }
30706 #else
vtrnq_p16(poly16x8_t __p0,poly16x8_t __p1)30707 __ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
30708   poly16x8x2_t __ret;
30709   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30710   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30711   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
30712 
30713   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30714   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30715   return __ret;
30716 }
30717 #endif
30718 
30719 #ifdef __LITTLE_ENDIAN__
vtrnq_u8(uint8x16_t __p0,uint8x16_t __p1)30720 __ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
30721   uint8x16x2_t __ret;
30722   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
30723   return __ret;
30724 }
30725 #else
vtrnq_u8(uint8x16_t __p0,uint8x16_t __p1)30726 __ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
30727   uint8x16x2_t __ret;
30728   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30729   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30730   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
30731 
30732   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30733   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30734   return __ret;
30735 }
30736 #endif
30737 
30738 #ifdef __LITTLE_ENDIAN__
vtrnq_u32(uint32x4_t __p0,uint32x4_t __p1)30739 __ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
30740   uint32x4x2_t __ret;
30741   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
30742   return __ret;
30743 }
30744 #else
vtrnq_u32(uint32x4_t __p0,uint32x4_t __p1)30745 __ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
30746   uint32x4x2_t __ret;
30747   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30748   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30749   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
30750 
30751   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30752   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30753   return __ret;
30754 }
30755 #endif
30756 
30757 #ifdef __LITTLE_ENDIAN__
vtrnq_u16(uint16x8_t __p0,uint16x8_t __p1)30758 __ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
30759   uint16x8x2_t __ret;
30760   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
30761   return __ret;
30762 }
30763 #else
vtrnq_u16(uint16x8_t __p0,uint16x8_t __p1)30764 __ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
30765   uint16x8x2_t __ret;
30766   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30767   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30768   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
30769 
30770   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30771   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30772   return __ret;
30773 }
30774 #endif
30775 
30776 #ifdef __LITTLE_ENDIAN__
vtrnq_s8(int8x16_t __p0,int8x16_t __p1)30777 __ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
30778   int8x16x2_t __ret;
30779   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
30780   return __ret;
30781 }
30782 #else
vtrnq_s8(int8x16_t __p0,int8x16_t __p1)30783 __ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
30784   int8x16x2_t __ret;
30785   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30786   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30787   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
30788 
30789   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30790   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30791   return __ret;
30792 }
30793 #endif
30794 
30795 #ifdef __LITTLE_ENDIAN__
vtrnq_f32(float32x4_t __p0,float32x4_t __p1)30796 __ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
30797   float32x4x2_t __ret;
30798   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
30799   return __ret;
30800 }
30801 #else
vtrnq_f32(float32x4_t __p0,float32x4_t __p1)30802 __ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
30803   float32x4x2_t __ret;
30804   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30805   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30806   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
30807 
30808   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30809   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30810   return __ret;
30811 }
30812 #endif
30813 
30814 #ifdef __LITTLE_ENDIAN__
vtrnq_s32(int32x4_t __p0,int32x4_t __p1)30815 __ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
30816   int32x4x2_t __ret;
30817   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
30818   return __ret;
30819 }
30820 #else
vtrnq_s32(int32x4_t __p0,int32x4_t __p1)30821 __ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
30822   int32x4x2_t __ret;
30823   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30824   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30825   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
30826 
30827   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30828   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30829   return __ret;
30830 }
30831 #endif
30832 
30833 #ifdef __LITTLE_ENDIAN__
vtrnq_s16(int16x8_t __p0,int16x8_t __p1)30834 __ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
30835   int16x8x2_t __ret;
30836   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
30837   return __ret;
30838 }
30839 #else
vtrnq_s16(int16x8_t __p0,int16x8_t __p1)30840 __ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
30841   int16x8x2_t __ret;
30842   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30843   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30844   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
30845 
30846   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30847   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30848   return __ret;
30849 }
30850 #endif
30851 
30852 #ifdef __LITTLE_ENDIAN__
vtrn_u8(uint8x8_t __p0,uint8x8_t __p1)30853 __ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
30854   uint8x8x2_t __ret;
30855   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
30856   return __ret;
30857 }
30858 #else
vtrn_u8(uint8x8_t __p0,uint8x8_t __p1)30859 __ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
30860   uint8x8x2_t __ret;
30861   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30862   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30863   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
30864 
30865   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30866   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30867   return __ret;
30868 }
30869 #endif
30870 
30871 #ifdef __LITTLE_ENDIAN__
vtrn_u32(uint32x2_t __p0,uint32x2_t __p1)30872 __ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
30873   uint32x2x2_t __ret;
30874   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
30875   return __ret;
30876 }
30877 #else
vtrn_u32(uint32x2_t __p0,uint32x2_t __p1)30878 __ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
30879   uint32x2x2_t __ret;
30880   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30881   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30882   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
30883 
30884   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
30885   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
30886   return __ret;
30887 }
30888 #endif
30889 
30890 #ifdef __LITTLE_ENDIAN__
vtrn_u16(uint16x4_t __p0,uint16x4_t __p1)30891 __ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
30892   uint16x4x2_t __ret;
30893   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
30894   return __ret;
30895 }
30896 #else
vtrn_u16(uint16x4_t __p0,uint16x4_t __p1)30897 __ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
30898   uint16x4x2_t __ret;
30899   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30900   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30901   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
30902 
30903   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30904   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30905   return __ret;
30906 }
30907 #endif
30908 
30909 #ifdef __LITTLE_ENDIAN__
vtrn_s8(int8x8_t __p0,int8x8_t __p1)30910 __ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
30911   int8x8x2_t __ret;
30912   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
30913   return __ret;
30914 }
30915 #else
vtrn_s8(int8x8_t __p0,int8x8_t __p1)30916 __ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
30917   int8x8x2_t __ret;
30918   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30919   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30920   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
30921 
30922   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30923   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30924   return __ret;
30925 }
30926 #endif
30927 
30928 #ifdef __LITTLE_ENDIAN__
vtrn_f32(float32x2_t __p0,float32x2_t __p1)30929 __ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
30930   float32x2x2_t __ret;
30931   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
30932   return __ret;
30933 }
30934 #else
vtrn_f32(float32x2_t __p0,float32x2_t __p1)30935 __ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
30936   float32x2x2_t __ret;
30937   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30938   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30939   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
30940 
30941   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
30942   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
30943   return __ret;
30944 }
30945 #endif
30946 
30947 #ifdef __LITTLE_ENDIAN__
vtrn_s32(int32x2_t __p0,int32x2_t __p1)30948 __ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
30949   int32x2x2_t __ret;
30950   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
30951   return __ret;
30952 }
30953 #else
vtrn_s32(int32x2_t __p0,int32x2_t __p1)30954 __ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
30955   int32x2x2_t __ret;
30956   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30957   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30958   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
30959 
30960   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
30961   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
30962   return __ret;
30963 }
30964 #endif
30965 
30966 #ifdef __LITTLE_ENDIAN__
vtrn_s16(int16x4_t __p0,int16x4_t __p1)30967 __ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
30968   int16x4x2_t __ret;
30969   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
30970   return __ret;
30971 }
30972 #else
vtrn_s16(int16x4_t __p0,int16x4_t __p1)30973 __ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
30974   int16x4x2_t __ret;
30975   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30976   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30977   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
30978 
30979   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30980   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30981   return __ret;
30982 }
30983 #endif
30984 
30985 #ifdef __LITTLE_ENDIAN__
vtst_p8(poly8x8_t __p0,poly8x8_t __p1)30986 __ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
30987   uint8x8_t __ret;
30988   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
30989   return __ret;
30990 }
30991 #else
vtst_p8(poly8x8_t __p0,poly8x8_t __p1)30992 __ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
30993   uint8x8_t __ret;
30994   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30995   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30996   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
30997   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30998   return __ret;
30999 }
31000 #endif
31001 
31002 #ifdef __LITTLE_ENDIAN__
vtst_p16(poly16x4_t __p0,poly16x4_t __p1)31003 __ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
31004   uint16x4_t __ret;
31005   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
31006   return __ret;
31007 }
31008 #else
vtst_p16(poly16x4_t __p0,poly16x4_t __p1)31009 __ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
31010   uint16x4_t __ret;
31011   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31012   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31013   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31014   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31015   return __ret;
31016 }
31017 #endif
31018 
31019 #ifdef __LITTLE_ENDIAN__
vtstq_p8(poly8x16_t __p0,poly8x16_t __p1)31020 __ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31021   uint8x16_t __ret;
31022   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
31023   return __ret;
31024 }
31025 #else
vtstq_p8(poly8x16_t __p0,poly8x16_t __p1)31026 __ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31027   uint8x16_t __ret;
31028   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31029   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31030   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31031   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31032   return __ret;
31033 }
31034 #endif
31035 
31036 #ifdef __LITTLE_ENDIAN__
vtstq_p16(poly16x8_t __p0,poly16x8_t __p1)31037 __ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31038   uint16x8_t __ret;
31039   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
31040   return __ret;
31041 }
31042 #else
vtstq_p16(poly16x8_t __p0,poly16x8_t __p1)31043 __ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31044   uint16x8_t __ret;
31045   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31046   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31047   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31048   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31049   return __ret;
31050 }
31051 #endif
31052 
31053 #ifdef __LITTLE_ENDIAN__
vtstq_u8(uint8x16_t __p0,uint8x16_t __p1)31054 __ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31055   uint8x16_t __ret;
31056   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
31057   return __ret;
31058 }
31059 #else
vtstq_u8(uint8x16_t __p0,uint8x16_t __p1)31060 __ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31061   uint8x16_t __ret;
31062   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31063   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31064   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31065   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31066   return __ret;
31067 }
31068 #endif
31069 
31070 #ifdef __LITTLE_ENDIAN__
vtstq_u32(uint32x4_t __p0,uint32x4_t __p1)31071 __ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31072   uint32x4_t __ret;
31073   __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
31074   return __ret;
31075 }
31076 #else
vtstq_u32(uint32x4_t __p0,uint32x4_t __p1)31077 __ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31078   uint32x4_t __ret;
31079   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31080   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31081   __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
31082   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31083   return __ret;
31084 }
31085 #endif
31086 
31087 #ifdef __LITTLE_ENDIAN__
vtstq_u16(uint16x8_t __p0,uint16x8_t __p1)31088 __ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31089   uint16x8_t __ret;
31090   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
31091   return __ret;
31092 }
31093 #else
vtstq_u16(uint16x8_t __p0,uint16x8_t __p1)31094 __ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31095   uint16x8_t __ret;
31096   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31097   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31098   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31099   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31100   return __ret;
31101 }
31102 #endif
31103 
31104 #ifdef __LITTLE_ENDIAN__
vtstq_s8(int8x16_t __p0,int8x16_t __p1)31105 __ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
31106   uint8x16_t __ret;
31107   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
31108   return __ret;
31109 }
31110 #else
vtstq_s8(int8x16_t __p0,int8x16_t __p1)31111 __ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
31112   uint8x16_t __ret;
31113   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31114   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31115   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31116   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31117   return __ret;
31118 }
31119 #endif
31120 
31121 #ifdef __LITTLE_ENDIAN__
vtstq_s32(int32x4_t __p0,int32x4_t __p1)31122 __ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
31123   uint32x4_t __ret;
31124   __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
31125   return __ret;
31126 }
31127 #else
vtstq_s32(int32x4_t __p0,int32x4_t __p1)31128 __ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
31129   uint32x4_t __ret;
31130   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31131   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31132   __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
31133   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31134   return __ret;
31135 }
31136 #endif
31137 
31138 #ifdef __LITTLE_ENDIAN__
vtstq_s16(int16x8_t __p0,int16x8_t __p1)31139 __ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
31140   uint16x8_t __ret;
31141   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
31142   return __ret;
31143 }
31144 #else
vtstq_s16(int16x8_t __p0,int16x8_t __p1)31145 __ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
31146   uint16x8_t __ret;
31147   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31148   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31149   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31150   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31151   return __ret;
31152 }
31153 #endif
31154 
31155 #ifdef __LITTLE_ENDIAN__
vtst_u8(uint8x8_t __p0,uint8x8_t __p1)31156 __ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
31157   uint8x8_t __ret;
31158   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
31159   return __ret;
31160 }
31161 #else
vtst_u8(uint8x8_t __p0,uint8x8_t __p1)31162 __ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
31163   uint8x8_t __ret;
31164   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31165   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31166   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31167   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31168   return __ret;
31169 }
31170 #endif
31171 
31172 #ifdef __LITTLE_ENDIAN__
vtst_u32(uint32x2_t __p0,uint32x2_t __p1)31173 __ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
31174   uint32x2_t __ret;
31175   __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
31176   return __ret;
31177 }
31178 #else
vtst_u32(uint32x2_t __p0,uint32x2_t __p1)31179 __ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
31180   uint32x2_t __ret;
31181   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31182   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31183   __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
31184   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
31185   return __ret;
31186 }
31187 #endif
31188 
31189 #ifdef __LITTLE_ENDIAN__
vtst_u16(uint16x4_t __p0,uint16x4_t __p1)31190 __ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
31191   uint16x4_t __ret;
31192   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
31193   return __ret;
31194 }
31195 #else
vtst_u16(uint16x4_t __p0,uint16x4_t __p1)31196 __ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
31197   uint16x4_t __ret;
31198   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31199   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31200   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31201   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31202   return __ret;
31203 }
31204 #endif
31205 
31206 #ifdef __LITTLE_ENDIAN__
vtst_s8(int8x8_t __p0,int8x8_t __p1)31207 __ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
31208   uint8x8_t __ret;
31209   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
31210   return __ret;
31211 }
31212 #else
vtst_s8(int8x8_t __p0,int8x8_t __p1)31213 __ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
31214   uint8x8_t __ret;
31215   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31216   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31217   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31218   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31219   return __ret;
31220 }
31221 #endif
31222 
31223 #ifdef __LITTLE_ENDIAN__
vtst_s32(int32x2_t __p0,int32x2_t __p1)31224 __ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
31225   uint32x2_t __ret;
31226   __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
31227   return __ret;
31228 }
31229 #else
vtst_s32(int32x2_t __p0,int32x2_t __p1)31230 __ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
31231   uint32x2_t __ret;
31232   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31233   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31234   __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
31235   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
31236   return __ret;
31237 }
31238 #endif
31239 
31240 #ifdef __LITTLE_ENDIAN__
vtst_s16(int16x4_t __p0,int16x4_t __p1)31241 __ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
31242   uint16x4_t __ret;
31243   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
31244   return __ret;
31245 }
31246 #else
vtst_s16(int16x4_t __p0,int16x4_t __p1)31247 __ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
31248   uint16x4_t __ret;
31249   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31250   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31251   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31252   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31253   return __ret;
31254 }
31255 #endif
31256 
31257 #ifdef __LITTLE_ENDIAN__
vuzp_p8(poly8x8_t __p0,poly8x8_t __p1)31258 __ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
31259   poly8x8x2_t __ret;
31260   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
31261   return __ret;
31262 }
31263 #else
vuzp_p8(poly8x8_t __p0,poly8x8_t __p1)31264 __ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
31265   poly8x8x2_t __ret;
31266   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31267   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31268   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
31269 
31270   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31271   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31272   return __ret;
31273 }
31274 #endif
31275 
31276 #ifdef __LITTLE_ENDIAN__
vuzp_p16(poly16x4_t __p0,poly16x4_t __p1)31277 __ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
31278   poly16x4x2_t __ret;
31279   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
31280   return __ret;
31281 }
31282 #else
vuzp_p16(poly16x4_t __p0,poly16x4_t __p1)31283 __ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
31284   poly16x4x2_t __ret;
31285   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31286   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31287   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
31288 
31289   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31290   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31291   return __ret;
31292 }
31293 #endif
31294 
31295 #ifdef __LITTLE_ENDIAN__
vuzpq_p8(poly8x16_t __p0,poly8x16_t __p1)31296 __ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31297   poly8x16x2_t __ret;
31298   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
31299   return __ret;
31300 }
31301 #else
vuzpq_p8(poly8x16_t __p0,poly8x16_t __p1)31302 __ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31303   poly8x16x2_t __ret;
31304   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31305   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31306   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
31307 
31308   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31309   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31310   return __ret;
31311 }
31312 #endif
31313 
31314 #ifdef __LITTLE_ENDIAN__
vuzpq_p16(poly16x8_t __p0,poly16x8_t __p1)31315 __ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31316   poly16x8x2_t __ret;
31317   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
31318   return __ret;
31319 }
31320 #else
vuzpq_p16(poly16x8_t __p0,poly16x8_t __p1)31321 __ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31322   poly16x8x2_t __ret;
31323   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31324   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31325   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
31326 
31327   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31328   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31329   return __ret;
31330 }
31331 #endif
31332 
31333 #ifdef __LITTLE_ENDIAN__
vuzpq_u8(uint8x16_t __p0,uint8x16_t __p1)31334 __ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31335   uint8x16x2_t __ret;
31336   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
31337   return __ret;
31338 }
31339 #else
vuzpq_u8(uint8x16_t __p0,uint8x16_t __p1)31340 __ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31341   uint8x16x2_t __ret;
31342   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31343   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31344   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31345 
31346   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31347   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31348   return __ret;
31349 }
31350 #endif
31351 
31352 #ifdef __LITTLE_ENDIAN__
vuzpq_u32(uint32x4_t __p0,uint32x4_t __p1)31353 __ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31354   uint32x4x2_t __ret;
31355   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
31356   return __ret;
31357 }
31358 #else
vuzpq_u32(uint32x4_t __p0,uint32x4_t __p1)31359 __ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31360   uint32x4x2_t __ret;
31361   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31362   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31363   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
31364 
31365   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31366   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31367   return __ret;
31368 }
31369 #endif
31370 
31371 #ifdef __LITTLE_ENDIAN__
vuzpq_u16(uint16x8_t __p0,uint16x8_t __p1)31372 __ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31373   uint16x8x2_t __ret;
31374   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
31375   return __ret;
31376 }
31377 #else
vuzpq_u16(uint16x8_t __p0,uint16x8_t __p1)31378 __ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31379   uint16x8x2_t __ret;
31380   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31381   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31382   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31383 
31384   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31385   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31386   return __ret;
31387 }
31388 #endif
31389 
31390 #ifdef __LITTLE_ENDIAN__
vuzpq_s8(int8x16_t __p0,int8x16_t __p1)31391 __ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
31392   int8x16x2_t __ret;
31393   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
31394   return __ret;
31395 }
31396 #else
vuzpq_s8(int8x16_t __p0,int8x16_t __p1)31397 __ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
31398   int8x16x2_t __ret;
31399   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31400   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31401   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
31402 
31403   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31404   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31405   return __ret;
31406 }
31407 #endif
31408 
31409 #ifdef __LITTLE_ENDIAN__
vuzpq_f32(float32x4_t __p0,float32x4_t __p1)31410 __ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
31411   float32x4x2_t __ret;
31412   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
31413   return __ret;
31414 }
31415 #else
vuzpq_f32(float32x4_t __p0,float32x4_t __p1)31416 __ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
31417   float32x4x2_t __ret;
31418   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31419   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31420   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
31421 
31422   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31423   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31424   return __ret;
31425 }
31426 #endif
31427 
31428 #ifdef __LITTLE_ENDIAN__
vuzpq_s32(int32x4_t __p0,int32x4_t __p1)31429 __ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
31430   int32x4x2_t __ret;
31431   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
31432   return __ret;
31433 }
31434 #else
vuzpq_s32(int32x4_t __p0,int32x4_t __p1)31435 __ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
31436   int32x4x2_t __ret;
31437   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31438   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31439   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
31440 
31441   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31442   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31443   return __ret;
31444 }
31445 #endif
31446 
31447 #ifdef __LITTLE_ENDIAN__
vuzpq_s16(int16x8_t __p0,int16x8_t __p1)31448 __ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
31449   int16x8x2_t __ret;
31450   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
31451   return __ret;
31452 }
31453 #else
vuzpq_s16(int16x8_t __p0,int16x8_t __p1)31454 __ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
31455   int16x8x2_t __ret;
31456   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31457   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31458   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
31459 
31460   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31461   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31462   return __ret;
31463 }
31464 #endif
31465 
31466 #ifdef __LITTLE_ENDIAN__
vuzp_u8(uint8x8_t __p0,uint8x8_t __p1)31467 __ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
31468   uint8x8x2_t __ret;
31469   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
31470   return __ret;
31471 }
31472 #else
vuzp_u8(uint8x8_t __p0,uint8x8_t __p1)31473 __ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
31474   uint8x8x2_t __ret;
31475   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31476   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31477   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31478 
31479   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31480   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31481   return __ret;
31482 }
31483 #endif
31484 
31485 #ifdef __LITTLE_ENDIAN__
vuzp_u32(uint32x2_t __p0,uint32x2_t __p1)31486 __ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
31487   uint32x2x2_t __ret;
31488   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
31489   return __ret;
31490 }
31491 #else
vuzp_u32(uint32x2_t __p0,uint32x2_t __p1)31492 __ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
31493   uint32x2x2_t __ret;
31494   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31495   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31496   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
31497 
31498   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31499   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31500   return __ret;
31501 }
31502 #endif
31503 
31504 #ifdef __LITTLE_ENDIAN__
vuzp_u16(uint16x4_t __p0,uint16x4_t __p1)31505 __ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
31506   uint16x4x2_t __ret;
31507   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
31508   return __ret;
31509 }
31510 #else
vuzp_u16(uint16x4_t __p0,uint16x4_t __p1)31511 __ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
31512   uint16x4x2_t __ret;
31513   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31514   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31515   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31516 
31517   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31518   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31519   return __ret;
31520 }
31521 #endif
31522 
31523 #ifdef __LITTLE_ENDIAN__
vuzp_s8(int8x8_t __p0,int8x8_t __p1)31524 __ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
31525   int8x8x2_t __ret;
31526   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
31527   return __ret;
31528 }
31529 #else
vuzp_s8(int8x8_t __p0,int8x8_t __p1)31530 __ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
31531   int8x8x2_t __ret;
31532   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31533   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31534   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
31535 
31536   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31537   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31538   return __ret;
31539 }
31540 #endif
31541 
31542 #ifdef __LITTLE_ENDIAN__
vuzp_f32(float32x2_t __p0,float32x2_t __p1)31543 __ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
31544   float32x2x2_t __ret;
31545   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
31546   return __ret;
31547 }
31548 #else
vuzp_f32(float32x2_t __p0,float32x2_t __p1)31549 __ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
31550   float32x2x2_t __ret;
31551   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31552   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31553   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
31554 
31555   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31556   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31557   return __ret;
31558 }
31559 #endif
31560 
31561 #ifdef __LITTLE_ENDIAN__
vuzp_s32(int32x2_t __p0,int32x2_t __p1)31562 __ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
31563   int32x2x2_t __ret;
31564   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
31565   return __ret;
31566 }
31567 #else
vuzp_s32(int32x2_t __p0,int32x2_t __p1)31568 __ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
31569   int32x2x2_t __ret;
31570   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31571   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31572   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
31573 
31574   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31575   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31576   return __ret;
31577 }
31578 #endif
31579 
31580 #ifdef __LITTLE_ENDIAN__
vuzp_s16(int16x4_t __p0,int16x4_t __p1)31581 __ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
31582   int16x4x2_t __ret;
31583   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
31584   return __ret;
31585 }
31586 #else
vuzp_s16(int16x4_t __p0,int16x4_t __p1)31587 __ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
31588   int16x4x2_t __ret;
31589   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31590   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31591   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
31592 
31593   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31594   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31595   return __ret;
31596 }
31597 #endif
31598 
31599 #ifdef __LITTLE_ENDIAN__
vzip_p8(poly8x8_t __p0,poly8x8_t __p1)31600 __ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
31601   poly8x8x2_t __ret;
31602   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
31603   return __ret;
31604 }
31605 #else
vzip_p8(poly8x8_t __p0,poly8x8_t __p1)31606 __ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
31607   poly8x8x2_t __ret;
31608   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31609   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31610   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
31611 
31612   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31613   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31614   return __ret;
31615 }
31616 #endif
31617 
31618 #ifdef __LITTLE_ENDIAN__
vzip_p16(poly16x4_t __p0,poly16x4_t __p1)31619 __ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
31620   poly16x4x2_t __ret;
31621   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
31622   return __ret;
31623 }
31624 #else
vzip_p16(poly16x4_t __p0,poly16x4_t __p1)31625 __ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
31626   poly16x4x2_t __ret;
31627   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31628   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31629   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
31630 
31631   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31632   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31633   return __ret;
31634 }
31635 #endif
31636 
31637 #ifdef __LITTLE_ENDIAN__
vzipq_p8(poly8x16_t __p0,poly8x16_t __p1)31638 __ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31639   poly8x16x2_t __ret;
31640   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
31641   return __ret;
31642 }
31643 #else
vzipq_p8(poly8x16_t __p0,poly8x16_t __p1)31644 __ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31645   poly8x16x2_t __ret;
31646   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31647   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31648   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
31649 
31650   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31651   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31652   return __ret;
31653 }
31654 #endif
31655 
31656 #ifdef __LITTLE_ENDIAN__
vzipq_p16(poly16x8_t __p0,poly16x8_t __p1)31657 __ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31658   poly16x8x2_t __ret;
31659   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
31660   return __ret;
31661 }
31662 #else
vzipq_p16(poly16x8_t __p0,poly16x8_t __p1)31663 __ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31664   poly16x8x2_t __ret;
31665   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31666   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31667   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
31668 
31669   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31670   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31671   return __ret;
31672 }
31673 #endif
31674 
31675 #ifdef __LITTLE_ENDIAN__
vzipq_u8(uint8x16_t __p0,uint8x16_t __p1)31676 __ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31677   uint8x16x2_t __ret;
31678   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
31679   return __ret;
31680 }
31681 #else
vzipq_u8(uint8x16_t __p0,uint8x16_t __p1)31682 __ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31683   uint8x16x2_t __ret;
31684   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31685   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31686   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31687 
31688   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31689   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31690   return __ret;
31691 }
31692 #endif
31693 
31694 #ifdef __LITTLE_ENDIAN__
vzipq_u32(uint32x4_t __p0,uint32x4_t __p1)31695 __ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31696   uint32x4x2_t __ret;
31697   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
31698   return __ret;
31699 }
31700 #else
vzipq_u32(uint32x4_t __p0,uint32x4_t __p1)31701 __ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31702   uint32x4x2_t __ret;
31703   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31704   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31705   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
31706 
31707   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31708   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31709   return __ret;
31710 }
31711 #endif
31712 
31713 #ifdef __LITTLE_ENDIAN__
vzipq_u16(uint16x8_t __p0,uint16x8_t __p1)31714 __ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31715   uint16x8x2_t __ret;
31716   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
31717   return __ret;
31718 }
31719 #else
vzipq_u16(uint16x8_t __p0,uint16x8_t __p1)31720 __ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31721   uint16x8x2_t __ret;
31722   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31723   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31724   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31725 
31726   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31727   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31728   return __ret;
31729 }
31730 #endif
31731 
31732 #ifdef __LITTLE_ENDIAN__
vzipq_s8(int8x16_t __p0,int8x16_t __p1)31733 __ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
31734   int8x16x2_t __ret;
31735   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
31736   return __ret;
31737 }
31738 #else
vzipq_s8(int8x16_t __p0,int8x16_t __p1)31739 __ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
31740   int8x16x2_t __ret;
31741   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31742   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31743   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
31744 
31745   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31746   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31747   return __ret;
31748 }
31749 #endif
31750 
31751 #ifdef __LITTLE_ENDIAN__
vzipq_f32(float32x4_t __p0,float32x4_t __p1)31752 __ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
31753   float32x4x2_t __ret;
31754   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
31755   return __ret;
31756 }
31757 #else
vzipq_f32(float32x4_t __p0,float32x4_t __p1)31758 __ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
31759   float32x4x2_t __ret;
31760   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31761   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31762   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
31763 
31764   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31765   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31766   return __ret;
31767 }
31768 #endif
31769 
31770 #ifdef __LITTLE_ENDIAN__
vzipq_s32(int32x4_t __p0,int32x4_t __p1)31771 __ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
31772   int32x4x2_t __ret;
31773   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
31774   return __ret;
31775 }
31776 #else
vzipq_s32(int32x4_t __p0,int32x4_t __p1)31777 __ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
31778   int32x4x2_t __ret;
31779   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31780   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31781   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
31782 
31783   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31784   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31785   return __ret;
31786 }
31787 #endif
31788 
31789 #ifdef __LITTLE_ENDIAN__
vzipq_s16(int16x8_t __p0,int16x8_t __p1)31790 __ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
31791   int16x8x2_t __ret;
31792   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
31793   return __ret;
31794 }
31795 #else
vzipq_s16(int16x8_t __p0,int16x8_t __p1)31796 __ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
31797   int16x8x2_t __ret;
31798   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31799   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31800   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
31801 
31802   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31803   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31804   return __ret;
31805 }
31806 #endif
31807 
31808 #ifdef __LITTLE_ENDIAN__
vzip_u8(uint8x8_t __p0,uint8x8_t __p1)31809 __ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
31810   uint8x8x2_t __ret;
31811   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
31812   return __ret;
31813 }
31814 #else
vzip_u8(uint8x8_t __p0,uint8x8_t __p1)31815 __ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
31816   uint8x8x2_t __ret;
31817   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31818   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31819   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31820 
31821   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31822   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31823   return __ret;
31824 }
31825 #endif
31826 
31827 #ifdef __LITTLE_ENDIAN__
vzip_u32(uint32x2_t __p0,uint32x2_t __p1)31828 __ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
31829   uint32x2x2_t __ret;
31830   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
31831   return __ret;
31832 }
31833 #else
vzip_u32(uint32x2_t __p0,uint32x2_t __p1)31834 __ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
31835   uint32x2x2_t __ret;
31836   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31837   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31838   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
31839 
31840   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31841   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31842   return __ret;
31843 }
31844 #endif
31845 
31846 #ifdef __LITTLE_ENDIAN__
vzip_u16(uint16x4_t __p0,uint16x4_t __p1)31847 __ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
31848   uint16x4x2_t __ret;
31849   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
31850   return __ret;
31851 }
31852 #else
vzip_u16(uint16x4_t __p0,uint16x4_t __p1)31853 __ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
31854   uint16x4x2_t __ret;
31855   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31856   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31857   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31858 
31859   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31860   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31861   return __ret;
31862 }
31863 #endif
31864 
31865 #ifdef __LITTLE_ENDIAN__
vzip_s8(int8x8_t __p0,int8x8_t __p1)31866 __ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
31867   int8x8x2_t __ret;
31868   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
31869   return __ret;
31870 }
31871 #else
vzip_s8(int8x8_t __p0,int8x8_t __p1)31872 __ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
31873   int8x8x2_t __ret;
31874   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31875   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31876   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
31877 
31878   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31879   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31880   return __ret;
31881 }
31882 #endif
31883 
31884 #ifdef __LITTLE_ENDIAN__
vzip_f32(float32x2_t __p0,float32x2_t __p1)31885 __ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
31886   float32x2x2_t __ret;
31887   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
31888   return __ret;
31889 }
31890 #else
vzip_f32(float32x2_t __p0,float32x2_t __p1)31891 __ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
31892   float32x2x2_t __ret;
31893   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31894   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31895   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
31896 
31897   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31898   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31899   return __ret;
31900 }
31901 #endif
31902 
31903 #ifdef __LITTLE_ENDIAN__
vzip_s32(int32x2_t __p0,int32x2_t __p1)31904 __ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
31905   int32x2x2_t __ret;
31906   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
31907   return __ret;
31908 }
31909 #else
vzip_s32(int32x2_t __p0,int32x2_t __p1)31910 __ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
31911   int32x2x2_t __ret;
31912   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31913   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31914   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
31915 
31916   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31917   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31918   return __ret;
31919 }
31920 #endif
31921 
31922 #ifdef __LITTLE_ENDIAN__
vzip_s16(int16x4_t __p0,int16x4_t __p1)31923 __ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
31924   int16x4x2_t __ret;
31925   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
31926   return __ret;
31927 }
31928 #else
vzip_s16(int16x4_t __p0,int16x4_t __p1)31929 __ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
31930   int16x4x2_t __ret;
31931   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31932   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31933   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
31934 
31935   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31936   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31937   return __ret;
31938 }
31939 #endif
31940 
31941 #ifdef __LITTLE_ENDIAN__
31942 #define splatq_lane_bf16(__p0, __p1) __extension__ ({ \
31943   bfloat16x8_t __ret; \
31944   bfloat16x4_t __s0 = __p0; \
31945   __ret = (bfloat16x8_t) __builtin_neon_splatq_lane_bf16((int8x8_t)__s0, __p1, 11); \
31946   __ret; \
31947 })
31948 #else
31949 #define splatq_lane_bf16(__p0, __p1) __extension__ ({ \
31950   bfloat16x8_t __ret; \
31951   bfloat16x4_t __s0 = __p0; \
31952   bfloat16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
31953   __ret = (bfloat16x8_t) __builtin_neon_splatq_lane_bf16((int8x8_t)__rev0, __p1, 11); \
31954   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
31955   __ret; \
31956 })
31957 #define __noswap_splatq_lane_bf16(__p0, __p1) __extension__ ({ \
31958   bfloat16x8_t __ret; \
31959   bfloat16x4_t __s0 = __p0; \
31960   __ret = (bfloat16x8_t) __builtin_neon_splatq_lane_bf16((int8x8_t)__s0, __p1, 11); \
31961   __ret; \
31962 })
31963 #endif
31964 
31965 #ifdef __LITTLE_ENDIAN__
31966 #define splat_lane_bf16(__p0, __p1) __extension__ ({ \
31967   bfloat16x4_t __ret; \
31968   bfloat16x4_t __s0 = __p0; \
31969   __ret = (bfloat16x4_t) __builtin_neon_splat_lane_bf16((int8x8_t)__s0, __p1, 11); \
31970   __ret; \
31971 })
31972 #else
31973 #define splat_lane_bf16(__p0, __p1) __extension__ ({ \
31974   bfloat16x4_t __ret; \
31975   bfloat16x4_t __s0 = __p0; \
31976   bfloat16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
31977   __ret = (bfloat16x4_t) __builtin_neon_splat_lane_bf16((int8x8_t)__rev0, __p1, 11); \
31978   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
31979   __ret; \
31980 })
31981 #define __noswap_splat_lane_bf16(__p0, __p1) __extension__ ({ \
31982   bfloat16x4_t __ret; \
31983   bfloat16x4_t __s0 = __p0; \
31984   __ret = (bfloat16x4_t) __builtin_neon_splat_lane_bf16((int8x8_t)__s0, __p1, 11); \
31985   __ret; \
31986 })
31987 #endif
31988 
31989 #ifdef __LITTLE_ENDIAN__
31990 #define splatq_laneq_bf16(__p0, __p1) __extension__ ({ \
31991   bfloat16x8_t __ret; \
31992   bfloat16x8_t __s0 = __p0; \
31993   __ret = (bfloat16x8_t) __builtin_neon_splatq_laneq_bf16((int8x16_t)__s0, __p1, 43); \
31994   __ret; \
31995 })
31996 #else
31997 #define splatq_laneq_bf16(__p0, __p1) __extension__ ({ \
31998   bfloat16x8_t __ret; \
31999   bfloat16x8_t __s0 = __p0; \
32000   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
32001   __ret = (bfloat16x8_t) __builtin_neon_splatq_laneq_bf16((int8x16_t)__rev0, __p1, 43); \
32002   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
32003   __ret; \
32004 })
32005 #define __noswap_splatq_laneq_bf16(__p0, __p1) __extension__ ({ \
32006   bfloat16x8_t __ret; \
32007   bfloat16x8_t __s0 = __p0; \
32008   __ret = (bfloat16x8_t) __builtin_neon_splatq_laneq_bf16((int8x16_t)__s0, __p1, 43); \
32009   __ret; \
32010 })
32011 #endif
32012 
32013 #ifdef __LITTLE_ENDIAN__
32014 #define splat_laneq_bf16(__p0, __p1) __extension__ ({ \
32015   bfloat16x4_t __ret; \
32016   bfloat16x8_t __s0 = __p0; \
32017   __ret = (bfloat16x4_t) __builtin_neon_splat_laneq_bf16((int8x16_t)__s0, __p1, 43); \
32018   __ret; \
32019 })
32020 #else
32021 #define splat_laneq_bf16(__p0, __p1) __extension__ ({ \
32022   bfloat16x4_t __ret; \
32023   bfloat16x8_t __s0 = __p0; \
32024   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
32025   __ret = (bfloat16x4_t) __builtin_neon_splat_laneq_bf16((int8x16_t)__rev0, __p1, 43); \
32026   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
32027   __ret; \
32028 })
32029 #define __noswap_splat_laneq_bf16(__p0, __p1) __extension__ ({ \
32030   bfloat16x4_t __ret; \
32031   bfloat16x8_t __s0 = __p0; \
32032   __ret = (bfloat16x4_t) __builtin_neon_splat_laneq_bf16((int8x16_t)__s0, __p1, 43); \
32033   __ret; \
32034 })
32035 #endif
32036 
32037 #ifdef __LITTLE_ENDIAN__
vbfdotq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32038 __ai __attribute__((target("bf16"))) float32x4_t vbfdotq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32039   float32x4_t __ret;
32040   __ret = (float32x4_t) __builtin_neon_vbfdotq_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
32041   return __ret;
32042 }
32043 #else
vbfdotq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32044 __ai __attribute__((target("bf16"))) float32x4_t vbfdotq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32045   float32x4_t __ret;
32046   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
32047   bfloat16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
32048   bfloat16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
32049   __ret = (float32x4_t) __builtin_neon_vbfdotq_f32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
32050   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
32051   return __ret;
32052 }
__noswap_vbfdotq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32053 __ai __attribute__((target("bf16"))) float32x4_t __noswap_vbfdotq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32054   float32x4_t __ret;
32055   __ret = (float32x4_t) __builtin_neon_vbfdotq_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
32056   return __ret;
32057 }
32058 #endif
32059 
32060 #ifdef __LITTLE_ENDIAN__
vbfdot_f32(float32x2_t __p0,bfloat16x4_t __p1,bfloat16x4_t __p2)32061 __ai __attribute__((target("bf16"))) float32x2_t vbfdot_f32(float32x2_t __p0, bfloat16x4_t __p1, bfloat16x4_t __p2) {
32062   float32x2_t __ret;
32063   __ret = (float32x2_t) __builtin_neon_vbfdot_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
32064   return __ret;
32065 }
32066 #else
vbfdot_f32(float32x2_t __p0,bfloat16x4_t __p1,bfloat16x4_t __p2)32067 __ai __attribute__((target("bf16"))) float32x2_t vbfdot_f32(float32x2_t __p0, bfloat16x4_t __p1, bfloat16x4_t __p2) {
32068   float32x2_t __ret;
32069   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
32070   bfloat16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
32071   bfloat16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
32072   __ret = (float32x2_t) __builtin_neon_vbfdot_f32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
32073   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
32074   return __ret;
32075 }
__noswap_vbfdot_f32(float32x2_t __p0,bfloat16x4_t __p1,bfloat16x4_t __p2)32076 __ai __attribute__((target("bf16"))) float32x2_t __noswap_vbfdot_f32(float32x2_t __p0, bfloat16x4_t __p1, bfloat16x4_t __p2) {
32077   float32x2_t __ret;
32078   __ret = (float32x2_t) __builtin_neon_vbfdot_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
32079   return __ret;
32080 }
32081 #endif
32082 
32083 #ifdef __LITTLE_ENDIAN__
32084 #define vbfdotq_lane_f32(__p0_126, __p1_126, __p2_126, __p3_126) __extension__ ({ \
32085   float32x4_t __ret_126; \
32086   float32x4_t __s0_126 = __p0_126; \
32087   bfloat16x8_t __s1_126 = __p1_126; \
32088   bfloat16x4_t __s2_126 = __p2_126; \
32089 bfloat16x4_t __reint_126 = __s2_126; \
32090 float32x4_t __reint1_126 = splatq_lane_f32(*(float32x2_t *) &__reint_126, __p3_126); \
32091   __ret_126 = vbfdotq_f32(__s0_126, __s1_126, *(bfloat16x8_t *) &__reint1_126); \
32092   __ret_126; \
32093 })
32094 #else
32095 #define vbfdotq_lane_f32(__p0_127, __p1_127, __p2_127, __p3_127) __extension__ ({ \
32096   float32x4_t __ret_127; \
32097   float32x4_t __s0_127 = __p0_127; \
32098   bfloat16x8_t __s1_127 = __p1_127; \
32099   bfloat16x4_t __s2_127 = __p2_127; \
32100   float32x4_t __rev0_127;  __rev0_127 = __builtin_shufflevector(__s0_127, __s0_127, 3, 2, 1, 0); \
32101   bfloat16x8_t __rev1_127;  __rev1_127 = __builtin_shufflevector(__s1_127, __s1_127, 7, 6, 5, 4, 3, 2, 1, 0); \
32102   bfloat16x4_t __rev2_127;  __rev2_127 = __builtin_shufflevector(__s2_127, __s2_127, 3, 2, 1, 0); \
32103 bfloat16x4_t __reint_127 = __rev2_127; \
32104 float32x4_t __reint1_127 = __noswap_splatq_lane_f32(*(float32x2_t *) &__reint_127, __p3_127); \
32105   __ret_127 = __noswap_vbfdotq_f32(__rev0_127, __rev1_127, *(bfloat16x8_t *) &__reint1_127); \
32106   __ret_127 = __builtin_shufflevector(__ret_127, __ret_127, 3, 2, 1, 0); \
32107   __ret_127; \
32108 })
32109 #endif
32110 
32111 #ifdef __LITTLE_ENDIAN__
32112 #define vbfdot_lane_f32(__p0_128, __p1_128, __p2_128, __p3_128) __extension__ ({ \
32113   float32x2_t __ret_128; \
32114   float32x2_t __s0_128 = __p0_128; \
32115   bfloat16x4_t __s1_128 = __p1_128; \
32116   bfloat16x4_t __s2_128 = __p2_128; \
32117 bfloat16x4_t __reint_128 = __s2_128; \
32118 float32x2_t __reint1_128 = splat_lane_f32(*(float32x2_t *) &__reint_128, __p3_128); \
32119   __ret_128 = vbfdot_f32(__s0_128, __s1_128, *(bfloat16x4_t *) &__reint1_128); \
32120   __ret_128; \
32121 })
32122 #else
32123 #define vbfdot_lane_f32(__p0_129, __p1_129, __p2_129, __p3_129) __extension__ ({ \
32124   float32x2_t __ret_129; \
32125   float32x2_t __s0_129 = __p0_129; \
32126   bfloat16x4_t __s1_129 = __p1_129; \
32127   bfloat16x4_t __s2_129 = __p2_129; \
32128   float32x2_t __rev0_129;  __rev0_129 = __builtin_shufflevector(__s0_129, __s0_129, 1, 0); \
32129   bfloat16x4_t __rev1_129;  __rev1_129 = __builtin_shufflevector(__s1_129, __s1_129, 3, 2, 1, 0); \
32130   bfloat16x4_t __rev2_129;  __rev2_129 = __builtin_shufflevector(__s2_129, __s2_129, 3, 2, 1, 0); \
32131 bfloat16x4_t __reint_129 = __rev2_129; \
32132 float32x2_t __reint1_129 = __noswap_splat_lane_f32(*(float32x2_t *) &__reint_129, __p3_129); \
32133   __ret_129 = __noswap_vbfdot_f32(__rev0_129, __rev1_129, *(bfloat16x4_t *) &__reint1_129); \
32134   __ret_129 = __builtin_shufflevector(__ret_129, __ret_129, 1, 0); \
32135   __ret_129; \
32136 })
32137 #endif
32138 
32139 #ifdef __LITTLE_ENDIAN__
32140 #define vbfdotq_laneq_f32(__p0_130, __p1_130, __p2_130, __p3_130) __extension__ ({ \
32141   float32x4_t __ret_130; \
32142   float32x4_t __s0_130 = __p0_130; \
32143   bfloat16x8_t __s1_130 = __p1_130; \
32144   bfloat16x8_t __s2_130 = __p2_130; \
32145 bfloat16x8_t __reint_130 = __s2_130; \
32146 float32x4_t __reint1_130 = splatq_laneq_f32(*(float32x4_t *) &__reint_130, __p3_130); \
32147   __ret_130 = vbfdotq_f32(__s0_130, __s1_130, *(bfloat16x8_t *) &__reint1_130); \
32148   __ret_130; \
32149 })
32150 #else
32151 #define vbfdotq_laneq_f32(__p0_131, __p1_131, __p2_131, __p3_131) __extension__ ({ \
32152   float32x4_t __ret_131; \
32153   float32x4_t __s0_131 = __p0_131; \
32154   bfloat16x8_t __s1_131 = __p1_131; \
32155   bfloat16x8_t __s2_131 = __p2_131; \
32156   float32x4_t __rev0_131;  __rev0_131 = __builtin_shufflevector(__s0_131, __s0_131, 3, 2, 1, 0); \
32157   bfloat16x8_t __rev1_131;  __rev1_131 = __builtin_shufflevector(__s1_131, __s1_131, 7, 6, 5, 4, 3, 2, 1, 0); \
32158   bfloat16x8_t __rev2_131;  __rev2_131 = __builtin_shufflevector(__s2_131, __s2_131, 7, 6, 5, 4, 3, 2, 1, 0); \
32159 bfloat16x8_t __reint_131 = __rev2_131; \
32160 float32x4_t __reint1_131 = __noswap_splatq_laneq_f32(*(float32x4_t *) &__reint_131, __p3_131); \
32161   __ret_131 = __noswap_vbfdotq_f32(__rev0_131, __rev1_131, *(bfloat16x8_t *) &__reint1_131); \
32162   __ret_131 = __builtin_shufflevector(__ret_131, __ret_131, 3, 2, 1, 0); \
32163   __ret_131; \
32164 })
32165 #endif
32166 
32167 #ifdef __LITTLE_ENDIAN__
32168 #define vbfdot_laneq_f32(__p0_132, __p1_132, __p2_132, __p3_132) __extension__ ({ \
32169   float32x2_t __ret_132; \
32170   float32x2_t __s0_132 = __p0_132; \
32171   bfloat16x4_t __s1_132 = __p1_132; \
32172   bfloat16x8_t __s2_132 = __p2_132; \
32173 bfloat16x8_t __reint_132 = __s2_132; \
32174 float32x2_t __reint1_132 = splat_laneq_f32(*(float32x4_t *) &__reint_132, __p3_132); \
32175   __ret_132 = vbfdot_f32(__s0_132, __s1_132, *(bfloat16x4_t *) &__reint1_132); \
32176   __ret_132; \
32177 })
32178 #else
32179 #define vbfdot_laneq_f32(__p0_133, __p1_133, __p2_133, __p3_133) __extension__ ({ \
32180   float32x2_t __ret_133; \
32181   float32x2_t __s0_133 = __p0_133; \
32182   bfloat16x4_t __s1_133 = __p1_133; \
32183   bfloat16x8_t __s2_133 = __p2_133; \
32184   float32x2_t __rev0_133;  __rev0_133 = __builtin_shufflevector(__s0_133, __s0_133, 1, 0); \
32185   bfloat16x4_t __rev1_133;  __rev1_133 = __builtin_shufflevector(__s1_133, __s1_133, 3, 2, 1, 0); \
32186   bfloat16x8_t __rev2_133;  __rev2_133 = __builtin_shufflevector(__s2_133, __s2_133, 7, 6, 5, 4, 3, 2, 1, 0); \
32187 bfloat16x8_t __reint_133 = __rev2_133; \
32188 float32x2_t __reint1_133 = __noswap_splat_laneq_f32(*(float32x4_t *) &__reint_133, __p3_133); \
32189   __ret_133 = __noswap_vbfdot_f32(__rev0_133, __rev1_133, *(bfloat16x4_t *) &__reint1_133); \
32190   __ret_133 = __builtin_shufflevector(__ret_133, __ret_133, 1, 0); \
32191   __ret_133; \
32192 })
32193 #endif
32194 
32195 #ifdef __LITTLE_ENDIAN__
vbfmlalbq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32196 __ai __attribute__((target("bf16"))) float32x4_t vbfmlalbq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32197   float32x4_t __ret;
32198   __ret = (float32x4_t) __builtin_neon_vbfmlalbq_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
32199   return __ret;
32200 }
32201 #else
vbfmlalbq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32202 __ai __attribute__((target("bf16"))) float32x4_t vbfmlalbq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32203   float32x4_t __ret;
32204   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
32205   bfloat16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
32206   bfloat16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
32207   __ret = (float32x4_t) __builtin_neon_vbfmlalbq_f32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
32208   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
32209   return __ret;
32210 }
__noswap_vbfmlalbq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32211 __ai __attribute__((target("bf16"))) float32x4_t __noswap_vbfmlalbq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32212   float32x4_t __ret;
32213   __ret = (float32x4_t) __builtin_neon_vbfmlalbq_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
32214   return __ret;
32215 }
32216 #endif
32217 
32218 #ifdef __LITTLE_ENDIAN__
vbfmlaltq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32219 __ai __attribute__((target("bf16"))) float32x4_t vbfmlaltq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32220   float32x4_t __ret;
32221   __ret = (float32x4_t) __builtin_neon_vbfmlaltq_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
32222   return __ret;
32223 }
32224 #else
vbfmlaltq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32225 __ai __attribute__((target("bf16"))) float32x4_t vbfmlaltq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32226   float32x4_t __ret;
32227   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
32228   bfloat16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
32229   bfloat16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
32230   __ret = (float32x4_t) __builtin_neon_vbfmlaltq_f32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
32231   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
32232   return __ret;
32233 }
__noswap_vbfmlaltq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32234 __ai __attribute__((target("bf16"))) float32x4_t __noswap_vbfmlaltq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32235   float32x4_t __ret;
32236   __ret = (float32x4_t) __builtin_neon_vbfmlaltq_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
32237   return __ret;
32238 }
32239 #endif
32240 
32241 #ifdef __LITTLE_ENDIAN__
vbfmmlaq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32242 __ai __attribute__((target("bf16"))) float32x4_t vbfmmlaq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32243   float32x4_t __ret;
32244   __ret = (float32x4_t) __builtin_neon_vbfmmlaq_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
32245   return __ret;
32246 }
32247 #else
vbfmmlaq_f32(float32x4_t __p0,bfloat16x8_t __p1,bfloat16x8_t __p2)32248 __ai __attribute__((target("bf16"))) float32x4_t vbfmmlaq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
32249   float32x4_t __ret;
32250   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
32251   bfloat16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
32252   bfloat16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
32253   __ret = (float32x4_t) __builtin_neon_vbfmmlaq_f32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
32254   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
32255   return __ret;
32256 }
32257 #endif
32258 
32259 #ifdef __LITTLE_ENDIAN__
vcombine_bf16(bfloat16x4_t __p0,bfloat16x4_t __p1)32260 __ai __attribute__((target("bf16"))) bfloat16x8_t vcombine_bf16(bfloat16x4_t __p0, bfloat16x4_t __p1) {
32261   bfloat16x8_t __ret;
32262   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
32263   return __ret;
32264 }
32265 #else
vcombine_bf16(bfloat16x4_t __p0,bfloat16x4_t __p1)32266 __ai __attribute__((target("bf16"))) bfloat16x8_t vcombine_bf16(bfloat16x4_t __p0, bfloat16x4_t __p1) {
32267   bfloat16x8_t __ret;
32268   bfloat16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
32269   bfloat16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
32270   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
32271   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
32272   return __ret;
32273 }
__noswap_vcombine_bf16(bfloat16x4_t __p0,bfloat16x4_t __p1)32274 __ai __attribute__((target("bf16"))) bfloat16x8_t __noswap_vcombine_bf16(bfloat16x4_t __p0, bfloat16x4_t __p1) {
32275   bfloat16x8_t __ret;
32276   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
32277   return __ret;
32278 }
32279 #endif
32280 
32281 #define vcreate_bf16(__p0) __extension__ ({ \
32282   bfloat16x4_t __ret; \
32283   uint64_t __promote = __p0; \
32284   __ret = (bfloat16x4_t)(__promote); \
32285   __ret; \
32286 })
32287 #ifdef __LITTLE_ENDIAN__
vcvt_f32_bf16(bfloat16x4_t __p0_134)32288 __ai __attribute__((target("bf16"))) float32x4_t vcvt_f32_bf16(bfloat16x4_t __p0_134) {
32289   float32x4_t __ret_134;
32290 bfloat16x4_t __reint_134 = __p0_134;
32291 int32x4_t __reint1_134 = vshll_n_s16(*(int16x4_t *) &__reint_134, 16);
32292   __ret_134 = *(float32x4_t *) &__reint1_134;
32293   return __ret_134;
32294 }
32295 #else
vcvt_f32_bf16(bfloat16x4_t __p0_135)32296 __ai __attribute__((target("bf16"))) float32x4_t vcvt_f32_bf16(bfloat16x4_t __p0_135) {
32297   float32x4_t __ret_135;
32298   bfloat16x4_t __rev0_135;  __rev0_135 = __builtin_shufflevector(__p0_135, __p0_135, 3, 2, 1, 0);
32299 bfloat16x4_t __reint_135 = __rev0_135;
32300 int32x4_t __reint1_135 = __noswap_vshll_n_s16(*(int16x4_t *) &__reint_135, 16);
32301   __ret_135 = *(float32x4_t *) &__reint1_135;
32302   __ret_135 = __builtin_shufflevector(__ret_135, __ret_135, 3, 2, 1, 0);
32303   return __ret_135;
32304 }
__noswap_vcvt_f32_bf16(bfloat16x4_t __p0_136)32305 __ai __attribute__((target("bf16"))) float32x4_t __noswap_vcvt_f32_bf16(bfloat16x4_t __p0_136) {
32306   float32x4_t __ret_136;
32307 bfloat16x4_t __reint_136 = __p0_136;
32308 int32x4_t __reint1_136 = __noswap_vshll_n_s16(*(int16x4_t *) &__reint_136, 16);
32309   __ret_136 = *(float32x4_t *) &__reint1_136;
32310   return __ret_136;
32311 }
32312 #endif
32313 
vcvtah_f32_bf16(bfloat16_t __p0)32314 __ai __attribute__((target("bf16"))) float32_t vcvtah_f32_bf16(bfloat16_t __p0) {
32315   float32_t __ret;
32316 bfloat16_t __reint = __p0;
32317 int32_t __reint1 = *(int32_t *) &__reint << 16;
32318   __ret = *(float32_t *) &__reint1;
32319   return __ret;
32320 }
vcvth_bf16_f32(float32_t __p0)32321 __ai __attribute__((target("bf16"))) bfloat16_t vcvth_bf16_f32(float32_t __p0) {
32322   bfloat16_t __ret;
32323   __ret = (bfloat16_t) __builtin_neon_vcvth_bf16_f32(__p0);
32324   return __ret;
32325 }
32326 #ifdef __LITTLE_ENDIAN__
32327 #define vduph_lane_bf16(__p0, __p1) __extension__ ({ \
32328   bfloat16_t __ret; \
32329   bfloat16x4_t __s0 = __p0; \
32330   __ret = (bfloat16_t) __builtin_neon_vduph_lane_bf16((bfloat16x4_t)__s0, __p1); \
32331   __ret; \
32332 })
32333 #else
32334 #define vduph_lane_bf16(__p0, __p1) __extension__ ({ \
32335   bfloat16_t __ret; \
32336   bfloat16x4_t __s0 = __p0; \
32337   bfloat16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
32338   __ret = (bfloat16_t) __builtin_neon_vduph_lane_bf16((bfloat16x4_t)__rev0, __p1); \
32339   __ret; \
32340 })
32341 #endif
32342 
32343 #ifdef __LITTLE_ENDIAN__
32344 #define vdupq_lane_bf16(__p0_137, __p1_137) __extension__ ({ \
32345   bfloat16x8_t __ret_137; \
32346   bfloat16x4_t __s0_137 = __p0_137; \
32347   __ret_137 = splatq_lane_bf16(__s0_137, __p1_137); \
32348   __ret_137; \
32349 })
32350 #else
32351 #define vdupq_lane_bf16(__p0_138, __p1_138) __extension__ ({ \
32352   bfloat16x8_t __ret_138; \
32353   bfloat16x4_t __s0_138 = __p0_138; \
32354   bfloat16x4_t __rev0_138;  __rev0_138 = __builtin_shufflevector(__s0_138, __s0_138, 3, 2, 1, 0); \
32355   __ret_138 = __noswap_splatq_lane_bf16(__rev0_138, __p1_138); \
32356   __ret_138 = __builtin_shufflevector(__ret_138, __ret_138, 7, 6, 5, 4, 3, 2, 1, 0); \
32357   __ret_138; \
32358 })
32359 #endif
32360 
32361 #ifdef __LITTLE_ENDIAN__
32362 #define vdup_lane_bf16(__p0_139, __p1_139) __extension__ ({ \
32363   bfloat16x4_t __ret_139; \
32364   bfloat16x4_t __s0_139 = __p0_139; \
32365   __ret_139 = splat_lane_bf16(__s0_139, __p1_139); \
32366   __ret_139; \
32367 })
32368 #else
32369 #define vdup_lane_bf16(__p0_140, __p1_140) __extension__ ({ \
32370   bfloat16x4_t __ret_140; \
32371   bfloat16x4_t __s0_140 = __p0_140; \
32372   bfloat16x4_t __rev0_140;  __rev0_140 = __builtin_shufflevector(__s0_140, __s0_140, 3, 2, 1, 0); \
32373   __ret_140 = __noswap_splat_lane_bf16(__rev0_140, __p1_140); \
32374   __ret_140 = __builtin_shufflevector(__ret_140, __ret_140, 3, 2, 1, 0); \
32375   __ret_140; \
32376 })
32377 #endif
32378 
32379 #ifdef __LITTLE_ENDIAN__
32380 #define vduph_laneq_bf16(__p0, __p1) __extension__ ({ \
32381   bfloat16_t __ret; \
32382   bfloat16x8_t __s0 = __p0; \
32383   __ret = (bfloat16_t) __builtin_neon_vduph_laneq_bf16((bfloat16x8_t)__s0, __p1); \
32384   __ret; \
32385 })
32386 #else
32387 #define vduph_laneq_bf16(__p0, __p1) __extension__ ({ \
32388   bfloat16_t __ret; \
32389   bfloat16x8_t __s0 = __p0; \
32390   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
32391   __ret = (bfloat16_t) __builtin_neon_vduph_laneq_bf16((bfloat16x8_t)__rev0, __p1); \
32392   __ret; \
32393 })
32394 #endif
32395 
32396 #ifdef __LITTLE_ENDIAN__
32397 #define vdupq_laneq_bf16(__p0_141, __p1_141) __extension__ ({ \
32398   bfloat16x8_t __ret_141; \
32399   bfloat16x8_t __s0_141 = __p0_141; \
32400   __ret_141 = splatq_laneq_bf16(__s0_141, __p1_141); \
32401   __ret_141; \
32402 })
32403 #else
32404 #define vdupq_laneq_bf16(__p0_142, __p1_142) __extension__ ({ \
32405   bfloat16x8_t __ret_142; \
32406   bfloat16x8_t __s0_142 = __p0_142; \
32407   bfloat16x8_t __rev0_142;  __rev0_142 = __builtin_shufflevector(__s0_142, __s0_142, 7, 6, 5, 4, 3, 2, 1, 0); \
32408   __ret_142 = __noswap_splatq_laneq_bf16(__rev0_142, __p1_142); \
32409   __ret_142 = __builtin_shufflevector(__ret_142, __ret_142, 7, 6, 5, 4, 3, 2, 1, 0); \
32410   __ret_142; \
32411 })
32412 #endif
32413 
32414 #ifdef __LITTLE_ENDIAN__
32415 #define vdup_laneq_bf16(__p0_143, __p1_143) __extension__ ({ \
32416   bfloat16x4_t __ret_143; \
32417   bfloat16x8_t __s0_143 = __p0_143; \
32418   __ret_143 = splat_laneq_bf16(__s0_143, __p1_143); \
32419   __ret_143; \
32420 })
32421 #else
32422 #define vdup_laneq_bf16(__p0_144, __p1_144) __extension__ ({ \
32423   bfloat16x4_t __ret_144; \
32424   bfloat16x8_t __s0_144 = __p0_144; \
32425   bfloat16x8_t __rev0_144;  __rev0_144 = __builtin_shufflevector(__s0_144, __s0_144, 7, 6, 5, 4, 3, 2, 1, 0); \
32426   __ret_144 = __noswap_splat_laneq_bf16(__rev0_144, __p1_144); \
32427   __ret_144 = __builtin_shufflevector(__ret_144, __ret_144, 3, 2, 1, 0); \
32428   __ret_144; \
32429 })
32430 #endif
32431 
32432 #ifdef __LITTLE_ENDIAN__
vdupq_n_bf16(bfloat16_t __p0)32433 __ai __attribute__((target("bf16"))) bfloat16x8_t vdupq_n_bf16(bfloat16_t __p0) {
32434   bfloat16x8_t __ret;
32435   __ret = (bfloat16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
32436   return __ret;
32437 }
32438 #else
vdupq_n_bf16(bfloat16_t __p0)32439 __ai __attribute__((target("bf16"))) bfloat16x8_t vdupq_n_bf16(bfloat16_t __p0) {
32440   bfloat16x8_t __ret;
32441   __ret = (bfloat16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
32442   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
32443   return __ret;
32444 }
32445 #endif
32446 
32447 #ifdef __LITTLE_ENDIAN__
vdup_n_bf16(bfloat16_t __p0)32448 __ai __attribute__((target("bf16"))) bfloat16x4_t vdup_n_bf16(bfloat16_t __p0) {
32449   bfloat16x4_t __ret;
32450   __ret = (bfloat16x4_t) {__p0, __p0, __p0, __p0};
32451   return __ret;
32452 }
32453 #else
vdup_n_bf16(bfloat16_t __p0)32454 __ai __attribute__((target("bf16"))) bfloat16x4_t vdup_n_bf16(bfloat16_t __p0) {
32455   bfloat16x4_t __ret;
32456   __ret = (bfloat16x4_t) {__p0, __p0, __p0, __p0};
32457   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
32458   return __ret;
32459 }
32460 #endif
32461 
32462 #ifdef __LITTLE_ENDIAN__
vget_high_bf16(bfloat16x8_t __p0)32463 __ai __attribute__((target("bf16"))) bfloat16x4_t vget_high_bf16(bfloat16x8_t __p0) {
32464   bfloat16x4_t __ret;
32465   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
32466   return __ret;
32467 }
32468 #else
vget_high_bf16(bfloat16x8_t __p0)32469 __ai __attribute__((target("bf16"))) bfloat16x4_t vget_high_bf16(bfloat16x8_t __p0) {
32470   bfloat16x4_t __ret;
32471   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
32472   __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
32473   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
32474   return __ret;
32475 }
__noswap_vget_high_bf16(bfloat16x8_t __p0)32476 __ai __attribute__((target("bf16"))) bfloat16x4_t __noswap_vget_high_bf16(bfloat16x8_t __p0) {
32477   bfloat16x4_t __ret;
32478   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
32479   return __ret;
32480 }
32481 #endif
32482 
32483 #ifdef __LITTLE_ENDIAN__
32484 #define vgetq_lane_bf16(__p0, __p1) __extension__ ({ \
32485   bfloat16_t __ret; \
32486   bfloat16x8_t __s0 = __p0; \
32487   __ret = (bfloat16_t) __builtin_neon_vgetq_lane_bf16((bfloat16x8_t)__s0, __p1); \
32488   __ret; \
32489 })
32490 #else
32491 #define vgetq_lane_bf16(__p0, __p1) __extension__ ({ \
32492   bfloat16_t __ret; \
32493   bfloat16x8_t __s0 = __p0; \
32494   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
32495   __ret = (bfloat16_t) __builtin_neon_vgetq_lane_bf16((bfloat16x8_t)__rev0, __p1); \
32496   __ret; \
32497 })
32498 #define __noswap_vgetq_lane_bf16(__p0, __p1) __extension__ ({ \
32499   bfloat16_t __ret; \
32500   bfloat16x8_t __s0 = __p0; \
32501   __ret = (bfloat16_t) __builtin_neon_vgetq_lane_bf16((bfloat16x8_t)__s0, __p1); \
32502   __ret; \
32503 })
32504 #endif
32505 
32506 #ifdef __LITTLE_ENDIAN__
32507 #define vget_lane_bf16(__p0, __p1) __extension__ ({ \
32508   bfloat16_t __ret; \
32509   bfloat16x4_t __s0 = __p0; \
32510   __ret = (bfloat16_t) __builtin_neon_vget_lane_bf16((bfloat16x4_t)__s0, __p1); \
32511   __ret; \
32512 })
32513 #else
32514 #define vget_lane_bf16(__p0, __p1) __extension__ ({ \
32515   bfloat16_t __ret; \
32516   bfloat16x4_t __s0 = __p0; \
32517   bfloat16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
32518   __ret = (bfloat16_t) __builtin_neon_vget_lane_bf16((bfloat16x4_t)__rev0, __p1); \
32519   __ret; \
32520 })
32521 #define __noswap_vget_lane_bf16(__p0, __p1) __extension__ ({ \
32522   bfloat16_t __ret; \
32523   bfloat16x4_t __s0 = __p0; \
32524   __ret = (bfloat16_t) __builtin_neon_vget_lane_bf16((bfloat16x4_t)__s0, __p1); \
32525   __ret; \
32526 })
32527 #endif
32528 
32529 #ifdef __LITTLE_ENDIAN__
vget_low_bf16(bfloat16x8_t __p0)32530 __ai __attribute__((target("bf16"))) bfloat16x4_t vget_low_bf16(bfloat16x8_t __p0) {
32531   bfloat16x4_t __ret;
32532   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
32533   return __ret;
32534 }
32535 #else
vget_low_bf16(bfloat16x8_t __p0)32536 __ai __attribute__((target("bf16"))) bfloat16x4_t vget_low_bf16(bfloat16x8_t __p0) {
32537   bfloat16x4_t __ret;
32538   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
32539   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
32540   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
32541   return __ret;
32542 }
__noswap_vget_low_bf16(bfloat16x8_t __p0)32543 __ai __attribute__((target("bf16"))) bfloat16x4_t __noswap_vget_low_bf16(bfloat16x8_t __p0) {
32544   bfloat16x4_t __ret;
32545   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
32546   return __ret;
32547 }
32548 #endif
32549 
32550 #ifdef __LITTLE_ENDIAN__
32551 #define vld1q_bf16(__p0) __extension__ ({ \
32552   bfloat16x8_t __ret; \
32553   __ret = (bfloat16x8_t) __builtin_neon_vld1q_bf16(__p0, 43); \
32554   __ret; \
32555 })
32556 #else
32557 #define vld1q_bf16(__p0) __extension__ ({ \
32558   bfloat16x8_t __ret; \
32559   __ret = (bfloat16x8_t) __builtin_neon_vld1q_bf16(__p0, 43); \
32560   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
32561   __ret; \
32562 })
32563 #endif
32564 
32565 #ifdef __LITTLE_ENDIAN__
32566 #define vld1_bf16(__p0) __extension__ ({ \
32567   bfloat16x4_t __ret; \
32568   __ret = (bfloat16x4_t) __builtin_neon_vld1_bf16(__p0, 11); \
32569   __ret; \
32570 })
32571 #else
32572 #define vld1_bf16(__p0) __extension__ ({ \
32573   bfloat16x4_t __ret; \
32574   __ret = (bfloat16x4_t) __builtin_neon_vld1_bf16(__p0, 11); \
32575   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
32576   __ret; \
32577 })
32578 #endif
32579 
32580 #ifdef __LITTLE_ENDIAN__
32581 #define vld1q_dup_bf16(__p0) __extension__ ({ \
32582   bfloat16x8_t __ret; \
32583   __ret = (bfloat16x8_t) __builtin_neon_vld1q_dup_bf16(__p0, 43); \
32584   __ret; \
32585 })
32586 #else
32587 #define vld1q_dup_bf16(__p0) __extension__ ({ \
32588   bfloat16x8_t __ret; \
32589   __ret = (bfloat16x8_t) __builtin_neon_vld1q_dup_bf16(__p0, 43); \
32590   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
32591   __ret; \
32592 })
32593 #endif
32594 
32595 #ifdef __LITTLE_ENDIAN__
32596 #define vld1_dup_bf16(__p0) __extension__ ({ \
32597   bfloat16x4_t __ret; \
32598   __ret = (bfloat16x4_t) __builtin_neon_vld1_dup_bf16(__p0, 11); \
32599   __ret; \
32600 })
32601 #else
32602 #define vld1_dup_bf16(__p0) __extension__ ({ \
32603   bfloat16x4_t __ret; \
32604   __ret = (bfloat16x4_t) __builtin_neon_vld1_dup_bf16(__p0, 11); \
32605   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
32606   __ret; \
32607 })
32608 #endif
32609 
32610 #ifdef __LITTLE_ENDIAN__
32611 #define vld1q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32612   bfloat16x8_t __ret; \
32613   bfloat16x8_t __s1 = __p1; \
32614   __ret = (bfloat16x8_t) __builtin_neon_vld1q_lane_bf16(__p0, (int8x16_t)__s1, __p2, 43); \
32615   __ret; \
32616 })
32617 #else
32618 #define vld1q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32619   bfloat16x8_t __ret; \
32620   bfloat16x8_t __s1 = __p1; \
32621   bfloat16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
32622   __ret = (bfloat16x8_t) __builtin_neon_vld1q_lane_bf16(__p0, (int8x16_t)__rev1, __p2, 43); \
32623   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
32624   __ret; \
32625 })
32626 #endif
32627 
32628 #ifdef __LITTLE_ENDIAN__
32629 #define vld1_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32630   bfloat16x4_t __ret; \
32631   bfloat16x4_t __s1 = __p1; \
32632   __ret = (bfloat16x4_t) __builtin_neon_vld1_lane_bf16(__p0, (int8x8_t)__s1, __p2, 11); \
32633   __ret; \
32634 })
32635 #else
32636 #define vld1_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32637   bfloat16x4_t __ret; \
32638   bfloat16x4_t __s1 = __p1; \
32639   bfloat16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
32640   __ret = (bfloat16x4_t) __builtin_neon_vld1_lane_bf16(__p0, (int8x8_t)__rev1, __p2, 11); \
32641   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
32642   __ret; \
32643 })
32644 #endif
32645 
32646 #ifdef __LITTLE_ENDIAN__
32647 #define vld1q_bf16_x2(__p0) __extension__ ({ \
32648   bfloat16x8x2_t __ret; \
32649   __builtin_neon_vld1q_bf16_x2(&__ret, __p0, 43); \
32650   __ret; \
32651 })
32652 #else
32653 #define vld1q_bf16_x2(__p0) __extension__ ({ \
32654   bfloat16x8x2_t __ret; \
32655   __builtin_neon_vld1q_bf16_x2(&__ret, __p0, 43); \
32656  \
32657   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32658   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32659   __ret; \
32660 })
32661 #endif
32662 
32663 #ifdef __LITTLE_ENDIAN__
32664 #define vld1_bf16_x2(__p0) __extension__ ({ \
32665   bfloat16x4x2_t __ret; \
32666   __builtin_neon_vld1_bf16_x2(&__ret, __p0, 11); \
32667   __ret; \
32668 })
32669 #else
32670 #define vld1_bf16_x2(__p0) __extension__ ({ \
32671   bfloat16x4x2_t __ret; \
32672   __builtin_neon_vld1_bf16_x2(&__ret, __p0, 11); \
32673  \
32674   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
32675   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
32676   __ret; \
32677 })
32678 #endif
32679 
32680 #ifdef __LITTLE_ENDIAN__
32681 #define vld1q_bf16_x3(__p0) __extension__ ({ \
32682   bfloat16x8x3_t __ret; \
32683   __builtin_neon_vld1q_bf16_x3(&__ret, __p0, 43); \
32684   __ret; \
32685 })
32686 #else
32687 #define vld1q_bf16_x3(__p0) __extension__ ({ \
32688   bfloat16x8x3_t __ret; \
32689   __builtin_neon_vld1q_bf16_x3(&__ret, __p0, 43); \
32690  \
32691   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32692   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32693   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
32694   __ret; \
32695 })
32696 #endif
32697 
32698 #ifdef __LITTLE_ENDIAN__
32699 #define vld1_bf16_x3(__p0) __extension__ ({ \
32700   bfloat16x4x3_t __ret; \
32701   __builtin_neon_vld1_bf16_x3(&__ret, __p0, 11); \
32702   __ret; \
32703 })
32704 #else
32705 #define vld1_bf16_x3(__p0) __extension__ ({ \
32706   bfloat16x4x3_t __ret; \
32707   __builtin_neon_vld1_bf16_x3(&__ret, __p0, 11); \
32708  \
32709   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
32710   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
32711   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
32712   __ret; \
32713 })
32714 #endif
32715 
32716 #ifdef __LITTLE_ENDIAN__
32717 #define vld1q_bf16_x4(__p0) __extension__ ({ \
32718   bfloat16x8x4_t __ret; \
32719   __builtin_neon_vld1q_bf16_x4(&__ret, __p0, 43); \
32720   __ret; \
32721 })
32722 #else
32723 #define vld1q_bf16_x4(__p0) __extension__ ({ \
32724   bfloat16x8x4_t __ret; \
32725   __builtin_neon_vld1q_bf16_x4(&__ret, __p0, 43); \
32726  \
32727   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32728   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32729   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
32730   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
32731   __ret; \
32732 })
32733 #endif
32734 
32735 #ifdef __LITTLE_ENDIAN__
32736 #define vld1_bf16_x4(__p0) __extension__ ({ \
32737   bfloat16x4x4_t __ret; \
32738   __builtin_neon_vld1_bf16_x4(&__ret, __p0, 11); \
32739   __ret; \
32740 })
32741 #else
32742 #define vld1_bf16_x4(__p0) __extension__ ({ \
32743   bfloat16x4x4_t __ret; \
32744   __builtin_neon_vld1_bf16_x4(&__ret, __p0, 11); \
32745  \
32746   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
32747   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
32748   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
32749   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
32750   __ret; \
32751 })
32752 #endif
32753 
32754 #ifdef __LITTLE_ENDIAN__
32755 #define vld2q_bf16(__p0) __extension__ ({ \
32756   bfloat16x8x2_t __ret; \
32757   __builtin_neon_vld2q_bf16(&__ret, __p0, 43); \
32758   __ret; \
32759 })
32760 #else
32761 #define vld2q_bf16(__p0) __extension__ ({ \
32762   bfloat16x8x2_t __ret; \
32763   __builtin_neon_vld2q_bf16(&__ret, __p0, 43); \
32764  \
32765   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32766   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32767   __ret; \
32768 })
32769 #endif
32770 
32771 #ifdef __LITTLE_ENDIAN__
32772 #define vld2_bf16(__p0) __extension__ ({ \
32773   bfloat16x4x2_t __ret; \
32774   __builtin_neon_vld2_bf16(&__ret, __p0, 11); \
32775   __ret; \
32776 })
32777 #else
32778 #define vld2_bf16(__p0) __extension__ ({ \
32779   bfloat16x4x2_t __ret; \
32780   __builtin_neon_vld2_bf16(&__ret, __p0, 11); \
32781  \
32782   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
32783   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
32784   __ret; \
32785 })
32786 #endif
32787 
32788 #ifdef __LITTLE_ENDIAN__
32789 #define vld2q_dup_bf16(__p0) __extension__ ({ \
32790   bfloat16x8x2_t __ret; \
32791   __builtin_neon_vld2q_dup_bf16(&__ret, __p0, 43); \
32792   __ret; \
32793 })
32794 #else
32795 #define vld2q_dup_bf16(__p0) __extension__ ({ \
32796   bfloat16x8x2_t __ret; \
32797   __builtin_neon_vld2q_dup_bf16(&__ret, __p0, 43); \
32798  \
32799   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32800   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32801   __ret; \
32802 })
32803 #endif
32804 
32805 #ifdef __LITTLE_ENDIAN__
32806 #define vld2_dup_bf16(__p0) __extension__ ({ \
32807   bfloat16x4x2_t __ret; \
32808   __builtin_neon_vld2_dup_bf16(&__ret, __p0, 11); \
32809   __ret; \
32810 })
32811 #else
32812 #define vld2_dup_bf16(__p0) __extension__ ({ \
32813   bfloat16x4x2_t __ret; \
32814   __builtin_neon_vld2_dup_bf16(&__ret, __p0, 11); \
32815  \
32816   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
32817   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
32818   __ret; \
32819 })
32820 #endif
32821 
32822 #ifdef __LITTLE_ENDIAN__
32823 #define vld2q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32824   bfloat16x8x2_t __ret; \
32825   bfloat16x8x2_t __s1 = __p1; \
32826   __builtin_neon_vld2q_lane_bf16(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 43); \
32827   __ret; \
32828 })
32829 #else
32830 #define vld2q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32831   bfloat16x8x2_t __ret; \
32832   bfloat16x8x2_t __s1 = __p1; \
32833   bfloat16x8x2_t __rev1; \
32834   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32835   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32836   __builtin_neon_vld2q_lane_bf16(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 43); \
32837  \
32838   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32839   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32840   __ret; \
32841 })
32842 #endif
32843 
32844 #ifdef __LITTLE_ENDIAN__
32845 #define vld2_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32846   bfloat16x4x2_t __ret; \
32847   bfloat16x4x2_t __s1 = __p1; \
32848   __builtin_neon_vld2_lane_bf16(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 11); \
32849   __ret; \
32850 })
32851 #else
32852 #define vld2_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32853   bfloat16x4x2_t __ret; \
32854   bfloat16x4x2_t __s1 = __p1; \
32855   bfloat16x4x2_t __rev1; \
32856   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
32857   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
32858   __builtin_neon_vld2_lane_bf16(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 11); \
32859  \
32860   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
32861   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
32862   __ret; \
32863 })
32864 #endif
32865 
32866 #ifdef __LITTLE_ENDIAN__
32867 #define vld3q_bf16(__p0) __extension__ ({ \
32868   bfloat16x8x3_t __ret; \
32869   __builtin_neon_vld3q_bf16(&__ret, __p0, 43); \
32870   __ret; \
32871 })
32872 #else
32873 #define vld3q_bf16(__p0) __extension__ ({ \
32874   bfloat16x8x3_t __ret; \
32875   __builtin_neon_vld3q_bf16(&__ret, __p0, 43); \
32876  \
32877   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32878   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32879   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
32880   __ret; \
32881 })
32882 #endif
32883 
32884 #ifdef __LITTLE_ENDIAN__
32885 #define vld3_bf16(__p0) __extension__ ({ \
32886   bfloat16x4x3_t __ret; \
32887   __builtin_neon_vld3_bf16(&__ret, __p0, 11); \
32888   __ret; \
32889 })
32890 #else
32891 #define vld3_bf16(__p0) __extension__ ({ \
32892   bfloat16x4x3_t __ret; \
32893   __builtin_neon_vld3_bf16(&__ret, __p0, 11); \
32894  \
32895   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
32896   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
32897   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
32898   __ret; \
32899 })
32900 #endif
32901 
32902 #ifdef __LITTLE_ENDIAN__
32903 #define vld3q_dup_bf16(__p0) __extension__ ({ \
32904   bfloat16x8x3_t __ret; \
32905   __builtin_neon_vld3q_dup_bf16(&__ret, __p0, 43); \
32906   __ret; \
32907 })
32908 #else
32909 #define vld3q_dup_bf16(__p0) __extension__ ({ \
32910   bfloat16x8x3_t __ret; \
32911   __builtin_neon_vld3q_dup_bf16(&__ret, __p0, 43); \
32912  \
32913   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32914   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32915   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
32916   __ret; \
32917 })
32918 #endif
32919 
32920 #ifdef __LITTLE_ENDIAN__
32921 #define vld3_dup_bf16(__p0) __extension__ ({ \
32922   bfloat16x4x3_t __ret; \
32923   __builtin_neon_vld3_dup_bf16(&__ret, __p0, 11); \
32924   __ret; \
32925 })
32926 #else
32927 #define vld3_dup_bf16(__p0) __extension__ ({ \
32928   bfloat16x4x3_t __ret; \
32929   __builtin_neon_vld3_dup_bf16(&__ret, __p0, 11); \
32930  \
32931   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
32932   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
32933   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
32934   __ret; \
32935 })
32936 #endif
32937 
32938 #ifdef __LITTLE_ENDIAN__
32939 #define vld3q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32940   bfloat16x8x3_t __ret; \
32941   bfloat16x8x3_t __s1 = __p1; \
32942   __builtin_neon_vld3q_lane_bf16(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 43); \
32943   __ret; \
32944 })
32945 #else
32946 #define vld3q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32947   bfloat16x8x3_t __ret; \
32948   bfloat16x8x3_t __s1 = __p1; \
32949   bfloat16x8x3_t __rev1; \
32950   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32951   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32952   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
32953   __builtin_neon_vld3q_lane_bf16(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 43); \
32954  \
32955   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32956   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32957   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
32958   __ret; \
32959 })
32960 #endif
32961 
32962 #ifdef __LITTLE_ENDIAN__
32963 #define vld3_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32964   bfloat16x4x3_t __ret; \
32965   bfloat16x4x3_t __s1 = __p1; \
32966   __builtin_neon_vld3_lane_bf16(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 11); \
32967   __ret; \
32968 })
32969 #else
32970 #define vld3_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
32971   bfloat16x4x3_t __ret; \
32972   bfloat16x4x3_t __s1 = __p1; \
32973   bfloat16x4x3_t __rev1; \
32974   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
32975   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
32976   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
32977   __builtin_neon_vld3_lane_bf16(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 11); \
32978  \
32979   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
32980   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
32981   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
32982   __ret; \
32983 })
32984 #endif
32985 
32986 #ifdef __LITTLE_ENDIAN__
32987 #define vld4q_bf16(__p0) __extension__ ({ \
32988   bfloat16x8x4_t __ret; \
32989   __builtin_neon_vld4q_bf16(&__ret, __p0, 43); \
32990   __ret; \
32991 })
32992 #else
32993 #define vld4q_bf16(__p0) __extension__ ({ \
32994   bfloat16x8x4_t __ret; \
32995   __builtin_neon_vld4q_bf16(&__ret, __p0, 43); \
32996  \
32997   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
32998   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
32999   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33000   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
33001   __ret; \
33002 })
33003 #endif
33004 
33005 #ifdef __LITTLE_ENDIAN__
33006 #define vld4_bf16(__p0) __extension__ ({ \
33007   bfloat16x4x4_t __ret; \
33008   __builtin_neon_vld4_bf16(&__ret, __p0, 11); \
33009   __ret; \
33010 })
33011 #else
33012 #define vld4_bf16(__p0) __extension__ ({ \
33013   bfloat16x4x4_t __ret; \
33014   __builtin_neon_vld4_bf16(&__ret, __p0, 11); \
33015  \
33016   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
33017   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
33018   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
33019   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
33020   __ret; \
33021 })
33022 #endif
33023 
33024 #ifdef __LITTLE_ENDIAN__
33025 #define vld4q_dup_bf16(__p0) __extension__ ({ \
33026   bfloat16x8x4_t __ret; \
33027   __builtin_neon_vld4q_dup_bf16(&__ret, __p0, 43); \
33028   __ret; \
33029 })
33030 #else
33031 #define vld4q_dup_bf16(__p0) __extension__ ({ \
33032   bfloat16x8x4_t __ret; \
33033   __builtin_neon_vld4q_dup_bf16(&__ret, __p0, 43); \
33034  \
33035   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33036   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33037   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33038   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
33039   __ret; \
33040 })
33041 #endif
33042 
33043 #ifdef __LITTLE_ENDIAN__
33044 #define vld4_dup_bf16(__p0) __extension__ ({ \
33045   bfloat16x4x4_t __ret; \
33046   __builtin_neon_vld4_dup_bf16(&__ret, __p0, 11); \
33047   __ret; \
33048 })
33049 #else
33050 #define vld4_dup_bf16(__p0) __extension__ ({ \
33051   bfloat16x4x4_t __ret; \
33052   __builtin_neon_vld4_dup_bf16(&__ret, __p0, 11); \
33053  \
33054   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
33055   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
33056   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
33057   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
33058   __ret; \
33059 })
33060 #endif
33061 
33062 #ifdef __LITTLE_ENDIAN__
33063 #define vld4q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33064   bfloat16x8x4_t __ret; \
33065   bfloat16x8x4_t __s1 = __p1; \
33066   __builtin_neon_vld4q_lane_bf16(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 43); \
33067   __ret; \
33068 })
33069 #else
33070 #define vld4q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33071   bfloat16x8x4_t __ret; \
33072   bfloat16x8x4_t __s1 = __p1; \
33073   bfloat16x8x4_t __rev1; \
33074   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33075   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33076   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33077   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
33078   __builtin_neon_vld4q_lane_bf16(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 43); \
33079  \
33080   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33081   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33082   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33083   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
33084   __ret; \
33085 })
33086 #endif
33087 
33088 #ifdef __LITTLE_ENDIAN__
33089 #define vld4_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33090   bfloat16x4x4_t __ret; \
33091   bfloat16x4x4_t __s1 = __p1; \
33092   __builtin_neon_vld4_lane_bf16(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 11); \
33093   __ret; \
33094 })
33095 #else
33096 #define vld4_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33097   bfloat16x4x4_t __ret; \
33098   bfloat16x4x4_t __s1 = __p1; \
33099   bfloat16x4x4_t __rev1; \
33100   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33101   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33102   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
33103   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
33104   __builtin_neon_vld4_lane_bf16(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 11); \
33105  \
33106   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
33107   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
33108   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
33109   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
33110   __ret; \
33111 })
33112 #endif
33113 
33114 #ifdef __LITTLE_ENDIAN__
33115 #define vsetq_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33116   bfloat16x8_t __ret; \
33117   bfloat16_t __s0 = __p0; \
33118   bfloat16x8_t __s1 = __p1; \
33119   __ret = (bfloat16x8_t) __builtin_neon_vsetq_lane_bf16(__s0, (bfloat16x8_t)__s1, __p2); \
33120   __ret; \
33121 })
33122 #else
33123 #define vsetq_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33124   bfloat16x8_t __ret; \
33125   bfloat16_t __s0 = __p0; \
33126   bfloat16x8_t __s1 = __p1; \
33127   bfloat16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
33128   __ret = (bfloat16x8_t) __builtin_neon_vsetq_lane_bf16(__s0, (bfloat16x8_t)__rev1, __p2); \
33129   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
33130   __ret; \
33131 })
33132 #define __noswap_vsetq_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33133   bfloat16x8_t __ret; \
33134   bfloat16_t __s0 = __p0; \
33135   bfloat16x8_t __s1 = __p1; \
33136   __ret = (bfloat16x8_t) __builtin_neon_vsetq_lane_bf16(__s0, (bfloat16x8_t)__s1, __p2); \
33137   __ret; \
33138 })
33139 #endif
33140 
33141 #ifdef __LITTLE_ENDIAN__
33142 #define vset_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33143   bfloat16x4_t __ret; \
33144   bfloat16_t __s0 = __p0; \
33145   bfloat16x4_t __s1 = __p1; \
33146   __ret = (bfloat16x4_t) __builtin_neon_vset_lane_bf16(__s0, (bfloat16x4_t)__s1, __p2); \
33147   __ret; \
33148 })
33149 #else
33150 #define vset_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33151   bfloat16x4_t __ret; \
33152   bfloat16_t __s0 = __p0; \
33153   bfloat16x4_t __s1 = __p1; \
33154   bfloat16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
33155   __ret = (bfloat16x4_t) __builtin_neon_vset_lane_bf16(__s0, (bfloat16x4_t)__rev1, __p2); \
33156   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
33157   __ret; \
33158 })
33159 #define __noswap_vset_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33160   bfloat16x4_t __ret; \
33161   bfloat16_t __s0 = __p0; \
33162   bfloat16x4_t __s1 = __p1; \
33163   __ret = (bfloat16x4_t) __builtin_neon_vset_lane_bf16(__s0, (bfloat16x4_t)__s1, __p2); \
33164   __ret; \
33165 })
33166 #endif
33167 
33168 #ifdef __LITTLE_ENDIAN__
33169 #define vst1q_bf16(__p0, __p1) __extension__ ({ \
33170   bfloat16x8_t __s1 = __p1; \
33171   __builtin_neon_vst1q_bf16(__p0, (int8x16_t)__s1, 43); \
33172 })
33173 #else
33174 #define vst1q_bf16(__p0, __p1) __extension__ ({ \
33175   bfloat16x8_t __s1 = __p1; \
33176   bfloat16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
33177   __builtin_neon_vst1q_bf16(__p0, (int8x16_t)__rev1, 43); \
33178 })
33179 #endif
33180 
33181 #ifdef __LITTLE_ENDIAN__
33182 #define vst1_bf16(__p0, __p1) __extension__ ({ \
33183   bfloat16x4_t __s1 = __p1; \
33184   __builtin_neon_vst1_bf16(__p0, (int8x8_t)__s1, 11); \
33185 })
33186 #else
33187 #define vst1_bf16(__p0, __p1) __extension__ ({ \
33188   bfloat16x4_t __s1 = __p1; \
33189   bfloat16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
33190   __builtin_neon_vst1_bf16(__p0, (int8x8_t)__rev1, 11); \
33191 })
33192 #endif
33193 
33194 #ifdef __LITTLE_ENDIAN__
33195 #define vst1q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33196   bfloat16x8_t __s1 = __p1; \
33197   __builtin_neon_vst1q_lane_bf16(__p0, (int8x16_t)__s1, __p2, 43); \
33198 })
33199 #else
33200 #define vst1q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33201   bfloat16x8_t __s1 = __p1; \
33202   bfloat16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
33203   __builtin_neon_vst1q_lane_bf16(__p0, (int8x16_t)__rev1, __p2, 43); \
33204 })
33205 #endif
33206 
33207 #ifdef __LITTLE_ENDIAN__
33208 #define vst1_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33209   bfloat16x4_t __s1 = __p1; \
33210   __builtin_neon_vst1_lane_bf16(__p0, (int8x8_t)__s1, __p2, 11); \
33211 })
33212 #else
33213 #define vst1_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33214   bfloat16x4_t __s1 = __p1; \
33215   bfloat16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
33216   __builtin_neon_vst1_lane_bf16(__p0, (int8x8_t)__rev1, __p2, 11); \
33217 })
33218 #endif
33219 
33220 #ifdef __LITTLE_ENDIAN__
33221 #define vst1q_bf16_x2(__p0, __p1) __extension__ ({ \
33222   bfloat16x8x2_t __s1 = __p1; \
33223   __builtin_neon_vst1q_bf16_x2(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 43); \
33224 })
33225 #else
33226 #define vst1q_bf16_x2(__p0, __p1) __extension__ ({ \
33227   bfloat16x8x2_t __s1 = __p1; \
33228   bfloat16x8x2_t __rev1; \
33229   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33230   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33231   __builtin_neon_vst1q_bf16_x2(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 43); \
33232 })
33233 #endif
33234 
33235 #ifdef __LITTLE_ENDIAN__
33236 #define vst1_bf16_x2(__p0, __p1) __extension__ ({ \
33237   bfloat16x4x2_t __s1 = __p1; \
33238   __builtin_neon_vst1_bf16_x2(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 11); \
33239 })
33240 #else
33241 #define vst1_bf16_x2(__p0, __p1) __extension__ ({ \
33242   bfloat16x4x2_t __s1 = __p1; \
33243   bfloat16x4x2_t __rev1; \
33244   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33245   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33246   __builtin_neon_vst1_bf16_x2(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 11); \
33247 })
33248 #endif
33249 
33250 #ifdef __LITTLE_ENDIAN__
33251 #define vst1q_bf16_x3(__p0, __p1) __extension__ ({ \
33252   bfloat16x8x3_t __s1 = __p1; \
33253   __builtin_neon_vst1q_bf16_x3(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 43); \
33254 })
33255 #else
33256 #define vst1q_bf16_x3(__p0, __p1) __extension__ ({ \
33257   bfloat16x8x3_t __s1 = __p1; \
33258   bfloat16x8x3_t __rev1; \
33259   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33260   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33261   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33262   __builtin_neon_vst1q_bf16_x3(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 43); \
33263 })
33264 #endif
33265 
33266 #ifdef __LITTLE_ENDIAN__
33267 #define vst1_bf16_x3(__p0, __p1) __extension__ ({ \
33268   bfloat16x4x3_t __s1 = __p1; \
33269   __builtin_neon_vst1_bf16_x3(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 11); \
33270 })
33271 #else
33272 #define vst1_bf16_x3(__p0, __p1) __extension__ ({ \
33273   bfloat16x4x3_t __s1 = __p1; \
33274   bfloat16x4x3_t __rev1; \
33275   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33276   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33277   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
33278   __builtin_neon_vst1_bf16_x3(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 11); \
33279 })
33280 #endif
33281 
33282 #ifdef __LITTLE_ENDIAN__
33283 #define vst1q_bf16_x4(__p0, __p1) __extension__ ({ \
33284   bfloat16x8x4_t __s1 = __p1; \
33285   __builtin_neon_vst1q_bf16_x4(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 43); \
33286 })
33287 #else
33288 #define vst1q_bf16_x4(__p0, __p1) __extension__ ({ \
33289   bfloat16x8x4_t __s1 = __p1; \
33290   bfloat16x8x4_t __rev1; \
33291   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33292   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33293   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33294   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
33295   __builtin_neon_vst1q_bf16_x4(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 43); \
33296 })
33297 #endif
33298 
33299 #ifdef __LITTLE_ENDIAN__
33300 #define vst1_bf16_x4(__p0, __p1) __extension__ ({ \
33301   bfloat16x4x4_t __s1 = __p1; \
33302   __builtin_neon_vst1_bf16_x4(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 11); \
33303 })
33304 #else
33305 #define vst1_bf16_x4(__p0, __p1) __extension__ ({ \
33306   bfloat16x4x4_t __s1 = __p1; \
33307   bfloat16x4x4_t __rev1; \
33308   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33309   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33310   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
33311   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
33312   __builtin_neon_vst1_bf16_x4(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 11); \
33313 })
33314 #endif
33315 
33316 #ifdef __LITTLE_ENDIAN__
33317 #define vst2q_bf16(__p0, __p1) __extension__ ({ \
33318   bfloat16x8x2_t __s1 = __p1; \
33319   __builtin_neon_vst2q_bf16(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 43); \
33320 })
33321 #else
33322 #define vst2q_bf16(__p0, __p1) __extension__ ({ \
33323   bfloat16x8x2_t __s1 = __p1; \
33324   bfloat16x8x2_t __rev1; \
33325   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33326   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33327   __builtin_neon_vst2q_bf16(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 43); \
33328 })
33329 #endif
33330 
33331 #ifdef __LITTLE_ENDIAN__
33332 #define vst2_bf16(__p0, __p1) __extension__ ({ \
33333   bfloat16x4x2_t __s1 = __p1; \
33334   __builtin_neon_vst2_bf16(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 11); \
33335 })
33336 #else
33337 #define vst2_bf16(__p0, __p1) __extension__ ({ \
33338   bfloat16x4x2_t __s1 = __p1; \
33339   bfloat16x4x2_t __rev1; \
33340   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33341   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33342   __builtin_neon_vst2_bf16(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 11); \
33343 })
33344 #endif
33345 
33346 #ifdef __LITTLE_ENDIAN__
33347 #define vst2q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33348   bfloat16x8x2_t __s1 = __p1; \
33349   __builtin_neon_vst2q_lane_bf16(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 43); \
33350 })
33351 #else
33352 #define vst2q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33353   bfloat16x8x2_t __s1 = __p1; \
33354   bfloat16x8x2_t __rev1; \
33355   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33356   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33357   __builtin_neon_vst2q_lane_bf16(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 43); \
33358 })
33359 #endif
33360 
33361 #ifdef __LITTLE_ENDIAN__
33362 #define vst2_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33363   bfloat16x4x2_t __s1 = __p1; \
33364   __builtin_neon_vst2_lane_bf16(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 11); \
33365 })
33366 #else
33367 #define vst2_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33368   bfloat16x4x2_t __s1 = __p1; \
33369   bfloat16x4x2_t __rev1; \
33370   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33371   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33372   __builtin_neon_vst2_lane_bf16(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 11); \
33373 })
33374 #endif
33375 
33376 #ifdef __LITTLE_ENDIAN__
33377 #define vst3q_bf16(__p0, __p1) __extension__ ({ \
33378   bfloat16x8x3_t __s1 = __p1; \
33379   __builtin_neon_vst3q_bf16(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 43); \
33380 })
33381 #else
33382 #define vst3q_bf16(__p0, __p1) __extension__ ({ \
33383   bfloat16x8x3_t __s1 = __p1; \
33384   bfloat16x8x3_t __rev1; \
33385   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33386   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33387   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33388   __builtin_neon_vst3q_bf16(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 43); \
33389 })
33390 #endif
33391 
33392 #ifdef __LITTLE_ENDIAN__
33393 #define vst3_bf16(__p0, __p1) __extension__ ({ \
33394   bfloat16x4x3_t __s1 = __p1; \
33395   __builtin_neon_vst3_bf16(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 11); \
33396 })
33397 #else
33398 #define vst3_bf16(__p0, __p1) __extension__ ({ \
33399   bfloat16x4x3_t __s1 = __p1; \
33400   bfloat16x4x3_t __rev1; \
33401   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33402   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33403   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
33404   __builtin_neon_vst3_bf16(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 11); \
33405 })
33406 #endif
33407 
33408 #ifdef __LITTLE_ENDIAN__
33409 #define vst3q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33410   bfloat16x8x3_t __s1 = __p1; \
33411   __builtin_neon_vst3q_lane_bf16(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 43); \
33412 })
33413 #else
33414 #define vst3q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33415   bfloat16x8x3_t __s1 = __p1; \
33416   bfloat16x8x3_t __rev1; \
33417   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33418   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33419   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33420   __builtin_neon_vst3q_lane_bf16(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 43); \
33421 })
33422 #endif
33423 
33424 #ifdef __LITTLE_ENDIAN__
33425 #define vst3_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33426   bfloat16x4x3_t __s1 = __p1; \
33427   __builtin_neon_vst3_lane_bf16(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 11); \
33428 })
33429 #else
33430 #define vst3_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33431   bfloat16x4x3_t __s1 = __p1; \
33432   bfloat16x4x3_t __rev1; \
33433   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33434   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33435   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
33436   __builtin_neon_vst3_lane_bf16(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 11); \
33437 })
33438 #endif
33439 
33440 #ifdef __LITTLE_ENDIAN__
33441 #define vst4q_bf16(__p0, __p1) __extension__ ({ \
33442   bfloat16x8x4_t __s1 = __p1; \
33443   __builtin_neon_vst4q_bf16(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 43); \
33444 })
33445 #else
33446 #define vst4q_bf16(__p0, __p1) __extension__ ({ \
33447   bfloat16x8x4_t __s1 = __p1; \
33448   bfloat16x8x4_t __rev1; \
33449   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33450   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33451   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33452   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
33453   __builtin_neon_vst4q_bf16(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 43); \
33454 })
33455 #endif
33456 
33457 #ifdef __LITTLE_ENDIAN__
33458 #define vst4_bf16(__p0, __p1) __extension__ ({ \
33459   bfloat16x4x4_t __s1 = __p1; \
33460   __builtin_neon_vst4_bf16(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 11); \
33461 })
33462 #else
33463 #define vst4_bf16(__p0, __p1) __extension__ ({ \
33464   bfloat16x4x4_t __s1 = __p1; \
33465   bfloat16x4x4_t __rev1; \
33466   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33467   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33468   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
33469   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
33470   __builtin_neon_vst4_bf16(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 11); \
33471 })
33472 #endif
33473 
33474 #ifdef __LITTLE_ENDIAN__
33475 #define vst4q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33476   bfloat16x8x4_t __s1 = __p1; \
33477   __builtin_neon_vst4q_lane_bf16(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 43); \
33478 })
33479 #else
33480 #define vst4q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33481   bfloat16x8x4_t __s1 = __p1; \
33482   bfloat16x8x4_t __rev1; \
33483   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33484   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33485   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33486   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
33487   __builtin_neon_vst4q_lane_bf16(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 43); \
33488 })
33489 #endif
33490 
33491 #ifdef __LITTLE_ENDIAN__
33492 #define vst4_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33493   bfloat16x4x4_t __s1 = __p1; \
33494   __builtin_neon_vst4_lane_bf16(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 11); \
33495 })
33496 #else
33497 #define vst4_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
33498   bfloat16x4x4_t __s1 = __p1; \
33499   bfloat16x4x4_t __rev1; \
33500   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
33501   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
33502   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
33503   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
33504   __builtin_neon_vst4_lane_bf16(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 11); \
33505 })
33506 #endif
33507 
33508 #ifdef __LITTLE_ENDIAN__
vdotq_u32(uint32x4_t __p0,uint8x16_t __p1,uint8x16_t __p2)33509 __ai __attribute__((target("dotprod"))) uint32x4_t vdotq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
33510   uint32x4_t __ret;
33511   __ret = (uint32x4_t) __builtin_neon_vdotq_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
33512   return __ret;
33513 }
33514 #else
vdotq_u32(uint32x4_t __p0,uint8x16_t __p1,uint8x16_t __p2)33515 __ai __attribute__((target("dotprod"))) uint32x4_t vdotq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
33516   uint32x4_t __ret;
33517   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33518   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
33519   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
33520   __ret = (uint32x4_t) __builtin_neon_vdotq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
33521   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33522   return __ret;
33523 }
__noswap_vdotq_u32(uint32x4_t __p0,uint8x16_t __p1,uint8x16_t __p2)33524 __ai __attribute__((target("dotprod"))) uint32x4_t __noswap_vdotq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
33525   uint32x4_t __ret;
33526   __ret = (uint32x4_t) __builtin_neon_vdotq_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
33527   return __ret;
33528 }
33529 #endif
33530 
33531 #ifdef __LITTLE_ENDIAN__
vdotq_s32(int32x4_t __p0,int8x16_t __p1,int8x16_t __p2)33532 __ai __attribute__((target("dotprod"))) int32x4_t vdotq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
33533   int32x4_t __ret;
33534   __ret = (int32x4_t) __builtin_neon_vdotq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
33535   return __ret;
33536 }
33537 #else
vdotq_s32(int32x4_t __p0,int8x16_t __p1,int8x16_t __p2)33538 __ai __attribute__((target("dotprod"))) int32x4_t vdotq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
33539   int32x4_t __ret;
33540   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33541   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
33542   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
33543   __ret = (int32x4_t) __builtin_neon_vdotq_s32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
33544   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33545   return __ret;
33546 }
__noswap_vdotq_s32(int32x4_t __p0,int8x16_t __p1,int8x16_t __p2)33547 __ai __attribute__((target("dotprod"))) int32x4_t __noswap_vdotq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
33548   int32x4_t __ret;
33549   __ret = (int32x4_t) __builtin_neon_vdotq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
33550   return __ret;
33551 }
33552 #endif
33553 
33554 #ifdef __LITTLE_ENDIAN__
vdot_u32(uint32x2_t __p0,uint8x8_t __p1,uint8x8_t __p2)33555 __ai __attribute__((target("dotprod"))) uint32x2_t vdot_u32(uint32x2_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
33556   uint32x2_t __ret;
33557   __ret = (uint32x2_t) __builtin_neon_vdot_u32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
33558   return __ret;
33559 }
33560 #else
vdot_u32(uint32x2_t __p0,uint8x8_t __p1,uint8x8_t __p2)33561 __ai __attribute__((target("dotprod"))) uint32x2_t vdot_u32(uint32x2_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
33562   uint32x2_t __ret;
33563   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33564   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33565   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
33566   __ret = (uint32x2_t) __builtin_neon_vdot_u32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 18);
33567   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33568   return __ret;
33569 }
__noswap_vdot_u32(uint32x2_t __p0,uint8x8_t __p1,uint8x8_t __p2)33570 __ai __attribute__((target("dotprod"))) uint32x2_t __noswap_vdot_u32(uint32x2_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
33571   uint32x2_t __ret;
33572   __ret = (uint32x2_t) __builtin_neon_vdot_u32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
33573   return __ret;
33574 }
33575 #endif
33576 
33577 #ifdef __LITTLE_ENDIAN__
vdot_s32(int32x2_t __p0,int8x8_t __p1,int8x8_t __p2)33578 __ai __attribute__((target("dotprod"))) int32x2_t vdot_s32(int32x2_t __p0, int8x8_t __p1, int8x8_t __p2) {
33579   int32x2_t __ret;
33580   __ret = (int32x2_t) __builtin_neon_vdot_s32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
33581   return __ret;
33582 }
33583 #else
vdot_s32(int32x2_t __p0,int8x8_t __p1,int8x8_t __p2)33584 __ai __attribute__((target("dotprod"))) int32x2_t vdot_s32(int32x2_t __p0, int8x8_t __p1, int8x8_t __p2) {
33585   int32x2_t __ret;
33586   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33587   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33588   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
33589   __ret = (int32x2_t) __builtin_neon_vdot_s32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
33590   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33591   return __ret;
33592 }
__noswap_vdot_s32(int32x2_t __p0,int8x8_t __p1,int8x8_t __p2)33593 __ai __attribute__((target("dotprod"))) int32x2_t __noswap_vdot_s32(int32x2_t __p0, int8x8_t __p1, int8x8_t __p2) {
33594   int32x2_t __ret;
33595   __ret = (int32x2_t) __builtin_neon_vdot_s32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
33596   return __ret;
33597 }
33598 #endif
33599 
33600 #ifdef __LITTLE_ENDIAN__
33601 #define vdotq_lane_u32(__p0_145, __p1_145, __p2_145, __p3_145) __extension__ ({ \
33602   uint32x4_t __ret_145; \
33603   uint32x4_t __s0_145 = __p0_145; \
33604   uint8x16_t __s1_145 = __p1_145; \
33605   uint8x8_t __s2_145 = __p2_145; \
33606 uint8x8_t __reint_145 = __s2_145; \
33607 uint32x4_t __reint1_145 = splatq_lane_u32(*(uint32x2_t *) &__reint_145, __p3_145); \
33608   __ret_145 = vdotq_u32(__s0_145, __s1_145, *(uint8x16_t *) &__reint1_145); \
33609   __ret_145; \
33610 })
33611 #else
33612 #define vdotq_lane_u32(__p0_146, __p1_146, __p2_146, __p3_146) __extension__ ({ \
33613   uint32x4_t __ret_146; \
33614   uint32x4_t __s0_146 = __p0_146; \
33615   uint8x16_t __s1_146 = __p1_146; \
33616   uint8x8_t __s2_146 = __p2_146; \
33617   uint32x4_t __rev0_146;  __rev0_146 = __builtin_shufflevector(__s0_146, __s0_146, 3, 2, 1, 0); \
33618   uint8x16_t __rev1_146;  __rev1_146 = __builtin_shufflevector(__s1_146, __s1_146, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
33619   uint8x8_t __rev2_146;  __rev2_146 = __builtin_shufflevector(__s2_146, __s2_146, 7, 6, 5, 4, 3, 2, 1, 0); \
33620 uint8x8_t __reint_146 = __rev2_146; \
33621 uint32x4_t __reint1_146 = __noswap_splatq_lane_u32(*(uint32x2_t *) &__reint_146, __p3_146); \
33622   __ret_146 = __noswap_vdotq_u32(__rev0_146, __rev1_146, *(uint8x16_t *) &__reint1_146); \
33623   __ret_146 = __builtin_shufflevector(__ret_146, __ret_146, 3, 2, 1, 0); \
33624   __ret_146; \
33625 })
33626 #endif
33627 
33628 #ifdef __LITTLE_ENDIAN__
33629 #define vdotq_lane_s32(__p0_147, __p1_147, __p2_147, __p3_147) __extension__ ({ \
33630   int32x4_t __ret_147; \
33631   int32x4_t __s0_147 = __p0_147; \
33632   int8x16_t __s1_147 = __p1_147; \
33633   int8x8_t __s2_147 = __p2_147; \
33634 int8x8_t __reint_147 = __s2_147; \
33635 int32x4_t __reint1_147 = splatq_lane_s32(*(int32x2_t *) &__reint_147, __p3_147); \
33636   __ret_147 = vdotq_s32(__s0_147, __s1_147, *(int8x16_t *) &__reint1_147); \
33637   __ret_147; \
33638 })
33639 #else
33640 #define vdotq_lane_s32(__p0_148, __p1_148, __p2_148, __p3_148) __extension__ ({ \
33641   int32x4_t __ret_148; \
33642   int32x4_t __s0_148 = __p0_148; \
33643   int8x16_t __s1_148 = __p1_148; \
33644   int8x8_t __s2_148 = __p2_148; \
33645   int32x4_t __rev0_148;  __rev0_148 = __builtin_shufflevector(__s0_148, __s0_148, 3, 2, 1, 0); \
33646   int8x16_t __rev1_148;  __rev1_148 = __builtin_shufflevector(__s1_148, __s1_148, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
33647   int8x8_t __rev2_148;  __rev2_148 = __builtin_shufflevector(__s2_148, __s2_148, 7, 6, 5, 4, 3, 2, 1, 0); \
33648 int8x8_t __reint_148 = __rev2_148; \
33649 int32x4_t __reint1_148 = __noswap_splatq_lane_s32(*(int32x2_t *) &__reint_148, __p3_148); \
33650   __ret_148 = __noswap_vdotq_s32(__rev0_148, __rev1_148, *(int8x16_t *) &__reint1_148); \
33651   __ret_148 = __builtin_shufflevector(__ret_148, __ret_148, 3, 2, 1, 0); \
33652   __ret_148; \
33653 })
33654 #endif
33655 
33656 #ifdef __LITTLE_ENDIAN__
33657 #define vdot_lane_u32(__p0_149, __p1_149, __p2_149, __p3_149) __extension__ ({ \
33658   uint32x2_t __ret_149; \
33659   uint32x2_t __s0_149 = __p0_149; \
33660   uint8x8_t __s1_149 = __p1_149; \
33661   uint8x8_t __s2_149 = __p2_149; \
33662 uint8x8_t __reint_149 = __s2_149; \
33663 uint32x2_t __reint1_149 = splat_lane_u32(*(uint32x2_t *) &__reint_149, __p3_149); \
33664   __ret_149 = vdot_u32(__s0_149, __s1_149, *(uint8x8_t *) &__reint1_149); \
33665   __ret_149; \
33666 })
33667 #else
33668 #define vdot_lane_u32(__p0_150, __p1_150, __p2_150, __p3_150) __extension__ ({ \
33669   uint32x2_t __ret_150; \
33670   uint32x2_t __s0_150 = __p0_150; \
33671   uint8x8_t __s1_150 = __p1_150; \
33672   uint8x8_t __s2_150 = __p2_150; \
33673   uint32x2_t __rev0_150;  __rev0_150 = __builtin_shufflevector(__s0_150, __s0_150, 1, 0); \
33674   uint8x8_t __rev1_150;  __rev1_150 = __builtin_shufflevector(__s1_150, __s1_150, 7, 6, 5, 4, 3, 2, 1, 0); \
33675   uint8x8_t __rev2_150;  __rev2_150 = __builtin_shufflevector(__s2_150, __s2_150, 7, 6, 5, 4, 3, 2, 1, 0); \
33676 uint8x8_t __reint_150 = __rev2_150; \
33677 uint32x2_t __reint1_150 = __noswap_splat_lane_u32(*(uint32x2_t *) &__reint_150, __p3_150); \
33678   __ret_150 = __noswap_vdot_u32(__rev0_150, __rev1_150, *(uint8x8_t *) &__reint1_150); \
33679   __ret_150 = __builtin_shufflevector(__ret_150, __ret_150, 1, 0); \
33680   __ret_150; \
33681 })
33682 #endif
33683 
33684 #ifdef __LITTLE_ENDIAN__
33685 #define vdot_lane_s32(__p0_151, __p1_151, __p2_151, __p3_151) __extension__ ({ \
33686   int32x2_t __ret_151; \
33687   int32x2_t __s0_151 = __p0_151; \
33688   int8x8_t __s1_151 = __p1_151; \
33689   int8x8_t __s2_151 = __p2_151; \
33690 int8x8_t __reint_151 = __s2_151; \
33691 int32x2_t __reint1_151 = splat_lane_s32(*(int32x2_t *) &__reint_151, __p3_151); \
33692   __ret_151 = vdot_s32(__s0_151, __s1_151, *(int8x8_t *) &__reint1_151); \
33693   __ret_151; \
33694 })
33695 #else
33696 #define vdot_lane_s32(__p0_152, __p1_152, __p2_152, __p3_152) __extension__ ({ \
33697   int32x2_t __ret_152; \
33698   int32x2_t __s0_152 = __p0_152; \
33699   int8x8_t __s1_152 = __p1_152; \
33700   int8x8_t __s2_152 = __p2_152; \
33701   int32x2_t __rev0_152;  __rev0_152 = __builtin_shufflevector(__s0_152, __s0_152, 1, 0); \
33702   int8x8_t __rev1_152;  __rev1_152 = __builtin_shufflevector(__s1_152, __s1_152, 7, 6, 5, 4, 3, 2, 1, 0); \
33703   int8x8_t __rev2_152;  __rev2_152 = __builtin_shufflevector(__s2_152, __s2_152, 7, 6, 5, 4, 3, 2, 1, 0); \
33704 int8x8_t __reint_152 = __rev2_152; \
33705 int32x2_t __reint1_152 = __noswap_splat_lane_s32(*(int32x2_t *) &__reint_152, __p3_152); \
33706   __ret_152 = __noswap_vdot_s32(__rev0_152, __rev1_152, *(int8x8_t *) &__reint1_152); \
33707   __ret_152 = __builtin_shufflevector(__ret_152, __ret_152, 1, 0); \
33708   __ret_152; \
33709 })
33710 #endif
33711 
33712 #ifdef __LITTLE_ENDIAN__
vabdq_f16(float16x8_t __p0,float16x8_t __p1)33713 __ai __attribute__((target("fullfp16"))) float16x8_t vabdq_f16(float16x8_t __p0, float16x8_t __p1) {
33714   float16x8_t __ret;
33715   __ret = (float16x8_t) __builtin_neon_vabdq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
33716   return __ret;
33717 }
33718 #else
vabdq_f16(float16x8_t __p0,float16x8_t __p1)33719 __ai __attribute__((target("fullfp16"))) float16x8_t vabdq_f16(float16x8_t __p0, float16x8_t __p1) {
33720   float16x8_t __ret;
33721   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
33722   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33723   __ret = (float16x8_t) __builtin_neon_vabdq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
33724   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
33725   return __ret;
33726 }
33727 #endif
33728 
33729 #ifdef __LITTLE_ENDIAN__
vabd_f16(float16x4_t __p0,float16x4_t __p1)33730 __ai __attribute__((target("fullfp16"))) float16x4_t vabd_f16(float16x4_t __p0, float16x4_t __p1) {
33731   float16x4_t __ret;
33732   __ret = (float16x4_t) __builtin_neon_vabd_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
33733   return __ret;
33734 }
33735 #else
vabd_f16(float16x4_t __p0,float16x4_t __p1)33736 __ai __attribute__((target("fullfp16"))) float16x4_t vabd_f16(float16x4_t __p0, float16x4_t __p1) {
33737   float16x4_t __ret;
33738   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33739   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
33740   __ret = (float16x4_t) __builtin_neon_vabd_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
33741   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33742   return __ret;
33743 }
33744 #endif
33745 
33746 #ifdef __LITTLE_ENDIAN__
vabsq_f16(float16x8_t __p0)33747 __ai __attribute__((target("fullfp16"))) float16x8_t vabsq_f16(float16x8_t __p0) {
33748   float16x8_t __ret;
33749   __ret = (float16x8_t) __builtin_neon_vabsq_f16((int8x16_t)__p0, 40);
33750   return __ret;
33751 }
33752 #else
vabsq_f16(float16x8_t __p0)33753 __ai __attribute__((target("fullfp16"))) float16x8_t vabsq_f16(float16x8_t __p0) {
33754   float16x8_t __ret;
33755   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
33756   __ret = (float16x8_t) __builtin_neon_vabsq_f16((int8x16_t)__rev0, 40);
33757   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
33758   return __ret;
33759 }
33760 #endif
33761 
33762 #ifdef __LITTLE_ENDIAN__
vabs_f16(float16x4_t __p0)33763 __ai __attribute__((target("fullfp16"))) float16x4_t vabs_f16(float16x4_t __p0) {
33764   float16x4_t __ret;
33765   __ret = (float16x4_t) __builtin_neon_vabs_f16((int8x8_t)__p0, 8);
33766   return __ret;
33767 }
33768 #else
vabs_f16(float16x4_t __p0)33769 __ai __attribute__((target("fullfp16"))) float16x4_t vabs_f16(float16x4_t __p0) {
33770   float16x4_t __ret;
33771   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33772   __ret = (float16x4_t) __builtin_neon_vabs_f16((int8x8_t)__rev0, 8);
33773   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33774   return __ret;
33775 }
33776 #endif
33777 
33778 #ifdef __LITTLE_ENDIAN__
vaddq_f16(float16x8_t __p0,float16x8_t __p1)33779 __ai __attribute__((target("fullfp16"))) float16x8_t vaddq_f16(float16x8_t __p0, float16x8_t __p1) {
33780   float16x8_t __ret;
33781   __ret = __p0 + __p1;
33782   return __ret;
33783 }
33784 #else
vaddq_f16(float16x8_t __p0,float16x8_t __p1)33785 __ai __attribute__((target("fullfp16"))) float16x8_t vaddq_f16(float16x8_t __p0, float16x8_t __p1) {
33786   float16x8_t __ret;
33787   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
33788   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33789   __ret = __rev0 + __rev1;
33790   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
33791   return __ret;
33792 }
33793 #endif
33794 
33795 #ifdef __LITTLE_ENDIAN__
vadd_f16(float16x4_t __p0,float16x4_t __p1)33796 __ai __attribute__((target("fullfp16"))) float16x4_t vadd_f16(float16x4_t __p0, float16x4_t __p1) {
33797   float16x4_t __ret;
33798   __ret = __p0 + __p1;
33799   return __ret;
33800 }
33801 #else
vadd_f16(float16x4_t __p0,float16x4_t __p1)33802 __ai __attribute__((target("fullfp16"))) float16x4_t vadd_f16(float16x4_t __p0, float16x4_t __p1) {
33803   float16x4_t __ret;
33804   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33805   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
33806   __ret = __rev0 + __rev1;
33807   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33808   return __ret;
33809 }
33810 #endif
33811 
33812 #ifdef __LITTLE_ENDIAN__
vbslq_f16(uint16x8_t __p0,float16x8_t __p1,float16x8_t __p2)33813 __ai __attribute__((target("fullfp16"))) float16x8_t vbslq_f16(uint16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
33814   float16x8_t __ret;
33815   __ret = (float16x8_t) __builtin_neon_vbslq_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
33816   return __ret;
33817 }
33818 #else
vbslq_f16(uint16x8_t __p0,float16x8_t __p1,float16x8_t __p2)33819 __ai __attribute__((target("fullfp16"))) float16x8_t vbslq_f16(uint16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
33820   float16x8_t __ret;
33821   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
33822   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33823   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
33824   __ret = (float16x8_t) __builtin_neon_vbslq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 40);
33825   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
33826   return __ret;
33827 }
33828 #endif
33829 
33830 #ifdef __LITTLE_ENDIAN__
vbsl_f16(uint16x4_t __p0,float16x4_t __p1,float16x4_t __p2)33831 __ai __attribute__((target("fullfp16"))) float16x4_t vbsl_f16(uint16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
33832   float16x4_t __ret;
33833   __ret = (float16x4_t) __builtin_neon_vbsl_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
33834   return __ret;
33835 }
33836 #else
vbsl_f16(uint16x4_t __p0,float16x4_t __p1,float16x4_t __p2)33837 __ai __attribute__((target("fullfp16"))) float16x4_t vbsl_f16(uint16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
33838   float16x4_t __ret;
33839   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33840   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
33841   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
33842   __ret = (float16x4_t) __builtin_neon_vbsl_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 8);
33843   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33844   return __ret;
33845 }
33846 #endif
33847 
33848 #ifdef __LITTLE_ENDIAN__
vcageq_f16(float16x8_t __p0,float16x8_t __p1)33849 __ai __attribute__((target("fullfp16"))) uint16x8_t vcageq_f16(float16x8_t __p0, float16x8_t __p1) {
33850   uint16x8_t __ret;
33851   __ret = (uint16x8_t) __builtin_neon_vcageq_f16((int8x16_t)__p0, (int8x16_t)__p1, 49);
33852   return __ret;
33853 }
33854 #else
vcageq_f16(float16x8_t __p0,float16x8_t __p1)33855 __ai __attribute__((target("fullfp16"))) uint16x8_t vcageq_f16(float16x8_t __p0, float16x8_t __p1) {
33856   uint16x8_t __ret;
33857   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
33858   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33859   __ret = (uint16x8_t) __builtin_neon_vcageq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
33860   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
33861   return __ret;
33862 }
33863 #endif
33864 
33865 #ifdef __LITTLE_ENDIAN__
vcage_f16(float16x4_t __p0,float16x4_t __p1)33866 __ai __attribute__((target("fullfp16"))) uint16x4_t vcage_f16(float16x4_t __p0, float16x4_t __p1) {
33867   uint16x4_t __ret;
33868   __ret = (uint16x4_t) __builtin_neon_vcage_f16((int8x8_t)__p0, (int8x8_t)__p1, 17);
33869   return __ret;
33870 }
33871 #else
vcage_f16(float16x4_t __p0,float16x4_t __p1)33872 __ai __attribute__((target("fullfp16"))) uint16x4_t vcage_f16(float16x4_t __p0, float16x4_t __p1) {
33873   uint16x4_t __ret;
33874   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33875   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
33876   __ret = (uint16x4_t) __builtin_neon_vcage_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
33877   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33878   return __ret;
33879 }
33880 #endif
33881 
33882 #ifdef __LITTLE_ENDIAN__
vcagtq_f16(float16x8_t __p0,float16x8_t __p1)33883 __ai __attribute__((target("fullfp16"))) uint16x8_t vcagtq_f16(float16x8_t __p0, float16x8_t __p1) {
33884   uint16x8_t __ret;
33885   __ret = (uint16x8_t) __builtin_neon_vcagtq_f16((int8x16_t)__p0, (int8x16_t)__p1, 49);
33886   return __ret;
33887 }
33888 #else
vcagtq_f16(float16x8_t __p0,float16x8_t __p1)33889 __ai __attribute__((target("fullfp16"))) uint16x8_t vcagtq_f16(float16x8_t __p0, float16x8_t __p1) {
33890   uint16x8_t __ret;
33891   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
33892   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33893   __ret = (uint16x8_t) __builtin_neon_vcagtq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
33894   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
33895   return __ret;
33896 }
33897 #endif
33898 
33899 #ifdef __LITTLE_ENDIAN__
vcagt_f16(float16x4_t __p0,float16x4_t __p1)33900 __ai __attribute__((target("fullfp16"))) uint16x4_t vcagt_f16(float16x4_t __p0, float16x4_t __p1) {
33901   uint16x4_t __ret;
33902   __ret = (uint16x4_t) __builtin_neon_vcagt_f16((int8x8_t)__p0, (int8x8_t)__p1, 17);
33903   return __ret;
33904 }
33905 #else
vcagt_f16(float16x4_t __p0,float16x4_t __p1)33906 __ai __attribute__((target("fullfp16"))) uint16x4_t vcagt_f16(float16x4_t __p0, float16x4_t __p1) {
33907   uint16x4_t __ret;
33908   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33909   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
33910   __ret = (uint16x4_t) __builtin_neon_vcagt_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
33911   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33912   return __ret;
33913 }
33914 #endif
33915 
33916 #ifdef __LITTLE_ENDIAN__
vcaleq_f16(float16x8_t __p0,float16x8_t __p1)33917 __ai __attribute__((target("fullfp16"))) uint16x8_t vcaleq_f16(float16x8_t __p0, float16x8_t __p1) {
33918   uint16x8_t __ret;
33919   __ret = (uint16x8_t) __builtin_neon_vcaleq_f16((int8x16_t)__p0, (int8x16_t)__p1, 49);
33920   return __ret;
33921 }
33922 #else
vcaleq_f16(float16x8_t __p0,float16x8_t __p1)33923 __ai __attribute__((target("fullfp16"))) uint16x8_t vcaleq_f16(float16x8_t __p0, float16x8_t __p1) {
33924   uint16x8_t __ret;
33925   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
33926   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33927   __ret = (uint16x8_t) __builtin_neon_vcaleq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
33928   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
33929   return __ret;
33930 }
33931 #endif
33932 
33933 #ifdef __LITTLE_ENDIAN__
vcale_f16(float16x4_t __p0,float16x4_t __p1)33934 __ai __attribute__((target("fullfp16"))) uint16x4_t vcale_f16(float16x4_t __p0, float16x4_t __p1) {
33935   uint16x4_t __ret;
33936   __ret = (uint16x4_t) __builtin_neon_vcale_f16((int8x8_t)__p0, (int8x8_t)__p1, 17);
33937   return __ret;
33938 }
33939 #else
vcale_f16(float16x4_t __p0,float16x4_t __p1)33940 __ai __attribute__((target("fullfp16"))) uint16x4_t vcale_f16(float16x4_t __p0, float16x4_t __p1) {
33941   uint16x4_t __ret;
33942   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33943   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
33944   __ret = (uint16x4_t) __builtin_neon_vcale_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
33945   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33946   return __ret;
33947 }
33948 #endif
33949 
33950 #ifdef __LITTLE_ENDIAN__
vcaltq_f16(float16x8_t __p0,float16x8_t __p1)33951 __ai __attribute__((target("fullfp16"))) uint16x8_t vcaltq_f16(float16x8_t __p0, float16x8_t __p1) {
33952   uint16x8_t __ret;
33953   __ret = (uint16x8_t) __builtin_neon_vcaltq_f16((int8x16_t)__p0, (int8x16_t)__p1, 49);
33954   return __ret;
33955 }
33956 #else
vcaltq_f16(float16x8_t __p0,float16x8_t __p1)33957 __ai __attribute__((target("fullfp16"))) uint16x8_t vcaltq_f16(float16x8_t __p0, float16x8_t __p1) {
33958   uint16x8_t __ret;
33959   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
33960   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33961   __ret = (uint16x8_t) __builtin_neon_vcaltq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
33962   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
33963   return __ret;
33964 }
33965 #endif
33966 
33967 #ifdef __LITTLE_ENDIAN__
vcalt_f16(float16x4_t __p0,float16x4_t __p1)33968 __ai __attribute__((target("fullfp16"))) uint16x4_t vcalt_f16(float16x4_t __p0, float16x4_t __p1) {
33969   uint16x4_t __ret;
33970   __ret = (uint16x4_t) __builtin_neon_vcalt_f16((int8x8_t)__p0, (int8x8_t)__p1, 17);
33971   return __ret;
33972 }
33973 #else
vcalt_f16(float16x4_t __p0,float16x4_t __p1)33974 __ai __attribute__((target("fullfp16"))) uint16x4_t vcalt_f16(float16x4_t __p0, float16x4_t __p1) {
33975   uint16x4_t __ret;
33976   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33977   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
33978   __ret = (uint16x4_t) __builtin_neon_vcalt_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
33979   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33980   return __ret;
33981 }
33982 #endif
33983 
33984 #ifdef __LITTLE_ENDIAN__
vceqq_f16(float16x8_t __p0,float16x8_t __p1)33985 __ai __attribute__((target("fullfp16"))) uint16x8_t vceqq_f16(float16x8_t __p0, float16x8_t __p1) {
33986   uint16x8_t __ret;
33987   __ret = (uint16x8_t)(__p0 == __p1);
33988   return __ret;
33989 }
33990 #else
vceqq_f16(float16x8_t __p0,float16x8_t __p1)33991 __ai __attribute__((target("fullfp16"))) uint16x8_t vceqq_f16(float16x8_t __p0, float16x8_t __p1) {
33992   uint16x8_t __ret;
33993   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
33994   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
33995   __ret = (uint16x8_t)(__rev0 == __rev1);
33996   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
33997   return __ret;
33998 }
33999 #endif
34000 
34001 #ifdef __LITTLE_ENDIAN__
vceq_f16(float16x4_t __p0,float16x4_t __p1)34002 __ai __attribute__((target("fullfp16"))) uint16x4_t vceq_f16(float16x4_t __p0, float16x4_t __p1) {
34003   uint16x4_t __ret;
34004   __ret = (uint16x4_t)(__p0 == __p1);
34005   return __ret;
34006 }
34007 #else
vceq_f16(float16x4_t __p0,float16x4_t __p1)34008 __ai __attribute__((target("fullfp16"))) uint16x4_t vceq_f16(float16x4_t __p0, float16x4_t __p1) {
34009   uint16x4_t __ret;
34010   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34011   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34012   __ret = (uint16x4_t)(__rev0 == __rev1);
34013   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34014   return __ret;
34015 }
34016 #endif
34017 
34018 #ifdef __LITTLE_ENDIAN__
vceqzq_f16(float16x8_t __p0)34019 __ai __attribute__((target("fullfp16"))) uint16x8_t vceqzq_f16(float16x8_t __p0) {
34020   uint16x8_t __ret;
34021   __ret = (uint16x8_t) __builtin_neon_vceqzq_f16((int8x16_t)__p0, 49);
34022   return __ret;
34023 }
34024 #else
vceqzq_f16(float16x8_t __p0)34025 __ai __attribute__((target("fullfp16"))) uint16x8_t vceqzq_f16(float16x8_t __p0) {
34026   uint16x8_t __ret;
34027   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34028   __ret = (uint16x8_t) __builtin_neon_vceqzq_f16((int8x16_t)__rev0, 49);
34029   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34030   return __ret;
34031 }
34032 #endif
34033 
34034 #ifdef __LITTLE_ENDIAN__
vceqz_f16(float16x4_t __p0)34035 __ai __attribute__((target("fullfp16"))) uint16x4_t vceqz_f16(float16x4_t __p0) {
34036   uint16x4_t __ret;
34037   __ret = (uint16x4_t) __builtin_neon_vceqz_f16((int8x8_t)__p0, 17);
34038   return __ret;
34039 }
34040 #else
vceqz_f16(float16x4_t __p0)34041 __ai __attribute__((target("fullfp16"))) uint16x4_t vceqz_f16(float16x4_t __p0) {
34042   uint16x4_t __ret;
34043   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34044   __ret = (uint16x4_t) __builtin_neon_vceqz_f16((int8x8_t)__rev0, 17);
34045   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34046   return __ret;
34047 }
34048 #endif
34049 
34050 #ifdef __LITTLE_ENDIAN__
vcgeq_f16(float16x8_t __p0,float16x8_t __p1)34051 __ai __attribute__((target("fullfp16"))) uint16x8_t vcgeq_f16(float16x8_t __p0, float16x8_t __p1) {
34052   uint16x8_t __ret;
34053   __ret = (uint16x8_t)(__p0 >= __p1);
34054   return __ret;
34055 }
34056 #else
vcgeq_f16(float16x8_t __p0,float16x8_t __p1)34057 __ai __attribute__((target("fullfp16"))) uint16x8_t vcgeq_f16(float16x8_t __p0, float16x8_t __p1) {
34058   uint16x8_t __ret;
34059   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34060   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
34061   __ret = (uint16x8_t)(__rev0 >= __rev1);
34062   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34063   return __ret;
34064 }
34065 #endif
34066 
34067 #ifdef __LITTLE_ENDIAN__
vcge_f16(float16x4_t __p0,float16x4_t __p1)34068 __ai __attribute__((target("fullfp16"))) uint16x4_t vcge_f16(float16x4_t __p0, float16x4_t __p1) {
34069   uint16x4_t __ret;
34070   __ret = (uint16x4_t)(__p0 >= __p1);
34071   return __ret;
34072 }
34073 #else
vcge_f16(float16x4_t __p0,float16x4_t __p1)34074 __ai __attribute__((target("fullfp16"))) uint16x4_t vcge_f16(float16x4_t __p0, float16x4_t __p1) {
34075   uint16x4_t __ret;
34076   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34077   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34078   __ret = (uint16x4_t)(__rev0 >= __rev1);
34079   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34080   return __ret;
34081 }
34082 #endif
34083 
34084 #ifdef __LITTLE_ENDIAN__
vcgezq_f16(float16x8_t __p0)34085 __ai __attribute__((target("fullfp16"))) uint16x8_t vcgezq_f16(float16x8_t __p0) {
34086   uint16x8_t __ret;
34087   __ret = (uint16x8_t) __builtin_neon_vcgezq_f16((int8x16_t)__p0, 49);
34088   return __ret;
34089 }
34090 #else
vcgezq_f16(float16x8_t __p0)34091 __ai __attribute__((target("fullfp16"))) uint16x8_t vcgezq_f16(float16x8_t __p0) {
34092   uint16x8_t __ret;
34093   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34094   __ret = (uint16x8_t) __builtin_neon_vcgezq_f16((int8x16_t)__rev0, 49);
34095   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34096   return __ret;
34097 }
34098 #endif
34099 
34100 #ifdef __LITTLE_ENDIAN__
vcgez_f16(float16x4_t __p0)34101 __ai __attribute__((target("fullfp16"))) uint16x4_t vcgez_f16(float16x4_t __p0) {
34102   uint16x4_t __ret;
34103   __ret = (uint16x4_t) __builtin_neon_vcgez_f16((int8x8_t)__p0, 17);
34104   return __ret;
34105 }
34106 #else
vcgez_f16(float16x4_t __p0)34107 __ai __attribute__((target("fullfp16"))) uint16x4_t vcgez_f16(float16x4_t __p0) {
34108   uint16x4_t __ret;
34109   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34110   __ret = (uint16x4_t) __builtin_neon_vcgez_f16((int8x8_t)__rev0, 17);
34111   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34112   return __ret;
34113 }
34114 #endif
34115 
34116 #ifdef __LITTLE_ENDIAN__
vcgtq_f16(float16x8_t __p0,float16x8_t __p1)34117 __ai __attribute__((target("fullfp16"))) uint16x8_t vcgtq_f16(float16x8_t __p0, float16x8_t __p1) {
34118   uint16x8_t __ret;
34119   __ret = (uint16x8_t)(__p0 > __p1);
34120   return __ret;
34121 }
34122 #else
vcgtq_f16(float16x8_t __p0,float16x8_t __p1)34123 __ai __attribute__((target("fullfp16"))) uint16x8_t vcgtq_f16(float16x8_t __p0, float16x8_t __p1) {
34124   uint16x8_t __ret;
34125   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34126   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
34127   __ret = (uint16x8_t)(__rev0 > __rev1);
34128   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34129   return __ret;
34130 }
34131 #endif
34132 
34133 #ifdef __LITTLE_ENDIAN__
vcgt_f16(float16x4_t __p0,float16x4_t __p1)34134 __ai __attribute__((target("fullfp16"))) uint16x4_t vcgt_f16(float16x4_t __p0, float16x4_t __p1) {
34135   uint16x4_t __ret;
34136   __ret = (uint16x4_t)(__p0 > __p1);
34137   return __ret;
34138 }
34139 #else
vcgt_f16(float16x4_t __p0,float16x4_t __p1)34140 __ai __attribute__((target("fullfp16"))) uint16x4_t vcgt_f16(float16x4_t __p0, float16x4_t __p1) {
34141   uint16x4_t __ret;
34142   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34143   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34144   __ret = (uint16x4_t)(__rev0 > __rev1);
34145   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34146   return __ret;
34147 }
34148 #endif
34149 
34150 #ifdef __LITTLE_ENDIAN__
vcgtzq_f16(float16x8_t __p0)34151 __ai __attribute__((target("fullfp16"))) uint16x8_t vcgtzq_f16(float16x8_t __p0) {
34152   uint16x8_t __ret;
34153   __ret = (uint16x8_t) __builtin_neon_vcgtzq_f16((int8x16_t)__p0, 49);
34154   return __ret;
34155 }
34156 #else
vcgtzq_f16(float16x8_t __p0)34157 __ai __attribute__((target("fullfp16"))) uint16x8_t vcgtzq_f16(float16x8_t __p0) {
34158   uint16x8_t __ret;
34159   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34160   __ret = (uint16x8_t) __builtin_neon_vcgtzq_f16((int8x16_t)__rev0, 49);
34161   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34162   return __ret;
34163 }
34164 #endif
34165 
34166 #ifdef __LITTLE_ENDIAN__
vcgtz_f16(float16x4_t __p0)34167 __ai __attribute__((target("fullfp16"))) uint16x4_t vcgtz_f16(float16x4_t __p0) {
34168   uint16x4_t __ret;
34169   __ret = (uint16x4_t) __builtin_neon_vcgtz_f16((int8x8_t)__p0, 17);
34170   return __ret;
34171 }
34172 #else
vcgtz_f16(float16x4_t __p0)34173 __ai __attribute__((target("fullfp16"))) uint16x4_t vcgtz_f16(float16x4_t __p0) {
34174   uint16x4_t __ret;
34175   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34176   __ret = (uint16x4_t) __builtin_neon_vcgtz_f16((int8x8_t)__rev0, 17);
34177   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34178   return __ret;
34179 }
34180 #endif
34181 
34182 #ifdef __LITTLE_ENDIAN__
vcleq_f16(float16x8_t __p0,float16x8_t __p1)34183 __ai __attribute__((target("fullfp16"))) uint16x8_t vcleq_f16(float16x8_t __p0, float16x8_t __p1) {
34184   uint16x8_t __ret;
34185   __ret = (uint16x8_t)(__p0 <= __p1);
34186   return __ret;
34187 }
34188 #else
vcleq_f16(float16x8_t __p0,float16x8_t __p1)34189 __ai __attribute__((target("fullfp16"))) uint16x8_t vcleq_f16(float16x8_t __p0, float16x8_t __p1) {
34190   uint16x8_t __ret;
34191   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34192   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
34193   __ret = (uint16x8_t)(__rev0 <= __rev1);
34194   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34195   return __ret;
34196 }
34197 #endif
34198 
34199 #ifdef __LITTLE_ENDIAN__
vcle_f16(float16x4_t __p0,float16x4_t __p1)34200 __ai __attribute__((target("fullfp16"))) uint16x4_t vcle_f16(float16x4_t __p0, float16x4_t __p1) {
34201   uint16x4_t __ret;
34202   __ret = (uint16x4_t)(__p0 <= __p1);
34203   return __ret;
34204 }
34205 #else
vcle_f16(float16x4_t __p0,float16x4_t __p1)34206 __ai __attribute__((target("fullfp16"))) uint16x4_t vcle_f16(float16x4_t __p0, float16x4_t __p1) {
34207   uint16x4_t __ret;
34208   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34209   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34210   __ret = (uint16x4_t)(__rev0 <= __rev1);
34211   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34212   return __ret;
34213 }
34214 #endif
34215 
34216 #ifdef __LITTLE_ENDIAN__
vclezq_f16(float16x8_t __p0)34217 __ai __attribute__((target("fullfp16"))) uint16x8_t vclezq_f16(float16x8_t __p0) {
34218   uint16x8_t __ret;
34219   __ret = (uint16x8_t) __builtin_neon_vclezq_f16((int8x16_t)__p0, 49);
34220   return __ret;
34221 }
34222 #else
vclezq_f16(float16x8_t __p0)34223 __ai __attribute__((target("fullfp16"))) uint16x8_t vclezq_f16(float16x8_t __p0) {
34224   uint16x8_t __ret;
34225   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34226   __ret = (uint16x8_t) __builtin_neon_vclezq_f16((int8x16_t)__rev0, 49);
34227   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34228   return __ret;
34229 }
34230 #endif
34231 
34232 #ifdef __LITTLE_ENDIAN__
vclez_f16(float16x4_t __p0)34233 __ai __attribute__((target("fullfp16"))) uint16x4_t vclez_f16(float16x4_t __p0) {
34234   uint16x4_t __ret;
34235   __ret = (uint16x4_t) __builtin_neon_vclez_f16((int8x8_t)__p0, 17);
34236   return __ret;
34237 }
34238 #else
vclez_f16(float16x4_t __p0)34239 __ai __attribute__((target("fullfp16"))) uint16x4_t vclez_f16(float16x4_t __p0) {
34240   uint16x4_t __ret;
34241   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34242   __ret = (uint16x4_t) __builtin_neon_vclez_f16((int8x8_t)__rev0, 17);
34243   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34244   return __ret;
34245 }
34246 #endif
34247 
34248 #ifdef __LITTLE_ENDIAN__
vcltq_f16(float16x8_t __p0,float16x8_t __p1)34249 __ai __attribute__((target("fullfp16"))) uint16x8_t vcltq_f16(float16x8_t __p0, float16x8_t __p1) {
34250   uint16x8_t __ret;
34251   __ret = (uint16x8_t)(__p0 < __p1);
34252   return __ret;
34253 }
34254 #else
vcltq_f16(float16x8_t __p0,float16x8_t __p1)34255 __ai __attribute__((target("fullfp16"))) uint16x8_t vcltq_f16(float16x8_t __p0, float16x8_t __p1) {
34256   uint16x8_t __ret;
34257   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34258   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
34259   __ret = (uint16x8_t)(__rev0 < __rev1);
34260   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34261   return __ret;
34262 }
34263 #endif
34264 
34265 #ifdef __LITTLE_ENDIAN__
vclt_f16(float16x4_t __p0,float16x4_t __p1)34266 __ai __attribute__((target("fullfp16"))) uint16x4_t vclt_f16(float16x4_t __p0, float16x4_t __p1) {
34267   uint16x4_t __ret;
34268   __ret = (uint16x4_t)(__p0 < __p1);
34269   return __ret;
34270 }
34271 #else
vclt_f16(float16x4_t __p0,float16x4_t __p1)34272 __ai __attribute__((target("fullfp16"))) uint16x4_t vclt_f16(float16x4_t __p0, float16x4_t __p1) {
34273   uint16x4_t __ret;
34274   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34275   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34276   __ret = (uint16x4_t)(__rev0 < __rev1);
34277   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34278   return __ret;
34279 }
34280 #endif
34281 
34282 #ifdef __LITTLE_ENDIAN__
vcltzq_f16(float16x8_t __p0)34283 __ai __attribute__((target("fullfp16"))) uint16x8_t vcltzq_f16(float16x8_t __p0) {
34284   uint16x8_t __ret;
34285   __ret = (uint16x8_t) __builtin_neon_vcltzq_f16((int8x16_t)__p0, 49);
34286   return __ret;
34287 }
34288 #else
vcltzq_f16(float16x8_t __p0)34289 __ai __attribute__((target("fullfp16"))) uint16x8_t vcltzq_f16(float16x8_t __p0) {
34290   uint16x8_t __ret;
34291   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34292   __ret = (uint16x8_t) __builtin_neon_vcltzq_f16((int8x16_t)__rev0, 49);
34293   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34294   return __ret;
34295 }
34296 #endif
34297 
34298 #ifdef __LITTLE_ENDIAN__
vcltz_f16(float16x4_t __p0)34299 __ai __attribute__((target("fullfp16"))) uint16x4_t vcltz_f16(float16x4_t __p0) {
34300   uint16x4_t __ret;
34301   __ret = (uint16x4_t) __builtin_neon_vcltz_f16((int8x8_t)__p0, 17);
34302   return __ret;
34303 }
34304 #else
vcltz_f16(float16x4_t __p0)34305 __ai __attribute__((target("fullfp16"))) uint16x4_t vcltz_f16(float16x4_t __p0) {
34306   uint16x4_t __ret;
34307   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34308   __ret = (uint16x4_t) __builtin_neon_vcltz_f16((int8x8_t)__rev0, 17);
34309   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34310   return __ret;
34311 }
34312 #endif
34313 
34314 #ifdef __LITTLE_ENDIAN__
vcvtq_f16_u16(uint16x8_t __p0)34315 __ai __attribute__((target("fullfp16"))) float16x8_t vcvtq_f16_u16(uint16x8_t __p0) {
34316   float16x8_t __ret;
34317   __ret = (float16x8_t) __builtin_neon_vcvtq_f16_u16((int8x16_t)__p0, 49);
34318   return __ret;
34319 }
34320 #else
vcvtq_f16_u16(uint16x8_t __p0)34321 __ai __attribute__((target("fullfp16"))) float16x8_t vcvtq_f16_u16(uint16x8_t __p0) {
34322   float16x8_t __ret;
34323   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34324   __ret = (float16x8_t) __builtin_neon_vcvtq_f16_u16((int8x16_t)__rev0, 49);
34325   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34326   return __ret;
34327 }
34328 #endif
34329 
34330 #ifdef __LITTLE_ENDIAN__
vcvtq_f16_s16(int16x8_t __p0)34331 __ai __attribute__((target("fullfp16"))) float16x8_t vcvtq_f16_s16(int16x8_t __p0) {
34332   float16x8_t __ret;
34333   __ret = (float16x8_t) __builtin_neon_vcvtq_f16_s16((int8x16_t)__p0, 33);
34334   return __ret;
34335 }
34336 #else
vcvtq_f16_s16(int16x8_t __p0)34337 __ai __attribute__((target("fullfp16"))) float16x8_t vcvtq_f16_s16(int16x8_t __p0) {
34338   float16x8_t __ret;
34339   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34340   __ret = (float16x8_t) __builtin_neon_vcvtq_f16_s16((int8x16_t)__rev0, 33);
34341   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34342   return __ret;
34343 }
34344 #endif
34345 
34346 #ifdef __LITTLE_ENDIAN__
vcvt_f16_u16(uint16x4_t __p0)34347 __ai __attribute__((target("fullfp16"))) float16x4_t vcvt_f16_u16(uint16x4_t __p0) {
34348   float16x4_t __ret;
34349   __ret = (float16x4_t) __builtin_neon_vcvt_f16_u16((int8x8_t)__p0, 17);
34350   return __ret;
34351 }
34352 #else
vcvt_f16_u16(uint16x4_t __p0)34353 __ai __attribute__((target("fullfp16"))) float16x4_t vcvt_f16_u16(uint16x4_t __p0) {
34354   float16x4_t __ret;
34355   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34356   __ret = (float16x4_t) __builtin_neon_vcvt_f16_u16((int8x8_t)__rev0, 17);
34357   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34358   return __ret;
34359 }
34360 #endif
34361 
34362 #ifdef __LITTLE_ENDIAN__
vcvt_f16_s16(int16x4_t __p0)34363 __ai __attribute__((target("fullfp16"))) float16x4_t vcvt_f16_s16(int16x4_t __p0) {
34364   float16x4_t __ret;
34365   __ret = (float16x4_t) __builtin_neon_vcvt_f16_s16((int8x8_t)__p0, 1);
34366   return __ret;
34367 }
34368 #else
vcvt_f16_s16(int16x4_t __p0)34369 __ai __attribute__((target("fullfp16"))) float16x4_t vcvt_f16_s16(int16x4_t __p0) {
34370   float16x4_t __ret;
34371   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34372   __ret = (float16x4_t) __builtin_neon_vcvt_f16_s16((int8x8_t)__rev0, 1);
34373   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34374   return __ret;
34375 }
34376 #endif
34377 
34378 #ifdef __LITTLE_ENDIAN__
34379 #define vcvtq_n_f16_u16(__p0, __p1) __extension__ ({ \
34380   float16x8_t __ret; \
34381   uint16x8_t __s0 = __p0; \
34382   __ret = (float16x8_t) __builtin_neon_vcvtq_n_f16_u16((int8x16_t)__s0, __p1, 49); \
34383   __ret; \
34384 })
34385 #else
34386 #define vcvtq_n_f16_u16(__p0, __p1) __extension__ ({ \
34387   float16x8_t __ret; \
34388   uint16x8_t __s0 = __p0; \
34389   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
34390   __ret = (float16x8_t) __builtin_neon_vcvtq_n_f16_u16((int8x16_t)__rev0, __p1, 49); \
34391   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
34392   __ret; \
34393 })
34394 #endif
34395 
34396 #ifdef __LITTLE_ENDIAN__
34397 #define vcvtq_n_f16_s16(__p0, __p1) __extension__ ({ \
34398   float16x8_t __ret; \
34399   int16x8_t __s0 = __p0; \
34400   __ret = (float16x8_t) __builtin_neon_vcvtq_n_f16_s16((int8x16_t)__s0, __p1, 33); \
34401   __ret; \
34402 })
34403 #else
34404 #define vcvtq_n_f16_s16(__p0, __p1) __extension__ ({ \
34405   float16x8_t __ret; \
34406   int16x8_t __s0 = __p0; \
34407   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
34408   __ret = (float16x8_t) __builtin_neon_vcvtq_n_f16_s16((int8x16_t)__rev0, __p1, 33); \
34409   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
34410   __ret; \
34411 })
34412 #endif
34413 
34414 #ifdef __LITTLE_ENDIAN__
34415 #define vcvt_n_f16_u16(__p0, __p1) __extension__ ({ \
34416   float16x4_t __ret; \
34417   uint16x4_t __s0 = __p0; \
34418   __ret = (float16x4_t) __builtin_neon_vcvt_n_f16_u16((int8x8_t)__s0, __p1, 17); \
34419   __ret; \
34420 })
34421 #else
34422 #define vcvt_n_f16_u16(__p0, __p1) __extension__ ({ \
34423   float16x4_t __ret; \
34424   uint16x4_t __s0 = __p0; \
34425   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
34426   __ret = (float16x4_t) __builtin_neon_vcvt_n_f16_u16((int8x8_t)__rev0, __p1, 17); \
34427   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
34428   __ret; \
34429 })
34430 #endif
34431 
34432 #ifdef __LITTLE_ENDIAN__
34433 #define vcvt_n_f16_s16(__p0, __p1) __extension__ ({ \
34434   float16x4_t __ret; \
34435   int16x4_t __s0 = __p0; \
34436   __ret = (float16x4_t) __builtin_neon_vcvt_n_f16_s16((int8x8_t)__s0, __p1, 1); \
34437   __ret; \
34438 })
34439 #else
34440 #define vcvt_n_f16_s16(__p0, __p1) __extension__ ({ \
34441   float16x4_t __ret; \
34442   int16x4_t __s0 = __p0; \
34443   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
34444   __ret = (float16x4_t) __builtin_neon_vcvt_n_f16_s16((int8x8_t)__rev0, __p1, 1); \
34445   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
34446   __ret; \
34447 })
34448 #endif
34449 
34450 #ifdef __LITTLE_ENDIAN__
34451 #define vcvtq_n_s16_f16(__p0, __p1) __extension__ ({ \
34452   int16x8_t __ret; \
34453   float16x8_t __s0 = __p0; \
34454   __ret = (int16x8_t) __builtin_neon_vcvtq_n_s16_f16((int8x16_t)__s0, __p1, 33); \
34455   __ret; \
34456 })
34457 #else
34458 #define vcvtq_n_s16_f16(__p0, __p1) __extension__ ({ \
34459   int16x8_t __ret; \
34460   float16x8_t __s0 = __p0; \
34461   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
34462   __ret = (int16x8_t) __builtin_neon_vcvtq_n_s16_f16((int8x16_t)__rev0, __p1, 33); \
34463   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
34464   __ret; \
34465 })
34466 #endif
34467 
34468 #ifdef __LITTLE_ENDIAN__
34469 #define vcvt_n_s16_f16(__p0, __p1) __extension__ ({ \
34470   int16x4_t __ret; \
34471   float16x4_t __s0 = __p0; \
34472   __ret = (int16x4_t) __builtin_neon_vcvt_n_s16_f16((int8x8_t)__s0, __p1, 1); \
34473   __ret; \
34474 })
34475 #else
34476 #define vcvt_n_s16_f16(__p0, __p1) __extension__ ({ \
34477   int16x4_t __ret; \
34478   float16x4_t __s0 = __p0; \
34479   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
34480   __ret = (int16x4_t) __builtin_neon_vcvt_n_s16_f16((int8x8_t)__rev0, __p1, 1); \
34481   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
34482   __ret; \
34483 })
34484 #endif
34485 
34486 #ifdef __LITTLE_ENDIAN__
34487 #define vcvtq_n_u16_f16(__p0, __p1) __extension__ ({ \
34488   uint16x8_t __ret; \
34489   float16x8_t __s0 = __p0; \
34490   __ret = (uint16x8_t) __builtin_neon_vcvtq_n_u16_f16((int8x16_t)__s0, __p1, 49); \
34491   __ret; \
34492 })
34493 #else
34494 #define vcvtq_n_u16_f16(__p0, __p1) __extension__ ({ \
34495   uint16x8_t __ret; \
34496   float16x8_t __s0 = __p0; \
34497   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
34498   __ret = (uint16x8_t) __builtin_neon_vcvtq_n_u16_f16((int8x16_t)__rev0, __p1, 49); \
34499   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
34500   __ret; \
34501 })
34502 #endif
34503 
34504 #ifdef __LITTLE_ENDIAN__
34505 #define vcvt_n_u16_f16(__p0, __p1) __extension__ ({ \
34506   uint16x4_t __ret; \
34507   float16x4_t __s0 = __p0; \
34508   __ret = (uint16x4_t) __builtin_neon_vcvt_n_u16_f16((int8x8_t)__s0, __p1, 17); \
34509   __ret; \
34510 })
34511 #else
34512 #define vcvt_n_u16_f16(__p0, __p1) __extension__ ({ \
34513   uint16x4_t __ret; \
34514   float16x4_t __s0 = __p0; \
34515   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
34516   __ret = (uint16x4_t) __builtin_neon_vcvt_n_u16_f16((int8x8_t)__rev0, __p1, 17); \
34517   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
34518   __ret; \
34519 })
34520 #endif
34521 
34522 #ifdef __LITTLE_ENDIAN__
vcvtq_s16_f16(float16x8_t __p0)34523 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtq_s16_f16(float16x8_t __p0) {
34524   int16x8_t __ret;
34525   __ret = (int16x8_t) __builtin_neon_vcvtq_s16_f16((int8x16_t)__p0, 33);
34526   return __ret;
34527 }
34528 #else
vcvtq_s16_f16(float16x8_t __p0)34529 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtq_s16_f16(float16x8_t __p0) {
34530   int16x8_t __ret;
34531   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34532   __ret = (int16x8_t) __builtin_neon_vcvtq_s16_f16((int8x16_t)__rev0, 33);
34533   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34534   return __ret;
34535 }
34536 #endif
34537 
34538 #ifdef __LITTLE_ENDIAN__
vcvt_s16_f16(float16x4_t __p0)34539 __ai __attribute__((target("fullfp16"))) int16x4_t vcvt_s16_f16(float16x4_t __p0) {
34540   int16x4_t __ret;
34541   __ret = (int16x4_t) __builtin_neon_vcvt_s16_f16((int8x8_t)__p0, 1);
34542   return __ret;
34543 }
34544 #else
vcvt_s16_f16(float16x4_t __p0)34545 __ai __attribute__((target("fullfp16"))) int16x4_t vcvt_s16_f16(float16x4_t __p0) {
34546   int16x4_t __ret;
34547   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34548   __ret = (int16x4_t) __builtin_neon_vcvt_s16_f16((int8x8_t)__rev0, 1);
34549   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34550   return __ret;
34551 }
34552 #endif
34553 
34554 #ifdef __LITTLE_ENDIAN__
vcvtq_u16_f16(float16x8_t __p0)34555 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtq_u16_f16(float16x8_t __p0) {
34556   uint16x8_t __ret;
34557   __ret = (uint16x8_t) __builtin_neon_vcvtq_u16_f16((int8x16_t)__p0, 49);
34558   return __ret;
34559 }
34560 #else
vcvtq_u16_f16(float16x8_t __p0)34561 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtq_u16_f16(float16x8_t __p0) {
34562   uint16x8_t __ret;
34563   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34564   __ret = (uint16x8_t) __builtin_neon_vcvtq_u16_f16((int8x16_t)__rev0, 49);
34565   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34566   return __ret;
34567 }
34568 #endif
34569 
34570 #ifdef __LITTLE_ENDIAN__
vcvt_u16_f16(float16x4_t __p0)34571 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvt_u16_f16(float16x4_t __p0) {
34572   uint16x4_t __ret;
34573   __ret = (uint16x4_t) __builtin_neon_vcvt_u16_f16((int8x8_t)__p0, 17);
34574   return __ret;
34575 }
34576 #else
vcvt_u16_f16(float16x4_t __p0)34577 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvt_u16_f16(float16x4_t __p0) {
34578   uint16x4_t __ret;
34579   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34580   __ret = (uint16x4_t) __builtin_neon_vcvt_u16_f16((int8x8_t)__rev0, 17);
34581   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34582   return __ret;
34583 }
34584 #endif
34585 
34586 #ifdef __LITTLE_ENDIAN__
vcvtaq_s16_f16(float16x8_t __p0)34587 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtaq_s16_f16(float16x8_t __p0) {
34588   int16x8_t __ret;
34589   __ret = (int16x8_t) __builtin_neon_vcvtaq_s16_f16((int8x16_t)__p0, 33);
34590   return __ret;
34591 }
34592 #else
vcvtaq_s16_f16(float16x8_t __p0)34593 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtaq_s16_f16(float16x8_t __p0) {
34594   int16x8_t __ret;
34595   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34596   __ret = (int16x8_t) __builtin_neon_vcvtaq_s16_f16((int8x16_t)__rev0, 33);
34597   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34598   return __ret;
34599 }
34600 #endif
34601 
34602 #ifdef __LITTLE_ENDIAN__
vcvta_s16_f16(float16x4_t __p0)34603 __ai __attribute__((target("fullfp16"))) int16x4_t vcvta_s16_f16(float16x4_t __p0) {
34604   int16x4_t __ret;
34605   __ret = (int16x4_t) __builtin_neon_vcvta_s16_f16((int8x8_t)__p0, 1);
34606   return __ret;
34607 }
34608 #else
vcvta_s16_f16(float16x4_t __p0)34609 __ai __attribute__((target("fullfp16"))) int16x4_t vcvta_s16_f16(float16x4_t __p0) {
34610   int16x4_t __ret;
34611   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34612   __ret = (int16x4_t) __builtin_neon_vcvta_s16_f16((int8x8_t)__rev0, 1);
34613   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34614   return __ret;
34615 }
34616 #endif
34617 
34618 #ifdef __LITTLE_ENDIAN__
vcvtaq_u16_f16(float16x8_t __p0)34619 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtaq_u16_f16(float16x8_t __p0) {
34620   uint16x8_t __ret;
34621   __ret = (uint16x8_t) __builtin_neon_vcvtaq_u16_f16((int8x16_t)__p0, 49);
34622   return __ret;
34623 }
34624 #else
vcvtaq_u16_f16(float16x8_t __p0)34625 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtaq_u16_f16(float16x8_t __p0) {
34626   uint16x8_t __ret;
34627   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34628   __ret = (uint16x8_t) __builtin_neon_vcvtaq_u16_f16((int8x16_t)__rev0, 49);
34629   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34630   return __ret;
34631 }
34632 #endif
34633 
34634 #ifdef __LITTLE_ENDIAN__
vcvta_u16_f16(float16x4_t __p0)34635 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvta_u16_f16(float16x4_t __p0) {
34636   uint16x4_t __ret;
34637   __ret = (uint16x4_t) __builtin_neon_vcvta_u16_f16((int8x8_t)__p0, 17);
34638   return __ret;
34639 }
34640 #else
vcvta_u16_f16(float16x4_t __p0)34641 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvta_u16_f16(float16x4_t __p0) {
34642   uint16x4_t __ret;
34643   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34644   __ret = (uint16x4_t) __builtin_neon_vcvta_u16_f16((int8x8_t)__rev0, 17);
34645   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34646   return __ret;
34647 }
34648 #endif
34649 
34650 #ifdef __LITTLE_ENDIAN__
vcvtmq_s16_f16(float16x8_t __p0)34651 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtmq_s16_f16(float16x8_t __p0) {
34652   int16x8_t __ret;
34653   __ret = (int16x8_t) __builtin_neon_vcvtmq_s16_f16((int8x16_t)__p0, 33);
34654   return __ret;
34655 }
34656 #else
vcvtmq_s16_f16(float16x8_t __p0)34657 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtmq_s16_f16(float16x8_t __p0) {
34658   int16x8_t __ret;
34659   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34660   __ret = (int16x8_t) __builtin_neon_vcvtmq_s16_f16((int8x16_t)__rev0, 33);
34661   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34662   return __ret;
34663 }
34664 #endif
34665 
34666 #ifdef __LITTLE_ENDIAN__
vcvtm_s16_f16(float16x4_t __p0)34667 __ai __attribute__((target("fullfp16"))) int16x4_t vcvtm_s16_f16(float16x4_t __p0) {
34668   int16x4_t __ret;
34669   __ret = (int16x4_t) __builtin_neon_vcvtm_s16_f16((int8x8_t)__p0, 1);
34670   return __ret;
34671 }
34672 #else
vcvtm_s16_f16(float16x4_t __p0)34673 __ai __attribute__((target("fullfp16"))) int16x4_t vcvtm_s16_f16(float16x4_t __p0) {
34674   int16x4_t __ret;
34675   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34676   __ret = (int16x4_t) __builtin_neon_vcvtm_s16_f16((int8x8_t)__rev0, 1);
34677   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34678   return __ret;
34679 }
34680 #endif
34681 
34682 #ifdef __LITTLE_ENDIAN__
vcvtmq_u16_f16(float16x8_t __p0)34683 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtmq_u16_f16(float16x8_t __p0) {
34684   uint16x8_t __ret;
34685   __ret = (uint16x8_t) __builtin_neon_vcvtmq_u16_f16((int8x16_t)__p0, 49);
34686   return __ret;
34687 }
34688 #else
vcvtmq_u16_f16(float16x8_t __p0)34689 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtmq_u16_f16(float16x8_t __p0) {
34690   uint16x8_t __ret;
34691   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34692   __ret = (uint16x8_t) __builtin_neon_vcvtmq_u16_f16((int8x16_t)__rev0, 49);
34693   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34694   return __ret;
34695 }
34696 #endif
34697 
34698 #ifdef __LITTLE_ENDIAN__
vcvtm_u16_f16(float16x4_t __p0)34699 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvtm_u16_f16(float16x4_t __p0) {
34700   uint16x4_t __ret;
34701   __ret = (uint16x4_t) __builtin_neon_vcvtm_u16_f16((int8x8_t)__p0, 17);
34702   return __ret;
34703 }
34704 #else
vcvtm_u16_f16(float16x4_t __p0)34705 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvtm_u16_f16(float16x4_t __p0) {
34706   uint16x4_t __ret;
34707   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34708   __ret = (uint16x4_t) __builtin_neon_vcvtm_u16_f16((int8x8_t)__rev0, 17);
34709   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34710   return __ret;
34711 }
34712 #endif
34713 
34714 #ifdef __LITTLE_ENDIAN__
vcvtnq_s16_f16(float16x8_t __p0)34715 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtnq_s16_f16(float16x8_t __p0) {
34716   int16x8_t __ret;
34717   __ret = (int16x8_t) __builtin_neon_vcvtnq_s16_f16((int8x16_t)__p0, 33);
34718   return __ret;
34719 }
34720 #else
vcvtnq_s16_f16(float16x8_t __p0)34721 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtnq_s16_f16(float16x8_t __p0) {
34722   int16x8_t __ret;
34723   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34724   __ret = (int16x8_t) __builtin_neon_vcvtnq_s16_f16((int8x16_t)__rev0, 33);
34725   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34726   return __ret;
34727 }
34728 #endif
34729 
34730 #ifdef __LITTLE_ENDIAN__
vcvtn_s16_f16(float16x4_t __p0)34731 __ai __attribute__((target("fullfp16"))) int16x4_t vcvtn_s16_f16(float16x4_t __p0) {
34732   int16x4_t __ret;
34733   __ret = (int16x4_t) __builtin_neon_vcvtn_s16_f16((int8x8_t)__p0, 1);
34734   return __ret;
34735 }
34736 #else
vcvtn_s16_f16(float16x4_t __p0)34737 __ai __attribute__((target("fullfp16"))) int16x4_t vcvtn_s16_f16(float16x4_t __p0) {
34738   int16x4_t __ret;
34739   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34740   __ret = (int16x4_t) __builtin_neon_vcvtn_s16_f16((int8x8_t)__rev0, 1);
34741   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34742   return __ret;
34743 }
34744 #endif
34745 
34746 #ifdef __LITTLE_ENDIAN__
vcvtnq_u16_f16(float16x8_t __p0)34747 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtnq_u16_f16(float16x8_t __p0) {
34748   uint16x8_t __ret;
34749   __ret = (uint16x8_t) __builtin_neon_vcvtnq_u16_f16((int8x16_t)__p0, 49);
34750   return __ret;
34751 }
34752 #else
vcvtnq_u16_f16(float16x8_t __p0)34753 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtnq_u16_f16(float16x8_t __p0) {
34754   uint16x8_t __ret;
34755   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34756   __ret = (uint16x8_t) __builtin_neon_vcvtnq_u16_f16((int8x16_t)__rev0, 49);
34757   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34758   return __ret;
34759 }
34760 #endif
34761 
34762 #ifdef __LITTLE_ENDIAN__
vcvtn_u16_f16(float16x4_t __p0)34763 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvtn_u16_f16(float16x4_t __p0) {
34764   uint16x4_t __ret;
34765   __ret = (uint16x4_t) __builtin_neon_vcvtn_u16_f16((int8x8_t)__p0, 17);
34766   return __ret;
34767 }
34768 #else
vcvtn_u16_f16(float16x4_t __p0)34769 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvtn_u16_f16(float16x4_t __p0) {
34770   uint16x4_t __ret;
34771   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34772   __ret = (uint16x4_t) __builtin_neon_vcvtn_u16_f16((int8x8_t)__rev0, 17);
34773   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34774   return __ret;
34775 }
34776 #endif
34777 
34778 #ifdef __LITTLE_ENDIAN__
vcvtpq_s16_f16(float16x8_t __p0)34779 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtpq_s16_f16(float16x8_t __p0) {
34780   int16x8_t __ret;
34781   __ret = (int16x8_t) __builtin_neon_vcvtpq_s16_f16((int8x16_t)__p0, 33);
34782   return __ret;
34783 }
34784 #else
vcvtpq_s16_f16(float16x8_t __p0)34785 __ai __attribute__((target("fullfp16"))) int16x8_t vcvtpq_s16_f16(float16x8_t __p0) {
34786   int16x8_t __ret;
34787   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34788   __ret = (int16x8_t) __builtin_neon_vcvtpq_s16_f16((int8x16_t)__rev0, 33);
34789   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34790   return __ret;
34791 }
34792 #endif
34793 
34794 #ifdef __LITTLE_ENDIAN__
vcvtp_s16_f16(float16x4_t __p0)34795 __ai __attribute__((target("fullfp16"))) int16x4_t vcvtp_s16_f16(float16x4_t __p0) {
34796   int16x4_t __ret;
34797   __ret = (int16x4_t) __builtin_neon_vcvtp_s16_f16((int8x8_t)__p0, 1);
34798   return __ret;
34799 }
34800 #else
vcvtp_s16_f16(float16x4_t __p0)34801 __ai __attribute__((target("fullfp16"))) int16x4_t vcvtp_s16_f16(float16x4_t __p0) {
34802   int16x4_t __ret;
34803   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34804   __ret = (int16x4_t) __builtin_neon_vcvtp_s16_f16((int8x8_t)__rev0, 1);
34805   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34806   return __ret;
34807 }
34808 #endif
34809 
34810 #ifdef __LITTLE_ENDIAN__
vcvtpq_u16_f16(float16x8_t __p0)34811 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtpq_u16_f16(float16x8_t __p0) {
34812   uint16x8_t __ret;
34813   __ret = (uint16x8_t) __builtin_neon_vcvtpq_u16_f16((int8x16_t)__p0, 49);
34814   return __ret;
34815 }
34816 #else
vcvtpq_u16_f16(float16x8_t __p0)34817 __ai __attribute__((target("fullfp16"))) uint16x8_t vcvtpq_u16_f16(float16x8_t __p0) {
34818   uint16x8_t __ret;
34819   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34820   __ret = (uint16x8_t) __builtin_neon_vcvtpq_u16_f16((int8x16_t)__rev0, 49);
34821   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34822   return __ret;
34823 }
34824 #endif
34825 
34826 #ifdef __LITTLE_ENDIAN__
vcvtp_u16_f16(float16x4_t __p0)34827 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvtp_u16_f16(float16x4_t __p0) {
34828   uint16x4_t __ret;
34829   __ret = (uint16x4_t) __builtin_neon_vcvtp_u16_f16((int8x8_t)__p0, 17);
34830   return __ret;
34831 }
34832 #else
vcvtp_u16_f16(float16x4_t __p0)34833 __ai __attribute__((target("fullfp16"))) uint16x4_t vcvtp_u16_f16(float16x4_t __p0) {
34834   uint16x4_t __ret;
34835   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34836   __ret = (uint16x4_t) __builtin_neon_vcvtp_u16_f16((int8x8_t)__rev0, 17);
34837   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34838   return __ret;
34839 }
34840 #endif
34841 
34842 #ifdef __LITTLE_ENDIAN__
34843 #define vextq_f16(__p0, __p1, __p2) __extension__ ({ \
34844   float16x8_t __ret; \
34845   float16x8_t __s0 = __p0; \
34846   float16x8_t __s1 = __p1; \
34847   __ret = (float16x8_t) __builtin_neon_vextq_f16((int8x16_t)__s0, (int8x16_t)__s1, __p2, 40); \
34848   __ret; \
34849 })
34850 #else
34851 #define vextq_f16(__p0, __p1, __p2) __extension__ ({ \
34852   float16x8_t __ret; \
34853   float16x8_t __s0 = __p0; \
34854   float16x8_t __s1 = __p1; \
34855   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
34856   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
34857   __ret = (float16x8_t) __builtin_neon_vextq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 40); \
34858   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
34859   __ret; \
34860 })
34861 #endif
34862 
34863 #ifdef __LITTLE_ENDIAN__
34864 #define vext_f16(__p0, __p1, __p2) __extension__ ({ \
34865   float16x4_t __ret; \
34866   float16x4_t __s0 = __p0; \
34867   float16x4_t __s1 = __p1; \
34868   __ret = (float16x4_t) __builtin_neon_vext_f16((int8x8_t)__s0, (int8x8_t)__s1, __p2, 8); \
34869   __ret; \
34870 })
34871 #else
34872 #define vext_f16(__p0, __p1, __p2) __extension__ ({ \
34873   float16x4_t __ret; \
34874   float16x4_t __s0 = __p0; \
34875   float16x4_t __s1 = __p1; \
34876   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
34877   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
34878   __ret = (float16x4_t) __builtin_neon_vext_f16((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 8); \
34879   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
34880   __ret; \
34881 })
34882 #endif
34883 
34884 #ifdef __LITTLE_ENDIAN__
vfmaq_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)34885 __ai __attribute__((target("fullfp16"))) float16x8_t vfmaq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
34886   float16x8_t __ret;
34887   __ret = (float16x8_t) __builtin_neon_vfmaq_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
34888   return __ret;
34889 }
34890 #else
vfmaq_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)34891 __ai __attribute__((target("fullfp16"))) float16x8_t vfmaq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
34892   float16x8_t __ret;
34893   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34894   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
34895   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
34896   __ret = (float16x8_t) __builtin_neon_vfmaq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 40);
34897   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34898   return __ret;
34899 }
__noswap_vfmaq_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)34900 __ai __attribute__((target("fullfp16"))) float16x8_t __noswap_vfmaq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
34901   float16x8_t __ret;
34902   __ret = (float16x8_t) __builtin_neon_vfmaq_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
34903   return __ret;
34904 }
34905 #endif
34906 
34907 #ifdef __LITTLE_ENDIAN__
vfma_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)34908 __ai __attribute__((target("fullfp16"))) float16x4_t vfma_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
34909   float16x4_t __ret;
34910   __ret = (float16x4_t) __builtin_neon_vfma_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
34911   return __ret;
34912 }
34913 #else
vfma_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)34914 __ai __attribute__((target("fullfp16"))) float16x4_t vfma_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
34915   float16x4_t __ret;
34916   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34917   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34918   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
34919   __ret = (float16x4_t) __builtin_neon_vfma_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 8);
34920   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34921   return __ret;
34922 }
__noswap_vfma_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)34923 __ai __attribute__((target("fullfp16"))) float16x4_t __noswap_vfma_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
34924   float16x4_t __ret;
34925   __ret = (float16x4_t) __builtin_neon_vfma_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
34926   return __ret;
34927 }
34928 #endif
34929 
34930 #ifdef __LITTLE_ENDIAN__
vfmsq_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)34931 __ai __attribute__((target("fullfp16"))) float16x8_t vfmsq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
34932   float16x8_t __ret;
34933   __ret = vfmaq_f16(__p0, -__p1, __p2);
34934   return __ret;
34935 }
34936 #else
vfmsq_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)34937 __ai __attribute__((target("fullfp16"))) float16x8_t vfmsq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
34938   float16x8_t __ret;
34939   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34940   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
34941   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
34942   __ret = __noswap_vfmaq_f16(__rev0, -__rev1, __rev2);
34943   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34944   return __ret;
34945 }
34946 #endif
34947 
34948 #ifdef __LITTLE_ENDIAN__
vfms_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)34949 __ai __attribute__((target("fullfp16"))) float16x4_t vfms_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
34950   float16x4_t __ret;
34951   __ret = vfma_f16(__p0, -__p1, __p2);
34952   return __ret;
34953 }
34954 #else
vfms_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)34955 __ai __attribute__((target("fullfp16"))) float16x4_t vfms_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
34956   float16x4_t __ret;
34957   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34958   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34959   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
34960   __ret = __noswap_vfma_f16(__rev0, -__rev1, __rev2);
34961   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34962   return __ret;
34963 }
34964 #endif
34965 
34966 #ifdef __LITTLE_ENDIAN__
vmaxq_f16(float16x8_t __p0,float16x8_t __p1)34967 __ai __attribute__((target("fullfp16"))) float16x8_t vmaxq_f16(float16x8_t __p0, float16x8_t __p1) {
34968   float16x8_t __ret;
34969   __ret = (float16x8_t) __builtin_neon_vmaxq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
34970   return __ret;
34971 }
34972 #else
vmaxq_f16(float16x8_t __p0,float16x8_t __p1)34973 __ai __attribute__((target("fullfp16"))) float16x8_t vmaxq_f16(float16x8_t __p0, float16x8_t __p1) {
34974   float16x8_t __ret;
34975   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
34976   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
34977   __ret = (float16x8_t) __builtin_neon_vmaxq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
34978   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
34979   return __ret;
34980 }
34981 #endif
34982 
34983 #ifdef __LITTLE_ENDIAN__
vmax_f16(float16x4_t __p0,float16x4_t __p1)34984 __ai __attribute__((target("fullfp16"))) float16x4_t vmax_f16(float16x4_t __p0, float16x4_t __p1) {
34985   float16x4_t __ret;
34986   __ret = (float16x4_t) __builtin_neon_vmax_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
34987   return __ret;
34988 }
34989 #else
vmax_f16(float16x4_t __p0,float16x4_t __p1)34990 __ai __attribute__((target("fullfp16"))) float16x4_t vmax_f16(float16x4_t __p0, float16x4_t __p1) {
34991   float16x4_t __ret;
34992   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34993   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34994   __ret = (float16x4_t) __builtin_neon_vmax_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
34995   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34996   return __ret;
34997 }
34998 #endif
34999 
35000 #ifdef __LITTLE_ENDIAN__
vminq_f16(float16x8_t __p0,float16x8_t __p1)35001 __ai __attribute__((target("fullfp16"))) float16x8_t vminq_f16(float16x8_t __p0, float16x8_t __p1) {
35002   float16x8_t __ret;
35003   __ret = (float16x8_t) __builtin_neon_vminq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
35004   return __ret;
35005 }
35006 #else
vminq_f16(float16x8_t __p0,float16x8_t __p1)35007 __ai __attribute__((target("fullfp16"))) float16x8_t vminq_f16(float16x8_t __p0, float16x8_t __p1) {
35008   float16x8_t __ret;
35009   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35010   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35011   __ret = (float16x8_t) __builtin_neon_vminq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
35012   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35013   return __ret;
35014 }
35015 #endif
35016 
35017 #ifdef __LITTLE_ENDIAN__
vmin_f16(float16x4_t __p0,float16x4_t __p1)35018 __ai __attribute__((target("fullfp16"))) float16x4_t vmin_f16(float16x4_t __p0, float16x4_t __p1) {
35019   float16x4_t __ret;
35020   __ret = (float16x4_t) __builtin_neon_vmin_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
35021   return __ret;
35022 }
35023 #else
vmin_f16(float16x4_t __p0,float16x4_t __p1)35024 __ai __attribute__((target("fullfp16"))) float16x4_t vmin_f16(float16x4_t __p0, float16x4_t __p1) {
35025   float16x4_t __ret;
35026   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35027   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35028   __ret = (float16x4_t) __builtin_neon_vmin_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35029   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35030   return __ret;
35031 }
35032 #endif
35033 
35034 #ifdef __LITTLE_ENDIAN__
vmulq_f16(float16x8_t __p0,float16x8_t __p1)35035 __ai __attribute__((target("fullfp16"))) float16x8_t vmulq_f16(float16x8_t __p0, float16x8_t __p1) {
35036   float16x8_t __ret;
35037   __ret = __p0 * __p1;
35038   return __ret;
35039 }
35040 #else
vmulq_f16(float16x8_t __p0,float16x8_t __p1)35041 __ai __attribute__((target("fullfp16"))) float16x8_t vmulq_f16(float16x8_t __p0, float16x8_t __p1) {
35042   float16x8_t __ret;
35043   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35044   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35045   __ret = __rev0 * __rev1;
35046   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35047   return __ret;
35048 }
35049 #endif
35050 
35051 #ifdef __LITTLE_ENDIAN__
vmul_f16(float16x4_t __p0,float16x4_t __p1)35052 __ai __attribute__((target("fullfp16"))) float16x4_t vmul_f16(float16x4_t __p0, float16x4_t __p1) {
35053   float16x4_t __ret;
35054   __ret = __p0 * __p1;
35055   return __ret;
35056 }
35057 #else
vmul_f16(float16x4_t __p0,float16x4_t __p1)35058 __ai __attribute__((target("fullfp16"))) float16x4_t vmul_f16(float16x4_t __p0, float16x4_t __p1) {
35059   float16x4_t __ret;
35060   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35061   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35062   __ret = __rev0 * __rev1;
35063   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35064   return __ret;
35065 }
35066 #endif
35067 
35068 #ifdef __LITTLE_ENDIAN__
35069 #define vmulq_lane_f16(__p0_153, __p1_153, __p2_153) __extension__ ({ \
35070   float16x8_t __ret_153; \
35071   float16x8_t __s0_153 = __p0_153; \
35072   float16x4_t __s1_153 = __p1_153; \
35073   __ret_153 = __s0_153 * splatq_lane_f16(__s1_153, __p2_153); \
35074   __ret_153; \
35075 })
35076 #else
35077 #define vmulq_lane_f16(__p0_154, __p1_154, __p2_154) __extension__ ({ \
35078   float16x8_t __ret_154; \
35079   float16x8_t __s0_154 = __p0_154; \
35080   float16x4_t __s1_154 = __p1_154; \
35081   float16x8_t __rev0_154;  __rev0_154 = __builtin_shufflevector(__s0_154, __s0_154, 7, 6, 5, 4, 3, 2, 1, 0); \
35082   float16x4_t __rev1_154;  __rev1_154 = __builtin_shufflevector(__s1_154, __s1_154, 3, 2, 1, 0); \
35083   __ret_154 = __rev0_154 * __noswap_splatq_lane_f16(__rev1_154, __p2_154); \
35084   __ret_154 = __builtin_shufflevector(__ret_154, __ret_154, 7, 6, 5, 4, 3, 2, 1, 0); \
35085   __ret_154; \
35086 })
35087 #endif
35088 
35089 #ifdef __LITTLE_ENDIAN__
35090 #define vmul_lane_f16(__p0_155, __p1_155, __p2_155) __extension__ ({ \
35091   float16x4_t __ret_155; \
35092   float16x4_t __s0_155 = __p0_155; \
35093   float16x4_t __s1_155 = __p1_155; \
35094   __ret_155 = __s0_155 * splat_lane_f16(__s1_155, __p2_155); \
35095   __ret_155; \
35096 })
35097 #else
35098 #define vmul_lane_f16(__p0_156, __p1_156, __p2_156) __extension__ ({ \
35099   float16x4_t __ret_156; \
35100   float16x4_t __s0_156 = __p0_156; \
35101   float16x4_t __s1_156 = __p1_156; \
35102   float16x4_t __rev0_156;  __rev0_156 = __builtin_shufflevector(__s0_156, __s0_156, 3, 2, 1, 0); \
35103   float16x4_t __rev1_156;  __rev1_156 = __builtin_shufflevector(__s1_156, __s1_156, 3, 2, 1, 0); \
35104   __ret_156 = __rev0_156 * __noswap_splat_lane_f16(__rev1_156, __p2_156); \
35105   __ret_156 = __builtin_shufflevector(__ret_156, __ret_156, 3, 2, 1, 0); \
35106   __ret_156; \
35107 })
35108 #endif
35109 
35110 #ifdef __LITTLE_ENDIAN__
35111 #define vmulq_n_f16(__p0, __p1) __extension__ ({ \
35112   float16x8_t __ret; \
35113   float16x8_t __s0 = __p0; \
35114   float16_t __s1 = __p1; \
35115   __ret = __s0 * (float16x8_t) {__s1, __s1, __s1, __s1, __s1, __s1, __s1, __s1}; \
35116   __ret; \
35117 })
35118 #else
35119 #define vmulq_n_f16(__p0, __p1) __extension__ ({ \
35120   float16x8_t __ret; \
35121   float16x8_t __s0 = __p0; \
35122   float16_t __s1 = __p1; \
35123   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
35124   __ret = __rev0 * (float16x8_t) {__s1, __s1, __s1, __s1, __s1, __s1, __s1, __s1}; \
35125   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
35126   __ret; \
35127 })
35128 #endif
35129 
35130 #ifdef __LITTLE_ENDIAN__
35131 #define vmul_n_f16(__p0, __p1) __extension__ ({ \
35132   float16x4_t __ret; \
35133   float16x4_t __s0 = __p0; \
35134   float16_t __s1 = __p1; \
35135   __ret = __s0 * (float16x4_t) {__s1, __s1, __s1, __s1}; \
35136   __ret; \
35137 })
35138 #else
35139 #define vmul_n_f16(__p0, __p1) __extension__ ({ \
35140   float16x4_t __ret; \
35141   float16x4_t __s0 = __p0; \
35142   float16_t __s1 = __p1; \
35143   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
35144   __ret = __rev0 * (float16x4_t) {__s1, __s1, __s1, __s1}; \
35145   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
35146   __ret; \
35147 })
35148 #endif
35149 
35150 #ifdef __LITTLE_ENDIAN__
vnegq_f16(float16x8_t __p0)35151 __ai __attribute__((target("fullfp16"))) float16x8_t vnegq_f16(float16x8_t __p0) {
35152   float16x8_t __ret;
35153   __ret = -__p0;
35154   return __ret;
35155 }
35156 #else
vnegq_f16(float16x8_t __p0)35157 __ai __attribute__((target("fullfp16"))) float16x8_t vnegq_f16(float16x8_t __p0) {
35158   float16x8_t __ret;
35159   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35160   __ret = -__rev0;
35161   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35162   return __ret;
35163 }
35164 #endif
35165 
35166 #ifdef __LITTLE_ENDIAN__
vneg_f16(float16x4_t __p0)35167 __ai __attribute__((target("fullfp16"))) float16x4_t vneg_f16(float16x4_t __p0) {
35168   float16x4_t __ret;
35169   __ret = -__p0;
35170   return __ret;
35171 }
35172 #else
vneg_f16(float16x4_t __p0)35173 __ai __attribute__((target("fullfp16"))) float16x4_t vneg_f16(float16x4_t __p0) {
35174   float16x4_t __ret;
35175   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35176   __ret = -__rev0;
35177   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35178   return __ret;
35179 }
35180 #endif
35181 
35182 #ifdef __LITTLE_ENDIAN__
vpadd_f16(float16x4_t __p0,float16x4_t __p1)35183 __ai __attribute__((target("fullfp16"))) float16x4_t vpadd_f16(float16x4_t __p0, float16x4_t __p1) {
35184   float16x4_t __ret;
35185   __ret = (float16x4_t) __builtin_neon_vpadd_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
35186   return __ret;
35187 }
35188 #else
vpadd_f16(float16x4_t __p0,float16x4_t __p1)35189 __ai __attribute__((target("fullfp16"))) float16x4_t vpadd_f16(float16x4_t __p0, float16x4_t __p1) {
35190   float16x4_t __ret;
35191   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35192   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35193   __ret = (float16x4_t) __builtin_neon_vpadd_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35194   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35195   return __ret;
35196 }
35197 #endif
35198 
35199 #ifdef __LITTLE_ENDIAN__
vpmax_f16(float16x4_t __p0,float16x4_t __p1)35200 __ai __attribute__((target("fullfp16"))) float16x4_t vpmax_f16(float16x4_t __p0, float16x4_t __p1) {
35201   float16x4_t __ret;
35202   __ret = (float16x4_t) __builtin_neon_vpmax_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
35203   return __ret;
35204 }
35205 #else
vpmax_f16(float16x4_t __p0,float16x4_t __p1)35206 __ai __attribute__((target("fullfp16"))) float16x4_t vpmax_f16(float16x4_t __p0, float16x4_t __p1) {
35207   float16x4_t __ret;
35208   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35209   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35210   __ret = (float16x4_t) __builtin_neon_vpmax_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35211   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35212   return __ret;
35213 }
35214 #endif
35215 
35216 #ifdef __LITTLE_ENDIAN__
vpmin_f16(float16x4_t __p0,float16x4_t __p1)35217 __ai __attribute__((target("fullfp16"))) float16x4_t vpmin_f16(float16x4_t __p0, float16x4_t __p1) {
35218   float16x4_t __ret;
35219   __ret = (float16x4_t) __builtin_neon_vpmin_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
35220   return __ret;
35221 }
35222 #else
vpmin_f16(float16x4_t __p0,float16x4_t __p1)35223 __ai __attribute__((target("fullfp16"))) float16x4_t vpmin_f16(float16x4_t __p0, float16x4_t __p1) {
35224   float16x4_t __ret;
35225   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35226   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35227   __ret = (float16x4_t) __builtin_neon_vpmin_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35228   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35229   return __ret;
35230 }
35231 #endif
35232 
35233 #ifdef __LITTLE_ENDIAN__
vrecpeq_f16(float16x8_t __p0)35234 __ai __attribute__((target("fullfp16"))) float16x8_t vrecpeq_f16(float16x8_t __p0) {
35235   float16x8_t __ret;
35236   __ret = (float16x8_t) __builtin_neon_vrecpeq_f16((int8x16_t)__p0, 40);
35237   return __ret;
35238 }
35239 #else
vrecpeq_f16(float16x8_t __p0)35240 __ai __attribute__((target("fullfp16"))) float16x8_t vrecpeq_f16(float16x8_t __p0) {
35241   float16x8_t __ret;
35242   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35243   __ret = (float16x8_t) __builtin_neon_vrecpeq_f16((int8x16_t)__rev0, 40);
35244   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35245   return __ret;
35246 }
35247 #endif
35248 
35249 #ifdef __LITTLE_ENDIAN__
vrecpe_f16(float16x4_t __p0)35250 __ai __attribute__((target("fullfp16"))) float16x4_t vrecpe_f16(float16x4_t __p0) {
35251   float16x4_t __ret;
35252   __ret = (float16x4_t) __builtin_neon_vrecpe_f16((int8x8_t)__p0, 8);
35253   return __ret;
35254 }
35255 #else
vrecpe_f16(float16x4_t __p0)35256 __ai __attribute__((target("fullfp16"))) float16x4_t vrecpe_f16(float16x4_t __p0) {
35257   float16x4_t __ret;
35258   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35259   __ret = (float16x4_t) __builtin_neon_vrecpe_f16((int8x8_t)__rev0, 8);
35260   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35261   return __ret;
35262 }
35263 #endif
35264 
35265 #ifdef __LITTLE_ENDIAN__
vrecpsq_f16(float16x8_t __p0,float16x8_t __p1)35266 __ai __attribute__((target("fullfp16"))) float16x8_t vrecpsq_f16(float16x8_t __p0, float16x8_t __p1) {
35267   float16x8_t __ret;
35268   __ret = (float16x8_t) __builtin_neon_vrecpsq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
35269   return __ret;
35270 }
35271 #else
vrecpsq_f16(float16x8_t __p0,float16x8_t __p1)35272 __ai __attribute__((target("fullfp16"))) float16x8_t vrecpsq_f16(float16x8_t __p0, float16x8_t __p1) {
35273   float16x8_t __ret;
35274   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35275   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35276   __ret = (float16x8_t) __builtin_neon_vrecpsq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
35277   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35278   return __ret;
35279 }
35280 #endif
35281 
35282 #ifdef __LITTLE_ENDIAN__
vrecps_f16(float16x4_t __p0,float16x4_t __p1)35283 __ai __attribute__((target("fullfp16"))) float16x4_t vrecps_f16(float16x4_t __p0, float16x4_t __p1) {
35284   float16x4_t __ret;
35285   __ret = (float16x4_t) __builtin_neon_vrecps_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
35286   return __ret;
35287 }
35288 #else
vrecps_f16(float16x4_t __p0,float16x4_t __p1)35289 __ai __attribute__((target("fullfp16"))) float16x4_t vrecps_f16(float16x4_t __p0, float16x4_t __p1) {
35290   float16x4_t __ret;
35291   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35292   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35293   __ret = (float16x4_t) __builtin_neon_vrecps_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35294   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35295   return __ret;
35296 }
35297 #endif
35298 
35299 #ifdef __LITTLE_ENDIAN__
vrev64q_f16(float16x8_t __p0)35300 __ai __attribute__((target("fullfp16"))) float16x8_t vrev64q_f16(float16x8_t __p0) {
35301   float16x8_t __ret;
35302   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
35303   return __ret;
35304 }
35305 #else
vrev64q_f16(float16x8_t __p0)35306 __ai __attribute__((target("fullfp16"))) float16x8_t vrev64q_f16(float16x8_t __p0) {
35307   float16x8_t __ret;
35308   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35309   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
35310   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35311   return __ret;
35312 }
35313 #endif
35314 
35315 #ifdef __LITTLE_ENDIAN__
vrev64_f16(float16x4_t __p0)35316 __ai __attribute__((target("fullfp16"))) float16x4_t vrev64_f16(float16x4_t __p0) {
35317   float16x4_t __ret;
35318   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35319   return __ret;
35320 }
35321 #else
vrev64_f16(float16x4_t __p0)35322 __ai __attribute__((target("fullfp16"))) float16x4_t vrev64_f16(float16x4_t __p0) {
35323   float16x4_t __ret;
35324   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35325   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
35326   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35327   return __ret;
35328 }
35329 #endif
35330 
35331 #ifdef __LITTLE_ENDIAN__
vrsqrteq_f16(float16x8_t __p0)35332 __ai __attribute__((target("fullfp16"))) float16x8_t vrsqrteq_f16(float16x8_t __p0) {
35333   float16x8_t __ret;
35334   __ret = (float16x8_t) __builtin_neon_vrsqrteq_f16((int8x16_t)__p0, 40);
35335   return __ret;
35336 }
35337 #else
vrsqrteq_f16(float16x8_t __p0)35338 __ai __attribute__((target("fullfp16"))) float16x8_t vrsqrteq_f16(float16x8_t __p0) {
35339   float16x8_t __ret;
35340   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35341   __ret = (float16x8_t) __builtin_neon_vrsqrteq_f16((int8x16_t)__rev0, 40);
35342   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35343   return __ret;
35344 }
35345 #endif
35346 
35347 #ifdef __LITTLE_ENDIAN__
vrsqrte_f16(float16x4_t __p0)35348 __ai __attribute__((target("fullfp16"))) float16x4_t vrsqrte_f16(float16x4_t __p0) {
35349   float16x4_t __ret;
35350   __ret = (float16x4_t) __builtin_neon_vrsqrte_f16((int8x8_t)__p0, 8);
35351   return __ret;
35352 }
35353 #else
vrsqrte_f16(float16x4_t __p0)35354 __ai __attribute__((target("fullfp16"))) float16x4_t vrsqrte_f16(float16x4_t __p0) {
35355   float16x4_t __ret;
35356   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35357   __ret = (float16x4_t) __builtin_neon_vrsqrte_f16((int8x8_t)__rev0, 8);
35358   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35359   return __ret;
35360 }
35361 #endif
35362 
35363 #ifdef __LITTLE_ENDIAN__
vrsqrtsq_f16(float16x8_t __p0,float16x8_t __p1)35364 __ai __attribute__((target("fullfp16"))) float16x8_t vrsqrtsq_f16(float16x8_t __p0, float16x8_t __p1) {
35365   float16x8_t __ret;
35366   __ret = (float16x8_t) __builtin_neon_vrsqrtsq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
35367   return __ret;
35368 }
35369 #else
vrsqrtsq_f16(float16x8_t __p0,float16x8_t __p1)35370 __ai __attribute__((target("fullfp16"))) float16x8_t vrsqrtsq_f16(float16x8_t __p0, float16x8_t __p1) {
35371   float16x8_t __ret;
35372   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35373   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35374   __ret = (float16x8_t) __builtin_neon_vrsqrtsq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
35375   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35376   return __ret;
35377 }
35378 #endif
35379 
35380 #ifdef __LITTLE_ENDIAN__
vrsqrts_f16(float16x4_t __p0,float16x4_t __p1)35381 __ai __attribute__((target("fullfp16"))) float16x4_t vrsqrts_f16(float16x4_t __p0, float16x4_t __p1) {
35382   float16x4_t __ret;
35383   __ret = (float16x4_t) __builtin_neon_vrsqrts_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
35384   return __ret;
35385 }
35386 #else
vrsqrts_f16(float16x4_t __p0,float16x4_t __p1)35387 __ai __attribute__((target("fullfp16"))) float16x4_t vrsqrts_f16(float16x4_t __p0, float16x4_t __p1) {
35388   float16x4_t __ret;
35389   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35390   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35391   __ret = (float16x4_t) __builtin_neon_vrsqrts_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35392   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35393   return __ret;
35394 }
35395 #endif
35396 
35397 #ifdef __LITTLE_ENDIAN__
vsubq_f16(float16x8_t __p0,float16x8_t __p1)35398 __ai __attribute__((target("fullfp16"))) float16x8_t vsubq_f16(float16x8_t __p0, float16x8_t __p1) {
35399   float16x8_t __ret;
35400   __ret = __p0 - __p1;
35401   return __ret;
35402 }
35403 #else
vsubq_f16(float16x8_t __p0,float16x8_t __p1)35404 __ai __attribute__((target("fullfp16"))) float16x8_t vsubq_f16(float16x8_t __p0, float16x8_t __p1) {
35405   float16x8_t __ret;
35406   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35407   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35408   __ret = __rev0 - __rev1;
35409   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35410   return __ret;
35411 }
35412 #endif
35413 
35414 #ifdef __LITTLE_ENDIAN__
vsub_f16(float16x4_t __p0,float16x4_t __p1)35415 __ai __attribute__((target("fullfp16"))) float16x4_t vsub_f16(float16x4_t __p0, float16x4_t __p1) {
35416   float16x4_t __ret;
35417   __ret = __p0 - __p1;
35418   return __ret;
35419 }
35420 #else
vsub_f16(float16x4_t __p0,float16x4_t __p1)35421 __ai __attribute__((target("fullfp16"))) float16x4_t vsub_f16(float16x4_t __p0, float16x4_t __p1) {
35422   float16x4_t __ret;
35423   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35424   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35425   __ret = __rev0 - __rev1;
35426   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35427   return __ret;
35428 }
35429 #endif
35430 
35431 #ifdef __LITTLE_ENDIAN__
vtrnq_f16(float16x8_t __p0,float16x8_t __p1)35432 __ai __attribute__((target("fullfp16"))) float16x8x2_t vtrnq_f16(float16x8_t __p0, float16x8_t __p1) {
35433   float16x8x2_t __ret;
35434   __builtin_neon_vtrnq_f16(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 40);
35435   return __ret;
35436 }
35437 #else
vtrnq_f16(float16x8_t __p0,float16x8_t __p1)35438 __ai __attribute__((target("fullfp16"))) float16x8x2_t vtrnq_f16(float16x8_t __p0, float16x8_t __p1) {
35439   float16x8x2_t __ret;
35440   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35441   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35442   __builtin_neon_vtrnq_f16(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 40);
35443 
35444   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
35445   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
35446   return __ret;
35447 }
35448 #endif
35449 
35450 #ifdef __LITTLE_ENDIAN__
vtrn_f16(float16x4_t __p0,float16x4_t __p1)35451 __ai __attribute__((target("fullfp16"))) float16x4x2_t vtrn_f16(float16x4_t __p0, float16x4_t __p1) {
35452   float16x4x2_t __ret;
35453   __builtin_neon_vtrn_f16(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 8);
35454   return __ret;
35455 }
35456 #else
vtrn_f16(float16x4_t __p0,float16x4_t __p1)35457 __ai __attribute__((target("fullfp16"))) float16x4x2_t vtrn_f16(float16x4_t __p0, float16x4_t __p1) {
35458   float16x4x2_t __ret;
35459   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35460   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35461   __builtin_neon_vtrn_f16(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35462 
35463   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
35464   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
35465   return __ret;
35466 }
35467 #endif
35468 
35469 #ifdef __LITTLE_ENDIAN__
vuzpq_f16(float16x8_t __p0,float16x8_t __p1)35470 __ai __attribute__((target("fullfp16"))) float16x8x2_t vuzpq_f16(float16x8_t __p0, float16x8_t __p1) {
35471   float16x8x2_t __ret;
35472   __builtin_neon_vuzpq_f16(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 40);
35473   return __ret;
35474 }
35475 #else
vuzpq_f16(float16x8_t __p0,float16x8_t __p1)35476 __ai __attribute__((target("fullfp16"))) float16x8x2_t vuzpq_f16(float16x8_t __p0, float16x8_t __p1) {
35477   float16x8x2_t __ret;
35478   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35479   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35480   __builtin_neon_vuzpq_f16(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 40);
35481 
35482   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
35483   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
35484   return __ret;
35485 }
35486 #endif
35487 
35488 #ifdef __LITTLE_ENDIAN__
vuzp_f16(float16x4_t __p0,float16x4_t __p1)35489 __ai __attribute__((target("fullfp16"))) float16x4x2_t vuzp_f16(float16x4_t __p0, float16x4_t __p1) {
35490   float16x4x2_t __ret;
35491   __builtin_neon_vuzp_f16(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 8);
35492   return __ret;
35493 }
35494 #else
vuzp_f16(float16x4_t __p0,float16x4_t __p1)35495 __ai __attribute__((target("fullfp16"))) float16x4x2_t vuzp_f16(float16x4_t __p0, float16x4_t __p1) {
35496   float16x4x2_t __ret;
35497   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35498   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35499   __builtin_neon_vuzp_f16(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35500 
35501   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
35502   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
35503   return __ret;
35504 }
35505 #endif
35506 
35507 #ifdef __LITTLE_ENDIAN__
vzipq_f16(float16x8_t __p0,float16x8_t __p1)35508 __ai __attribute__((target("fullfp16"))) float16x8x2_t vzipq_f16(float16x8_t __p0, float16x8_t __p1) {
35509   float16x8x2_t __ret;
35510   __builtin_neon_vzipq_f16(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 40);
35511   return __ret;
35512 }
35513 #else
vzipq_f16(float16x8_t __p0,float16x8_t __p1)35514 __ai __attribute__((target("fullfp16"))) float16x8x2_t vzipq_f16(float16x8_t __p0, float16x8_t __p1) {
35515   float16x8x2_t __ret;
35516   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35517   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35518   __builtin_neon_vzipq_f16(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 40);
35519 
35520   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
35521   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
35522   return __ret;
35523 }
35524 #endif
35525 
35526 #ifdef __LITTLE_ENDIAN__
vzip_f16(float16x4_t __p0,float16x4_t __p1)35527 __ai __attribute__((target("fullfp16"))) float16x4x2_t vzip_f16(float16x4_t __p0, float16x4_t __p1) {
35528   float16x4x2_t __ret;
35529   __builtin_neon_vzip_f16(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 8);
35530   return __ret;
35531 }
35532 #else
vzip_f16(float16x4_t __p0,float16x4_t __p1)35533 __ai __attribute__((target("fullfp16"))) float16x4x2_t vzip_f16(float16x4_t __p0, float16x4_t __p1) {
35534   float16x4x2_t __ret;
35535   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35536   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35537   __builtin_neon_vzip_f16(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35538 
35539   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
35540   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
35541   return __ret;
35542 }
35543 #endif
35544 
35545 #ifdef __LITTLE_ENDIAN__
vmmlaq_u32(uint32x4_t __p0,uint8x16_t __p1,uint8x16_t __p2)35546 __ai __attribute__((target("i8mm"))) uint32x4_t vmmlaq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
35547   uint32x4_t __ret;
35548   __ret = (uint32x4_t) __builtin_neon_vmmlaq_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
35549   return __ret;
35550 }
35551 #else
vmmlaq_u32(uint32x4_t __p0,uint8x16_t __p1,uint8x16_t __p2)35552 __ai __attribute__((target("i8mm"))) uint32x4_t vmmlaq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
35553   uint32x4_t __ret;
35554   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35555   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
35556   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
35557   __ret = (uint32x4_t) __builtin_neon_vmmlaq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
35558   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35559   return __ret;
35560 }
35561 #endif
35562 
35563 #ifdef __LITTLE_ENDIAN__
vmmlaq_s32(int32x4_t __p0,int8x16_t __p1,int8x16_t __p2)35564 __ai __attribute__((target("i8mm"))) int32x4_t vmmlaq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
35565   int32x4_t __ret;
35566   __ret = (int32x4_t) __builtin_neon_vmmlaq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
35567   return __ret;
35568 }
35569 #else
vmmlaq_s32(int32x4_t __p0,int8x16_t __p1,int8x16_t __p2)35570 __ai __attribute__((target("i8mm"))) int32x4_t vmmlaq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
35571   int32x4_t __ret;
35572   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35573   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
35574   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
35575   __ret = (int32x4_t) __builtin_neon_vmmlaq_s32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
35576   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35577   return __ret;
35578 }
35579 #endif
35580 
35581 #ifdef __LITTLE_ENDIAN__
vusdotq_s32(int32x4_t __p0,uint8x16_t __p1,int8x16_t __p2)35582 __ai __attribute__((target("i8mm"))) int32x4_t vusdotq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
35583   int32x4_t __ret;
35584   __ret = (int32x4_t) __builtin_neon_vusdotq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
35585   return __ret;
35586 }
35587 #else
vusdotq_s32(int32x4_t __p0,uint8x16_t __p1,int8x16_t __p2)35588 __ai __attribute__((target("i8mm"))) int32x4_t vusdotq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
35589   int32x4_t __ret;
35590   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35591   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
35592   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
35593   __ret = (int32x4_t) __builtin_neon_vusdotq_s32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
35594   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35595   return __ret;
35596 }
__noswap_vusdotq_s32(int32x4_t __p0,uint8x16_t __p1,int8x16_t __p2)35597 __ai __attribute__((target("i8mm"))) int32x4_t __noswap_vusdotq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
35598   int32x4_t __ret;
35599   __ret = (int32x4_t) __builtin_neon_vusdotq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
35600   return __ret;
35601 }
35602 #endif
35603 
35604 #ifdef __LITTLE_ENDIAN__
vusdot_s32(int32x2_t __p0,uint8x8_t __p1,int8x8_t __p2)35605 __ai __attribute__((target("i8mm"))) int32x2_t vusdot_s32(int32x2_t __p0, uint8x8_t __p1, int8x8_t __p2) {
35606   int32x2_t __ret;
35607   __ret = (int32x2_t) __builtin_neon_vusdot_s32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
35608   return __ret;
35609 }
35610 #else
vusdot_s32(int32x2_t __p0,uint8x8_t __p1,int8x8_t __p2)35611 __ai __attribute__((target("i8mm"))) int32x2_t vusdot_s32(int32x2_t __p0, uint8x8_t __p1, int8x8_t __p2) {
35612   int32x2_t __ret;
35613   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35614   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35615   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
35616   __ret = (int32x2_t) __builtin_neon_vusdot_s32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
35617   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35618   return __ret;
35619 }
__noswap_vusdot_s32(int32x2_t __p0,uint8x8_t __p1,int8x8_t __p2)35620 __ai __attribute__((target("i8mm"))) int32x2_t __noswap_vusdot_s32(int32x2_t __p0, uint8x8_t __p1, int8x8_t __p2) {
35621   int32x2_t __ret;
35622   __ret = (int32x2_t) __builtin_neon_vusdot_s32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
35623   return __ret;
35624 }
35625 #endif
35626 
35627 #ifdef __LITTLE_ENDIAN__
35628 #define vusdotq_lane_s32(__p0_157, __p1_157, __p2_157, __p3_157) __extension__ ({ \
35629   int32x4_t __ret_157; \
35630   int32x4_t __s0_157 = __p0_157; \
35631   uint8x16_t __s1_157 = __p1_157; \
35632   int8x8_t __s2_157 = __p2_157; \
35633 int8x8_t __reint_157 = __s2_157; \
35634   __ret_157 = vusdotq_s32(__s0_157, __s1_157, (int8x16_t)(splatq_lane_s32(*(int32x2_t *) &__reint_157, __p3_157))); \
35635   __ret_157; \
35636 })
35637 #else
35638 #define vusdotq_lane_s32(__p0_158, __p1_158, __p2_158, __p3_158) __extension__ ({ \
35639   int32x4_t __ret_158; \
35640   int32x4_t __s0_158 = __p0_158; \
35641   uint8x16_t __s1_158 = __p1_158; \
35642   int8x8_t __s2_158 = __p2_158; \
35643   int32x4_t __rev0_158;  __rev0_158 = __builtin_shufflevector(__s0_158, __s0_158, 3, 2, 1, 0); \
35644   uint8x16_t __rev1_158;  __rev1_158 = __builtin_shufflevector(__s1_158, __s1_158, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
35645   int8x8_t __rev2_158;  __rev2_158 = __builtin_shufflevector(__s2_158, __s2_158, 7, 6, 5, 4, 3, 2, 1, 0); \
35646 int8x8_t __reint_158 = __rev2_158; \
35647   __ret_158 = __noswap_vusdotq_s32(__rev0_158, __rev1_158, (int8x16_t)(__noswap_splatq_lane_s32(*(int32x2_t *) &__reint_158, __p3_158))); \
35648   __ret_158 = __builtin_shufflevector(__ret_158, __ret_158, 3, 2, 1, 0); \
35649   __ret_158; \
35650 })
35651 #endif
35652 
35653 #ifdef __LITTLE_ENDIAN__
35654 #define vusdot_lane_s32(__p0_159, __p1_159, __p2_159, __p3_159) __extension__ ({ \
35655   int32x2_t __ret_159; \
35656   int32x2_t __s0_159 = __p0_159; \
35657   uint8x8_t __s1_159 = __p1_159; \
35658   int8x8_t __s2_159 = __p2_159; \
35659 int8x8_t __reint_159 = __s2_159; \
35660   __ret_159 = vusdot_s32(__s0_159, __s1_159, (int8x8_t)(splat_lane_s32(*(int32x2_t *) &__reint_159, __p3_159))); \
35661   __ret_159; \
35662 })
35663 #else
35664 #define vusdot_lane_s32(__p0_160, __p1_160, __p2_160, __p3_160) __extension__ ({ \
35665   int32x2_t __ret_160; \
35666   int32x2_t __s0_160 = __p0_160; \
35667   uint8x8_t __s1_160 = __p1_160; \
35668   int8x8_t __s2_160 = __p2_160; \
35669   int32x2_t __rev0_160;  __rev0_160 = __builtin_shufflevector(__s0_160, __s0_160, 1, 0); \
35670   uint8x8_t __rev1_160;  __rev1_160 = __builtin_shufflevector(__s1_160, __s1_160, 7, 6, 5, 4, 3, 2, 1, 0); \
35671   int8x8_t __rev2_160;  __rev2_160 = __builtin_shufflevector(__s2_160, __s2_160, 7, 6, 5, 4, 3, 2, 1, 0); \
35672 int8x8_t __reint_160 = __rev2_160; \
35673   __ret_160 = __noswap_vusdot_s32(__rev0_160, __rev1_160, (int8x8_t)(__noswap_splat_lane_s32(*(int32x2_t *) &__reint_160, __p3_160))); \
35674   __ret_160 = __builtin_shufflevector(__ret_160, __ret_160, 1, 0); \
35675   __ret_160; \
35676 })
35677 #endif
35678 
35679 #ifdef __LITTLE_ENDIAN__
vusmmlaq_s32(int32x4_t __p0,uint8x16_t __p1,int8x16_t __p2)35680 __ai __attribute__((target("i8mm"))) int32x4_t vusmmlaq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
35681   int32x4_t __ret;
35682   __ret = (int32x4_t) __builtin_neon_vusmmlaq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
35683   return __ret;
35684 }
35685 #else
vusmmlaq_s32(int32x4_t __p0,uint8x16_t __p1,int8x16_t __p2)35686 __ai __attribute__((target("i8mm"))) int32x4_t vusmmlaq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
35687   int32x4_t __ret;
35688   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35689   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
35690   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
35691   __ret = (int32x4_t) __builtin_neon_vusmmlaq_s32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
35692   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35693   return __ret;
35694 }
35695 #endif
35696 
35697 #ifdef __LITTLE_ENDIAN__
vqrdmlahq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)35698 __ai __attribute__((target("v8.1a"))) int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
35699   int32x4_t __ret;
35700   __ret = (int32x4_t) __builtin_neon_vqrdmlahq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
35701   return __ret;
35702 }
35703 #else
vqrdmlahq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)35704 __ai __attribute__((target("v8.1a"))) int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
35705   int32x4_t __ret;
35706   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35707   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35708   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
35709   __ret = (int32x4_t) __builtin_neon_vqrdmlahq_s32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
35710   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35711   return __ret;
35712 }
__noswap_vqrdmlahq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)35713 __ai __attribute__((target("v8.1a"))) int32x4_t __noswap_vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
35714   int32x4_t __ret;
35715   __ret = (int32x4_t) __builtin_neon_vqrdmlahq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
35716   return __ret;
35717 }
35718 #endif
35719 
35720 #ifdef __LITTLE_ENDIAN__
vqrdmlahq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)35721 __ai __attribute__((target("v8.1a"))) int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
35722   int16x8_t __ret;
35723   __ret = (int16x8_t) __builtin_neon_vqrdmlahq_s16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
35724   return __ret;
35725 }
35726 #else
vqrdmlahq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)35727 __ai __attribute__((target("v8.1a"))) int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
35728   int16x8_t __ret;
35729   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35730   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35731   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
35732   __ret = (int16x8_t) __builtin_neon_vqrdmlahq_s16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
35733   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35734   return __ret;
35735 }
__noswap_vqrdmlahq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)35736 __ai __attribute__((target("v8.1a"))) int16x8_t __noswap_vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
35737   int16x8_t __ret;
35738   __ret = (int16x8_t) __builtin_neon_vqrdmlahq_s16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
35739   return __ret;
35740 }
35741 #endif
35742 
35743 #ifdef __LITTLE_ENDIAN__
vqrdmlah_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)35744 __ai __attribute__((target("v8.1a"))) int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
35745   int32x2_t __ret;
35746   __ret = (int32x2_t) __builtin_neon_vqrdmlah_s32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
35747   return __ret;
35748 }
35749 #else
vqrdmlah_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)35750 __ai __attribute__((target("v8.1a"))) int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
35751   int32x2_t __ret;
35752   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35753   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
35754   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
35755   __ret = (int32x2_t) __builtin_neon_vqrdmlah_s32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
35756   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35757   return __ret;
35758 }
__noswap_vqrdmlah_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)35759 __ai __attribute__((target("v8.1a"))) int32x2_t __noswap_vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
35760   int32x2_t __ret;
35761   __ret = (int32x2_t) __builtin_neon_vqrdmlah_s32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
35762   return __ret;
35763 }
35764 #endif
35765 
35766 #ifdef __LITTLE_ENDIAN__
vqrdmlah_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)35767 __ai __attribute__((target("v8.1a"))) int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
35768   int16x4_t __ret;
35769   __ret = (int16x4_t) __builtin_neon_vqrdmlah_s16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
35770   return __ret;
35771 }
35772 #else
vqrdmlah_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)35773 __ai __attribute__((target("v8.1a"))) int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
35774   int16x4_t __ret;
35775   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35776   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35777   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
35778   __ret = (int16x4_t) __builtin_neon_vqrdmlah_s16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 1);
35779   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35780   return __ret;
35781 }
__noswap_vqrdmlah_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)35782 __ai __attribute__((target("v8.1a"))) int16x4_t __noswap_vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
35783   int16x4_t __ret;
35784   __ret = (int16x4_t) __builtin_neon_vqrdmlah_s16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
35785   return __ret;
35786 }
35787 #endif
35788 
35789 #ifdef __LITTLE_ENDIAN__
35790 #define vqrdmlahq_lane_s32(__p0_161, __p1_161, __p2_161, __p3_161) __extension__ ({ \
35791   int32x4_t __ret_161; \
35792   int32x4_t __s0_161 = __p0_161; \
35793   int32x4_t __s1_161 = __p1_161; \
35794   int32x2_t __s2_161 = __p2_161; \
35795   __ret_161 = vqrdmlahq_s32(__s0_161, __s1_161, splatq_lane_s32(__s2_161, __p3_161)); \
35796   __ret_161; \
35797 })
35798 #else
35799 #define vqrdmlahq_lane_s32(__p0_162, __p1_162, __p2_162, __p3_162) __extension__ ({ \
35800   int32x4_t __ret_162; \
35801   int32x4_t __s0_162 = __p0_162; \
35802   int32x4_t __s1_162 = __p1_162; \
35803   int32x2_t __s2_162 = __p2_162; \
35804   int32x4_t __rev0_162;  __rev0_162 = __builtin_shufflevector(__s0_162, __s0_162, 3, 2, 1, 0); \
35805   int32x4_t __rev1_162;  __rev1_162 = __builtin_shufflevector(__s1_162, __s1_162, 3, 2, 1, 0); \
35806   int32x2_t __rev2_162;  __rev2_162 = __builtin_shufflevector(__s2_162, __s2_162, 1, 0); \
35807   __ret_162 = __noswap_vqrdmlahq_s32(__rev0_162, __rev1_162, __noswap_splatq_lane_s32(__rev2_162, __p3_162)); \
35808   __ret_162 = __builtin_shufflevector(__ret_162, __ret_162, 3, 2, 1, 0); \
35809   __ret_162; \
35810 })
35811 #endif
35812 
35813 #ifdef __LITTLE_ENDIAN__
35814 #define vqrdmlahq_lane_s16(__p0_163, __p1_163, __p2_163, __p3_163) __extension__ ({ \
35815   int16x8_t __ret_163; \
35816   int16x8_t __s0_163 = __p0_163; \
35817   int16x8_t __s1_163 = __p1_163; \
35818   int16x4_t __s2_163 = __p2_163; \
35819   __ret_163 = vqrdmlahq_s16(__s0_163, __s1_163, splatq_lane_s16(__s2_163, __p3_163)); \
35820   __ret_163; \
35821 })
35822 #else
35823 #define vqrdmlahq_lane_s16(__p0_164, __p1_164, __p2_164, __p3_164) __extension__ ({ \
35824   int16x8_t __ret_164; \
35825   int16x8_t __s0_164 = __p0_164; \
35826   int16x8_t __s1_164 = __p1_164; \
35827   int16x4_t __s2_164 = __p2_164; \
35828   int16x8_t __rev0_164;  __rev0_164 = __builtin_shufflevector(__s0_164, __s0_164, 7, 6, 5, 4, 3, 2, 1, 0); \
35829   int16x8_t __rev1_164;  __rev1_164 = __builtin_shufflevector(__s1_164, __s1_164, 7, 6, 5, 4, 3, 2, 1, 0); \
35830   int16x4_t __rev2_164;  __rev2_164 = __builtin_shufflevector(__s2_164, __s2_164, 3, 2, 1, 0); \
35831   __ret_164 = __noswap_vqrdmlahq_s16(__rev0_164, __rev1_164, __noswap_splatq_lane_s16(__rev2_164, __p3_164)); \
35832   __ret_164 = __builtin_shufflevector(__ret_164, __ret_164, 7, 6, 5, 4, 3, 2, 1, 0); \
35833   __ret_164; \
35834 })
35835 #endif
35836 
35837 #ifdef __LITTLE_ENDIAN__
35838 #define vqrdmlah_lane_s32(__p0_165, __p1_165, __p2_165, __p3_165) __extension__ ({ \
35839   int32x2_t __ret_165; \
35840   int32x2_t __s0_165 = __p0_165; \
35841   int32x2_t __s1_165 = __p1_165; \
35842   int32x2_t __s2_165 = __p2_165; \
35843   __ret_165 = vqrdmlah_s32(__s0_165, __s1_165, splat_lane_s32(__s2_165, __p3_165)); \
35844   __ret_165; \
35845 })
35846 #else
35847 #define vqrdmlah_lane_s32(__p0_166, __p1_166, __p2_166, __p3_166) __extension__ ({ \
35848   int32x2_t __ret_166; \
35849   int32x2_t __s0_166 = __p0_166; \
35850   int32x2_t __s1_166 = __p1_166; \
35851   int32x2_t __s2_166 = __p2_166; \
35852   int32x2_t __rev0_166;  __rev0_166 = __builtin_shufflevector(__s0_166, __s0_166, 1, 0); \
35853   int32x2_t __rev1_166;  __rev1_166 = __builtin_shufflevector(__s1_166, __s1_166, 1, 0); \
35854   int32x2_t __rev2_166;  __rev2_166 = __builtin_shufflevector(__s2_166, __s2_166, 1, 0); \
35855   __ret_166 = __noswap_vqrdmlah_s32(__rev0_166, __rev1_166, __noswap_splat_lane_s32(__rev2_166, __p3_166)); \
35856   __ret_166 = __builtin_shufflevector(__ret_166, __ret_166, 1, 0); \
35857   __ret_166; \
35858 })
35859 #endif
35860 
35861 #ifdef __LITTLE_ENDIAN__
35862 #define vqrdmlah_lane_s16(__p0_167, __p1_167, __p2_167, __p3_167) __extension__ ({ \
35863   int16x4_t __ret_167; \
35864   int16x4_t __s0_167 = __p0_167; \
35865   int16x4_t __s1_167 = __p1_167; \
35866   int16x4_t __s2_167 = __p2_167; \
35867   __ret_167 = vqrdmlah_s16(__s0_167, __s1_167, splat_lane_s16(__s2_167, __p3_167)); \
35868   __ret_167; \
35869 })
35870 #else
35871 #define vqrdmlah_lane_s16(__p0_168, __p1_168, __p2_168, __p3_168) __extension__ ({ \
35872   int16x4_t __ret_168; \
35873   int16x4_t __s0_168 = __p0_168; \
35874   int16x4_t __s1_168 = __p1_168; \
35875   int16x4_t __s2_168 = __p2_168; \
35876   int16x4_t __rev0_168;  __rev0_168 = __builtin_shufflevector(__s0_168, __s0_168, 3, 2, 1, 0); \
35877   int16x4_t __rev1_168;  __rev1_168 = __builtin_shufflevector(__s1_168, __s1_168, 3, 2, 1, 0); \
35878   int16x4_t __rev2_168;  __rev2_168 = __builtin_shufflevector(__s2_168, __s2_168, 3, 2, 1, 0); \
35879   __ret_168 = __noswap_vqrdmlah_s16(__rev0_168, __rev1_168, __noswap_splat_lane_s16(__rev2_168, __p3_168)); \
35880   __ret_168 = __builtin_shufflevector(__ret_168, __ret_168, 3, 2, 1, 0); \
35881   __ret_168; \
35882 })
35883 #endif
35884 
35885 #ifdef __LITTLE_ENDIAN__
vqrdmlshq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)35886 __ai __attribute__((target("v8.1a"))) int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
35887   int32x4_t __ret;
35888   __ret = (int32x4_t) __builtin_neon_vqrdmlshq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
35889   return __ret;
35890 }
35891 #else
vqrdmlshq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)35892 __ai __attribute__((target("v8.1a"))) int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
35893   int32x4_t __ret;
35894   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35895   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35896   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
35897   __ret = (int32x4_t) __builtin_neon_vqrdmlshq_s32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
35898   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35899   return __ret;
35900 }
__noswap_vqrdmlshq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)35901 __ai __attribute__((target("v8.1a"))) int32x4_t __noswap_vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
35902   int32x4_t __ret;
35903   __ret = (int32x4_t) __builtin_neon_vqrdmlshq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
35904   return __ret;
35905 }
35906 #endif
35907 
35908 #ifdef __LITTLE_ENDIAN__
vqrdmlshq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)35909 __ai __attribute__((target("v8.1a"))) int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
35910   int16x8_t __ret;
35911   __ret = (int16x8_t) __builtin_neon_vqrdmlshq_s16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
35912   return __ret;
35913 }
35914 #else
vqrdmlshq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)35915 __ai __attribute__((target("v8.1a"))) int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
35916   int16x8_t __ret;
35917   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35918   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35919   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
35920   __ret = (int16x8_t) __builtin_neon_vqrdmlshq_s16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
35921   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35922   return __ret;
35923 }
__noswap_vqrdmlshq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)35924 __ai __attribute__((target("v8.1a"))) int16x8_t __noswap_vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
35925   int16x8_t __ret;
35926   __ret = (int16x8_t) __builtin_neon_vqrdmlshq_s16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
35927   return __ret;
35928 }
35929 #endif
35930 
35931 #ifdef __LITTLE_ENDIAN__
vqrdmlsh_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)35932 __ai __attribute__((target("v8.1a"))) int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
35933   int32x2_t __ret;
35934   __ret = (int32x2_t) __builtin_neon_vqrdmlsh_s32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
35935   return __ret;
35936 }
35937 #else
vqrdmlsh_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)35938 __ai __attribute__((target("v8.1a"))) int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
35939   int32x2_t __ret;
35940   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35941   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
35942   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
35943   __ret = (int32x2_t) __builtin_neon_vqrdmlsh_s32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
35944   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35945   return __ret;
35946 }
__noswap_vqrdmlsh_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)35947 __ai __attribute__((target("v8.1a"))) int32x2_t __noswap_vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
35948   int32x2_t __ret;
35949   __ret = (int32x2_t) __builtin_neon_vqrdmlsh_s32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
35950   return __ret;
35951 }
35952 #endif
35953 
35954 #ifdef __LITTLE_ENDIAN__
vqrdmlsh_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)35955 __ai __attribute__((target("v8.1a"))) int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
35956   int16x4_t __ret;
35957   __ret = (int16x4_t) __builtin_neon_vqrdmlsh_s16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
35958   return __ret;
35959 }
35960 #else
vqrdmlsh_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)35961 __ai __attribute__((target("v8.1a"))) int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
35962   int16x4_t __ret;
35963   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35964   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35965   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
35966   __ret = (int16x4_t) __builtin_neon_vqrdmlsh_s16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 1);
35967   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35968   return __ret;
35969 }
__noswap_vqrdmlsh_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)35970 __ai __attribute__((target("v8.1a"))) int16x4_t __noswap_vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
35971   int16x4_t __ret;
35972   __ret = (int16x4_t) __builtin_neon_vqrdmlsh_s16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
35973   return __ret;
35974 }
35975 #endif
35976 
35977 #ifdef __LITTLE_ENDIAN__
35978 #define vqrdmlshq_lane_s32(__p0_169, __p1_169, __p2_169, __p3_169) __extension__ ({ \
35979   int32x4_t __ret_169; \
35980   int32x4_t __s0_169 = __p0_169; \
35981   int32x4_t __s1_169 = __p1_169; \
35982   int32x2_t __s2_169 = __p2_169; \
35983   __ret_169 = vqrdmlshq_s32(__s0_169, __s1_169, splatq_lane_s32(__s2_169, __p3_169)); \
35984   __ret_169; \
35985 })
35986 #else
35987 #define vqrdmlshq_lane_s32(__p0_170, __p1_170, __p2_170, __p3_170) __extension__ ({ \
35988   int32x4_t __ret_170; \
35989   int32x4_t __s0_170 = __p0_170; \
35990   int32x4_t __s1_170 = __p1_170; \
35991   int32x2_t __s2_170 = __p2_170; \
35992   int32x4_t __rev0_170;  __rev0_170 = __builtin_shufflevector(__s0_170, __s0_170, 3, 2, 1, 0); \
35993   int32x4_t __rev1_170;  __rev1_170 = __builtin_shufflevector(__s1_170, __s1_170, 3, 2, 1, 0); \
35994   int32x2_t __rev2_170;  __rev2_170 = __builtin_shufflevector(__s2_170, __s2_170, 1, 0); \
35995   __ret_170 = __noswap_vqrdmlshq_s32(__rev0_170, __rev1_170, __noswap_splatq_lane_s32(__rev2_170, __p3_170)); \
35996   __ret_170 = __builtin_shufflevector(__ret_170, __ret_170, 3, 2, 1, 0); \
35997   __ret_170; \
35998 })
35999 #endif
36000 
36001 #ifdef __LITTLE_ENDIAN__
36002 #define vqrdmlshq_lane_s16(__p0_171, __p1_171, __p2_171, __p3_171) __extension__ ({ \
36003   int16x8_t __ret_171; \
36004   int16x8_t __s0_171 = __p0_171; \
36005   int16x8_t __s1_171 = __p1_171; \
36006   int16x4_t __s2_171 = __p2_171; \
36007   __ret_171 = vqrdmlshq_s16(__s0_171, __s1_171, splatq_lane_s16(__s2_171, __p3_171)); \
36008   __ret_171; \
36009 })
36010 #else
36011 #define vqrdmlshq_lane_s16(__p0_172, __p1_172, __p2_172, __p3_172) __extension__ ({ \
36012   int16x8_t __ret_172; \
36013   int16x8_t __s0_172 = __p0_172; \
36014   int16x8_t __s1_172 = __p1_172; \
36015   int16x4_t __s2_172 = __p2_172; \
36016   int16x8_t __rev0_172;  __rev0_172 = __builtin_shufflevector(__s0_172, __s0_172, 7, 6, 5, 4, 3, 2, 1, 0); \
36017   int16x8_t __rev1_172;  __rev1_172 = __builtin_shufflevector(__s1_172, __s1_172, 7, 6, 5, 4, 3, 2, 1, 0); \
36018   int16x4_t __rev2_172;  __rev2_172 = __builtin_shufflevector(__s2_172, __s2_172, 3, 2, 1, 0); \
36019   __ret_172 = __noswap_vqrdmlshq_s16(__rev0_172, __rev1_172, __noswap_splatq_lane_s16(__rev2_172, __p3_172)); \
36020   __ret_172 = __builtin_shufflevector(__ret_172, __ret_172, 7, 6, 5, 4, 3, 2, 1, 0); \
36021   __ret_172; \
36022 })
36023 #endif
36024 
36025 #ifdef __LITTLE_ENDIAN__
36026 #define vqrdmlsh_lane_s32(__p0_173, __p1_173, __p2_173, __p3_173) __extension__ ({ \
36027   int32x2_t __ret_173; \
36028   int32x2_t __s0_173 = __p0_173; \
36029   int32x2_t __s1_173 = __p1_173; \
36030   int32x2_t __s2_173 = __p2_173; \
36031   __ret_173 = vqrdmlsh_s32(__s0_173, __s1_173, splat_lane_s32(__s2_173, __p3_173)); \
36032   __ret_173; \
36033 })
36034 #else
36035 #define vqrdmlsh_lane_s32(__p0_174, __p1_174, __p2_174, __p3_174) __extension__ ({ \
36036   int32x2_t __ret_174; \
36037   int32x2_t __s0_174 = __p0_174; \
36038   int32x2_t __s1_174 = __p1_174; \
36039   int32x2_t __s2_174 = __p2_174; \
36040   int32x2_t __rev0_174;  __rev0_174 = __builtin_shufflevector(__s0_174, __s0_174, 1, 0); \
36041   int32x2_t __rev1_174;  __rev1_174 = __builtin_shufflevector(__s1_174, __s1_174, 1, 0); \
36042   int32x2_t __rev2_174;  __rev2_174 = __builtin_shufflevector(__s2_174, __s2_174, 1, 0); \
36043   __ret_174 = __noswap_vqrdmlsh_s32(__rev0_174, __rev1_174, __noswap_splat_lane_s32(__rev2_174, __p3_174)); \
36044   __ret_174 = __builtin_shufflevector(__ret_174, __ret_174, 1, 0); \
36045   __ret_174; \
36046 })
36047 #endif
36048 
36049 #ifdef __LITTLE_ENDIAN__
36050 #define vqrdmlsh_lane_s16(__p0_175, __p1_175, __p2_175, __p3_175) __extension__ ({ \
36051   int16x4_t __ret_175; \
36052   int16x4_t __s0_175 = __p0_175; \
36053   int16x4_t __s1_175 = __p1_175; \
36054   int16x4_t __s2_175 = __p2_175; \
36055   __ret_175 = vqrdmlsh_s16(__s0_175, __s1_175, splat_lane_s16(__s2_175, __p3_175)); \
36056   __ret_175; \
36057 })
36058 #else
36059 #define vqrdmlsh_lane_s16(__p0_176, __p1_176, __p2_176, __p3_176) __extension__ ({ \
36060   int16x4_t __ret_176; \
36061   int16x4_t __s0_176 = __p0_176; \
36062   int16x4_t __s1_176 = __p1_176; \
36063   int16x4_t __s2_176 = __p2_176; \
36064   int16x4_t __rev0_176;  __rev0_176 = __builtin_shufflevector(__s0_176, __s0_176, 3, 2, 1, 0); \
36065   int16x4_t __rev1_176;  __rev1_176 = __builtin_shufflevector(__s1_176, __s1_176, 3, 2, 1, 0); \
36066   int16x4_t __rev2_176;  __rev2_176 = __builtin_shufflevector(__s2_176, __s2_176, 3, 2, 1, 0); \
36067   __ret_176 = __noswap_vqrdmlsh_s16(__rev0_176, __rev1_176, __noswap_splat_lane_s16(__rev2_176, __p3_176)); \
36068   __ret_176 = __builtin_shufflevector(__ret_176, __ret_176, 3, 2, 1, 0); \
36069   __ret_176; \
36070 })
36071 #endif
36072 
36073 #ifdef __LITTLE_ENDIAN__
vcadd_rot270_f32(float32x2_t __p0,float32x2_t __p1)36074 __ai __attribute__((target("v8.3a"))) float32x2_t vcadd_rot270_f32(float32x2_t __p0, float32x2_t __p1) {
36075   float32x2_t __ret;
36076   __ret = (float32x2_t) __builtin_neon_vcadd_rot270_f32((int8x8_t)__p0, (int8x8_t)__p1, 9);
36077   return __ret;
36078 }
36079 #else
vcadd_rot270_f32(float32x2_t __p0,float32x2_t __p1)36080 __ai __attribute__((target("v8.3a"))) float32x2_t vcadd_rot270_f32(float32x2_t __p0, float32x2_t __p1) {
36081   float32x2_t __ret;
36082   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
36083   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
36084   __ret = (float32x2_t) __builtin_neon_vcadd_rot270_f32((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
36085   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
36086   return __ret;
36087 }
36088 #endif
36089 
36090 #ifdef __LITTLE_ENDIAN__
vcadd_rot90_f32(float32x2_t __p0,float32x2_t __p1)36091 __ai __attribute__((target("v8.3a"))) float32x2_t vcadd_rot90_f32(float32x2_t __p0, float32x2_t __p1) {
36092   float32x2_t __ret;
36093   __ret = (float32x2_t) __builtin_neon_vcadd_rot90_f32((int8x8_t)__p0, (int8x8_t)__p1, 9);
36094   return __ret;
36095 }
36096 #else
vcadd_rot90_f32(float32x2_t __p0,float32x2_t __p1)36097 __ai __attribute__((target("v8.3a"))) float32x2_t vcadd_rot90_f32(float32x2_t __p0, float32x2_t __p1) {
36098   float32x2_t __ret;
36099   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
36100   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
36101   __ret = (float32x2_t) __builtin_neon_vcadd_rot90_f32((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
36102   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
36103   return __ret;
36104 }
36105 #endif
36106 
36107 #ifdef __LITTLE_ENDIAN__
vcaddq_rot270_f32(float32x4_t __p0,float32x4_t __p1)36108 __ai __attribute__((target("v8.3a"))) float32x4_t vcaddq_rot270_f32(float32x4_t __p0, float32x4_t __p1) {
36109   float32x4_t __ret;
36110   __ret = (float32x4_t) __builtin_neon_vcaddq_rot270_f32((int8x16_t)__p0, (int8x16_t)__p1, 41);
36111   return __ret;
36112 }
36113 #else
vcaddq_rot270_f32(float32x4_t __p0,float32x4_t __p1)36114 __ai __attribute__((target("v8.3a"))) float32x4_t vcaddq_rot270_f32(float32x4_t __p0, float32x4_t __p1) {
36115   float32x4_t __ret;
36116   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
36117   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
36118   __ret = (float32x4_t) __builtin_neon_vcaddq_rot270_f32((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
36119   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
36120   return __ret;
36121 }
36122 #endif
36123 
36124 #ifdef __LITTLE_ENDIAN__
vcaddq_rot90_f32(float32x4_t __p0,float32x4_t __p1)36125 __ai __attribute__((target("v8.3a"))) float32x4_t vcaddq_rot90_f32(float32x4_t __p0, float32x4_t __p1) {
36126   float32x4_t __ret;
36127   __ret = (float32x4_t) __builtin_neon_vcaddq_rot90_f32((int8x16_t)__p0, (int8x16_t)__p1, 41);
36128   return __ret;
36129 }
36130 #else
vcaddq_rot90_f32(float32x4_t __p0,float32x4_t __p1)36131 __ai __attribute__((target("v8.3a"))) float32x4_t vcaddq_rot90_f32(float32x4_t __p0, float32x4_t __p1) {
36132   float32x4_t __ret;
36133   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
36134   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
36135   __ret = (float32x4_t) __builtin_neon_vcaddq_rot90_f32((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
36136   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
36137   return __ret;
36138 }
36139 #endif
36140 
36141 #ifdef __LITTLE_ENDIAN__
vcmlaq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36142 __ai __attribute__((target("v8.3a"))) float32x4_t vcmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36143   float32x4_t __ret;
36144   __ret = (float32x4_t) __builtin_neon_vcmlaq_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
36145   return __ret;
36146 }
36147 #else
vcmlaq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36148 __ai __attribute__((target("v8.3a"))) float32x4_t vcmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36149   float32x4_t __ret;
36150   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
36151   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
36152   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
36153   __ret = (float32x4_t) __builtin_neon_vcmlaq_f32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
36154   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
36155   return __ret;
36156 }
__noswap_vcmlaq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36157 __ai __attribute__((target("v8.3a"))) float32x4_t __noswap_vcmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36158   float32x4_t __ret;
36159   __ret = (float32x4_t) __builtin_neon_vcmlaq_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
36160   return __ret;
36161 }
36162 #endif
36163 
36164 #ifdef __LITTLE_ENDIAN__
vcmla_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36165 __ai __attribute__((target("v8.3a"))) float32x2_t vcmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36166   float32x2_t __ret;
36167   __ret = (float32x2_t) __builtin_neon_vcmla_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
36168   return __ret;
36169 }
36170 #else
vcmla_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36171 __ai __attribute__((target("v8.3a"))) float32x2_t vcmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36172   float32x2_t __ret;
36173   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
36174   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
36175   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
36176   __ret = (float32x2_t) __builtin_neon_vcmla_f32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
36177   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
36178   return __ret;
36179 }
__noswap_vcmla_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36180 __ai __attribute__((target("v8.3a"))) float32x2_t __noswap_vcmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36181   float32x2_t __ret;
36182   __ret = (float32x2_t) __builtin_neon_vcmla_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
36183   return __ret;
36184 }
36185 #endif
36186 
36187 #ifdef __LITTLE_ENDIAN__
36188 #define vcmla_lane_f32(__p0_177, __p1_177, __p2_177, __p3_177) __extension__ ({ \
36189   float32x2_t __ret_177; \
36190   float32x2_t __s0_177 = __p0_177; \
36191   float32x2_t __s1_177 = __p1_177; \
36192   float32x2_t __s2_177 = __p2_177; \
36193 float32x2_t __reint_177 = __s2_177; \
36194 uint64x1_t __reint1_177 = (uint64x1_t) {vget_lane_u64(*(uint64x1_t *) &__reint_177, __p3_177)}; \
36195   __ret_177 = vcmla_f32(__s0_177, __s1_177, *(float32x2_t *) &__reint1_177); \
36196   __ret_177; \
36197 })
36198 #else
36199 #define vcmla_lane_f32(__p0_178, __p1_178, __p2_178, __p3_178) __extension__ ({ \
36200   float32x2_t __ret_178; \
36201   float32x2_t __s0_178 = __p0_178; \
36202   float32x2_t __s1_178 = __p1_178; \
36203   float32x2_t __s2_178 = __p2_178; \
36204   float32x2_t __rev0_178;  __rev0_178 = __builtin_shufflevector(__s0_178, __s0_178, 1, 0); \
36205   float32x2_t __rev1_178;  __rev1_178 = __builtin_shufflevector(__s1_178, __s1_178, 1, 0); \
36206   float32x2_t __rev2_178;  __rev2_178 = __builtin_shufflevector(__s2_178, __s2_178, 1, 0); \
36207 float32x2_t __reint_178 = __rev2_178; \
36208 uint64x1_t __reint1_178 = (uint64x1_t) {vget_lane_u64(*(uint64x1_t *) &__reint_178, __p3_178)}; \
36209   __ret_178 = __noswap_vcmla_f32(__rev0_178, __rev1_178, *(float32x2_t *) &__reint1_178); \
36210   __ret_178 = __builtin_shufflevector(__ret_178, __ret_178, 1, 0); \
36211   __ret_178; \
36212 })
36213 #endif
36214 
36215 #ifdef __LITTLE_ENDIAN__
36216 #define vcmlaq_lane_f32(__p0_179, __p1_179, __p2_179, __p3_179) __extension__ ({ \
36217   float32x4_t __ret_179; \
36218   float32x4_t __s0_179 = __p0_179; \
36219   float32x4_t __s1_179 = __p1_179; \
36220   float32x2_t __s2_179 = __p2_179; \
36221 float32x2_t __reint_179 = __s2_179; \
36222 uint64x2_t __reint1_179 = (uint64x2_t) {vget_lane_u64(*(uint64x1_t *) &__reint_179, __p3_179), vget_lane_u64(*(uint64x1_t *) &__reint_179, __p3_179)}; \
36223   __ret_179 = vcmlaq_f32(__s0_179, __s1_179, *(float32x4_t *) &__reint1_179); \
36224   __ret_179; \
36225 })
36226 #else
36227 #define vcmlaq_lane_f32(__p0_180, __p1_180, __p2_180, __p3_180) __extension__ ({ \
36228   float32x4_t __ret_180; \
36229   float32x4_t __s0_180 = __p0_180; \
36230   float32x4_t __s1_180 = __p1_180; \
36231   float32x2_t __s2_180 = __p2_180; \
36232   float32x4_t __rev0_180;  __rev0_180 = __builtin_shufflevector(__s0_180, __s0_180, 3, 2, 1, 0); \
36233   float32x4_t __rev1_180;  __rev1_180 = __builtin_shufflevector(__s1_180, __s1_180, 3, 2, 1, 0); \
36234   float32x2_t __rev2_180;  __rev2_180 = __builtin_shufflevector(__s2_180, __s2_180, 1, 0); \
36235 float32x2_t __reint_180 = __rev2_180; \
36236 uint64x2_t __reint1_180 = (uint64x2_t) {vget_lane_u64(*(uint64x1_t *) &__reint_180, __p3_180), vget_lane_u64(*(uint64x1_t *) &__reint_180, __p3_180)}; \
36237   __ret_180 = __noswap_vcmlaq_f32(__rev0_180, __rev1_180, *(float32x4_t *) &__reint1_180); \
36238   __ret_180 = __builtin_shufflevector(__ret_180, __ret_180, 3, 2, 1, 0); \
36239   __ret_180; \
36240 })
36241 #endif
36242 
36243 #ifdef __LITTLE_ENDIAN__
36244 #define vcmla_laneq_f32(__p0_181, __p1_181, __p2_181, __p3_181) __extension__ ({ \
36245   float32x2_t __ret_181; \
36246   float32x2_t __s0_181 = __p0_181; \
36247   float32x2_t __s1_181 = __p1_181; \
36248   float32x4_t __s2_181 = __p2_181; \
36249 float32x4_t __reint_181 = __s2_181; \
36250 uint64x1_t __reint1_181 = (uint64x1_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_181, __p3_181)}; \
36251   __ret_181 = vcmla_f32(__s0_181, __s1_181, *(float32x2_t *) &__reint1_181); \
36252   __ret_181; \
36253 })
36254 #else
36255 #define vcmla_laneq_f32(__p0_182, __p1_182, __p2_182, __p3_182) __extension__ ({ \
36256   float32x2_t __ret_182; \
36257   float32x2_t __s0_182 = __p0_182; \
36258   float32x2_t __s1_182 = __p1_182; \
36259   float32x4_t __s2_182 = __p2_182; \
36260   float32x2_t __rev0_182;  __rev0_182 = __builtin_shufflevector(__s0_182, __s0_182, 1, 0); \
36261   float32x2_t __rev1_182;  __rev1_182 = __builtin_shufflevector(__s1_182, __s1_182, 1, 0); \
36262   float32x4_t __rev2_182;  __rev2_182 = __builtin_shufflevector(__s2_182, __s2_182, 3, 2, 1, 0); \
36263 float32x4_t __reint_182 = __rev2_182; \
36264 uint64x1_t __reint1_182 = (uint64x1_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_182, __p3_182)}; \
36265   __ret_182 = __noswap_vcmla_f32(__rev0_182, __rev1_182, *(float32x2_t *) &__reint1_182); \
36266   __ret_182 = __builtin_shufflevector(__ret_182, __ret_182, 1, 0); \
36267   __ret_182; \
36268 })
36269 #endif
36270 
36271 #ifdef __LITTLE_ENDIAN__
36272 #define vcmlaq_laneq_f32(__p0_183, __p1_183, __p2_183, __p3_183) __extension__ ({ \
36273   float32x4_t __ret_183; \
36274   float32x4_t __s0_183 = __p0_183; \
36275   float32x4_t __s1_183 = __p1_183; \
36276   float32x4_t __s2_183 = __p2_183; \
36277 float32x4_t __reint_183 = __s2_183; \
36278 uint64x2_t __reint1_183 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_183, __p3_183), vgetq_lane_u64(*(uint64x2_t *) &__reint_183, __p3_183)}; \
36279   __ret_183 = vcmlaq_f32(__s0_183, __s1_183, *(float32x4_t *) &__reint1_183); \
36280   __ret_183; \
36281 })
36282 #else
36283 #define vcmlaq_laneq_f32(__p0_184, __p1_184, __p2_184, __p3_184) __extension__ ({ \
36284   float32x4_t __ret_184; \
36285   float32x4_t __s0_184 = __p0_184; \
36286   float32x4_t __s1_184 = __p1_184; \
36287   float32x4_t __s2_184 = __p2_184; \
36288   float32x4_t __rev0_184;  __rev0_184 = __builtin_shufflevector(__s0_184, __s0_184, 3, 2, 1, 0); \
36289   float32x4_t __rev1_184;  __rev1_184 = __builtin_shufflevector(__s1_184, __s1_184, 3, 2, 1, 0); \
36290   float32x4_t __rev2_184;  __rev2_184 = __builtin_shufflevector(__s2_184, __s2_184, 3, 2, 1, 0); \
36291 float32x4_t __reint_184 = __rev2_184; \
36292 uint64x2_t __reint1_184 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_184, __p3_184), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_184, __p3_184)}; \
36293   __ret_184 = __noswap_vcmlaq_f32(__rev0_184, __rev1_184, *(float32x4_t *) &__reint1_184); \
36294   __ret_184 = __builtin_shufflevector(__ret_184, __ret_184, 3, 2, 1, 0); \
36295   __ret_184; \
36296 })
36297 #endif
36298 
36299 #ifdef __LITTLE_ENDIAN__
vcmlaq_rot180_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36300 __ai __attribute__((target("v8.3a"))) float32x4_t vcmlaq_rot180_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36301   float32x4_t __ret;
36302   __ret = (float32x4_t) __builtin_neon_vcmlaq_rot180_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
36303   return __ret;
36304 }
36305 #else
vcmlaq_rot180_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36306 __ai __attribute__((target("v8.3a"))) float32x4_t vcmlaq_rot180_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36307   float32x4_t __ret;
36308   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
36309   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
36310   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
36311   __ret = (float32x4_t) __builtin_neon_vcmlaq_rot180_f32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
36312   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
36313   return __ret;
36314 }
__noswap_vcmlaq_rot180_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36315 __ai __attribute__((target("v8.3a"))) float32x4_t __noswap_vcmlaq_rot180_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36316   float32x4_t __ret;
36317   __ret = (float32x4_t) __builtin_neon_vcmlaq_rot180_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
36318   return __ret;
36319 }
36320 #endif
36321 
36322 #ifdef __LITTLE_ENDIAN__
vcmla_rot180_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36323 __ai __attribute__((target("v8.3a"))) float32x2_t vcmla_rot180_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36324   float32x2_t __ret;
36325   __ret = (float32x2_t) __builtin_neon_vcmla_rot180_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
36326   return __ret;
36327 }
36328 #else
vcmla_rot180_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36329 __ai __attribute__((target("v8.3a"))) float32x2_t vcmla_rot180_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36330   float32x2_t __ret;
36331   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
36332   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
36333   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
36334   __ret = (float32x2_t) __builtin_neon_vcmla_rot180_f32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
36335   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
36336   return __ret;
36337 }
__noswap_vcmla_rot180_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36338 __ai __attribute__((target("v8.3a"))) float32x2_t __noswap_vcmla_rot180_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36339   float32x2_t __ret;
36340   __ret = (float32x2_t) __builtin_neon_vcmla_rot180_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
36341   return __ret;
36342 }
36343 #endif
36344 
36345 #ifdef __LITTLE_ENDIAN__
36346 #define vcmla_rot180_lane_f32(__p0_185, __p1_185, __p2_185, __p3_185) __extension__ ({ \
36347   float32x2_t __ret_185; \
36348   float32x2_t __s0_185 = __p0_185; \
36349   float32x2_t __s1_185 = __p1_185; \
36350   float32x2_t __s2_185 = __p2_185; \
36351 float32x2_t __reint_185 = __s2_185; \
36352 uint64x1_t __reint1_185 = (uint64x1_t) {vget_lane_u64(*(uint64x1_t *) &__reint_185, __p3_185)}; \
36353   __ret_185 = vcmla_rot180_f32(__s0_185, __s1_185, *(float32x2_t *) &__reint1_185); \
36354   __ret_185; \
36355 })
36356 #else
36357 #define vcmla_rot180_lane_f32(__p0_186, __p1_186, __p2_186, __p3_186) __extension__ ({ \
36358   float32x2_t __ret_186; \
36359   float32x2_t __s0_186 = __p0_186; \
36360   float32x2_t __s1_186 = __p1_186; \
36361   float32x2_t __s2_186 = __p2_186; \
36362   float32x2_t __rev0_186;  __rev0_186 = __builtin_shufflevector(__s0_186, __s0_186, 1, 0); \
36363   float32x2_t __rev1_186;  __rev1_186 = __builtin_shufflevector(__s1_186, __s1_186, 1, 0); \
36364   float32x2_t __rev2_186;  __rev2_186 = __builtin_shufflevector(__s2_186, __s2_186, 1, 0); \
36365 float32x2_t __reint_186 = __rev2_186; \
36366 uint64x1_t __reint1_186 = (uint64x1_t) {vget_lane_u64(*(uint64x1_t *) &__reint_186, __p3_186)}; \
36367   __ret_186 = __noswap_vcmla_rot180_f32(__rev0_186, __rev1_186, *(float32x2_t *) &__reint1_186); \
36368   __ret_186 = __builtin_shufflevector(__ret_186, __ret_186, 1, 0); \
36369   __ret_186; \
36370 })
36371 #endif
36372 
36373 #ifdef __LITTLE_ENDIAN__
36374 #define vcmlaq_rot180_lane_f32(__p0_187, __p1_187, __p2_187, __p3_187) __extension__ ({ \
36375   float32x4_t __ret_187; \
36376   float32x4_t __s0_187 = __p0_187; \
36377   float32x4_t __s1_187 = __p1_187; \
36378   float32x2_t __s2_187 = __p2_187; \
36379 float32x2_t __reint_187 = __s2_187; \
36380 uint64x2_t __reint1_187 = (uint64x2_t) {vget_lane_u64(*(uint64x1_t *) &__reint_187, __p3_187), vget_lane_u64(*(uint64x1_t *) &__reint_187, __p3_187)}; \
36381   __ret_187 = vcmlaq_rot180_f32(__s0_187, __s1_187, *(float32x4_t *) &__reint1_187); \
36382   __ret_187; \
36383 })
36384 #else
36385 #define vcmlaq_rot180_lane_f32(__p0_188, __p1_188, __p2_188, __p3_188) __extension__ ({ \
36386   float32x4_t __ret_188; \
36387   float32x4_t __s0_188 = __p0_188; \
36388   float32x4_t __s1_188 = __p1_188; \
36389   float32x2_t __s2_188 = __p2_188; \
36390   float32x4_t __rev0_188;  __rev0_188 = __builtin_shufflevector(__s0_188, __s0_188, 3, 2, 1, 0); \
36391   float32x4_t __rev1_188;  __rev1_188 = __builtin_shufflevector(__s1_188, __s1_188, 3, 2, 1, 0); \
36392   float32x2_t __rev2_188;  __rev2_188 = __builtin_shufflevector(__s2_188, __s2_188, 1, 0); \
36393 float32x2_t __reint_188 = __rev2_188; \
36394 uint64x2_t __reint1_188 = (uint64x2_t) {vget_lane_u64(*(uint64x1_t *) &__reint_188, __p3_188), vget_lane_u64(*(uint64x1_t *) &__reint_188, __p3_188)}; \
36395   __ret_188 = __noswap_vcmlaq_rot180_f32(__rev0_188, __rev1_188, *(float32x4_t *) &__reint1_188); \
36396   __ret_188 = __builtin_shufflevector(__ret_188, __ret_188, 3, 2, 1, 0); \
36397   __ret_188; \
36398 })
36399 #endif
36400 
36401 #ifdef __LITTLE_ENDIAN__
36402 #define vcmla_rot180_laneq_f32(__p0_189, __p1_189, __p2_189, __p3_189) __extension__ ({ \
36403   float32x2_t __ret_189; \
36404   float32x2_t __s0_189 = __p0_189; \
36405   float32x2_t __s1_189 = __p1_189; \
36406   float32x4_t __s2_189 = __p2_189; \
36407 float32x4_t __reint_189 = __s2_189; \
36408 uint64x1_t __reint1_189 = (uint64x1_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_189, __p3_189)}; \
36409   __ret_189 = vcmla_rot180_f32(__s0_189, __s1_189, *(float32x2_t *) &__reint1_189); \
36410   __ret_189; \
36411 })
36412 #else
36413 #define vcmla_rot180_laneq_f32(__p0_190, __p1_190, __p2_190, __p3_190) __extension__ ({ \
36414   float32x2_t __ret_190; \
36415   float32x2_t __s0_190 = __p0_190; \
36416   float32x2_t __s1_190 = __p1_190; \
36417   float32x4_t __s2_190 = __p2_190; \
36418   float32x2_t __rev0_190;  __rev0_190 = __builtin_shufflevector(__s0_190, __s0_190, 1, 0); \
36419   float32x2_t __rev1_190;  __rev1_190 = __builtin_shufflevector(__s1_190, __s1_190, 1, 0); \
36420   float32x4_t __rev2_190;  __rev2_190 = __builtin_shufflevector(__s2_190, __s2_190, 3, 2, 1, 0); \
36421 float32x4_t __reint_190 = __rev2_190; \
36422 uint64x1_t __reint1_190 = (uint64x1_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_190, __p3_190)}; \
36423   __ret_190 = __noswap_vcmla_rot180_f32(__rev0_190, __rev1_190, *(float32x2_t *) &__reint1_190); \
36424   __ret_190 = __builtin_shufflevector(__ret_190, __ret_190, 1, 0); \
36425   __ret_190; \
36426 })
36427 #endif
36428 
36429 #ifdef __LITTLE_ENDIAN__
36430 #define vcmlaq_rot180_laneq_f32(__p0_191, __p1_191, __p2_191, __p3_191) __extension__ ({ \
36431   float32x4_t __ret_191; \
36432   float32x4_t __s0_191 = __p0_191; \
36433   float32x4_t __s1_191 = __p1_191; \
36434   float32x4_t __s2_191 = __p2_191; \
36435 float32x4_t __reint_191 = __s2_191; \
36436 uint64x2_t __reint1_191 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_191, __p3_191), vgetq_lane_u64(*(uint64x2_t *) &__reint_191, __p3_191)}; \
36437   __ret_191 = vcmlaq_rot180_f32(__s0_191, __s1_191, *(float32x4_t *) &__reint1_191); \
36438   __ret_191; \
36439 })
36440 #else
36441 #define vcmlaq_rot180_laneq_f32(__p0_192, __p1_192, __p2_192, __p3_192) __extension__ ({ \
36442   float32x4_t __ret_192; \
36443   float32x4_t __s0_192 = __p0_192; \
36444   float32x4_t __s1_192 = __p1_192; \
36445   float32x4_t __s2_192 = __p2_192; \
36446   float32x4_t __rev0_192;  __rev0_192 = __builtin_shufflevector(__s0_192, __s0_192, 3, 2, 1, 0); \
36447   float32x4_t __rev1_192;  __rev1_192 = __builtin_shufflevector(__s1_192, __s1_192, 3, 2, 1, 0); \
36448   float32x4_t __rev2_192;  __rev2_192 = __builtin_shufflevector(__s2_192, __s2_192, 3, 2, 1, 0); \
36449 float32x4_t __reint_192 = __rev2_192; \
36450 uint64x2_t __reint1_192 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_192, __p3_192), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_192, __p3_192)}; \
36451   __ret_192 = __noswap_vcmlaq_rot180_f32(__rev0_192, __rev1_192, *(float32x4_t *) &__reint1_192); \
36452   __ret_192 = __builtin_shufflevector(__ret_192, __ret_192, 3, 2, 1, 0); \
36453   __ret_192; \
36454 })
36455 #endif
36456 
36457 #ifdef __LITTLE_ENDIAN__
vcmlaq_rot270_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36458 __ai __attribute__((target("v8.3a"))) float32x4_t vcmlaq_rot270_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36459   float32x4_t __ret;
36460   __ret = (float32x4_t) __builtin_neon_vcmlaq_rot270_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
36461   return __ret;
36462 }
36463 #else
vcmlaq_rot270_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36464 __ai __attribute__((target("v8.3a"))) float32x4_t vcmlaq_rot270_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36465   float32x4_t __ret;
36466   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
36467   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
36468   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
36469   __ret = (float32x4_t) __builtin_neon_vcmlaq_rot270_f32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
36470   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
36471   return __ret;
36472 }
__noswap_vcmlaq_rot270_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36473 __ai __attribute__((target("v8.3a"))) float32x4_t __noswap_vcmlaq_rot270_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36474   float32x4_t __ret;
36475   __ret = (float32x4_t) __builtin_neon_vcmlaq_rot270_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
36476   return __ret;
36477 }
36478 #endif
36479 
36480 #ifdef __LITTLE_ENDIAN__
vcmla_rot270_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36481 __ai __attribute__((target("v8.3a"))) float32x2_t vcmla_rot270_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36482   float32x2_t __ret;
36483   __ret = (float32x2_t) __builtin_neon_vcmla_rot270_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
36484   return __ret;
36485 }
36486 #else
vcmla_rot270_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36487 __ai __attribute__((target("v8.3a"))) float32x2_t vcmla_rot270_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36488   float32x2_t __ret;
36489   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
36490   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
36491   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
36492   __ret = (float32x2_t) __builtin_neon_vcmla_rot270_f32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
36493   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
36494   return __ret;
36495 }
__noswap_vcmla_rot270_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36496 __ai __attribute__((target("v8.3a"))) float32x2_t __noswap_vcmla_rot270_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36497   float32x2_t __ret;
36498   __ret = (float32x2_t) __builtin_neon_vcmla_rot270_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
36499   return __ret;
36500 }
36501 #endif
36502 
36503 #ifdef __LITTLE_ENDIAN__
36504 #define vcmla_rot270_lane_f32(__p0_193, __p1_193, __p2_193, __p3_193) __extension__ ({ \
36505   float32x2_t __ret_193; \
36506   float32x2_t __s0_193 = __p0_193; \
36507   float32x2_t __s1_193 = __p1_193; \
36508   float32x2_t __s2_193 = __p2_193; \
36509 float32x2_t __reint_193 = __s2_193; \
36510 uint64x1_t __reint1_193 = (uint64x1_t) {vget_lane_u64(*(uint64x1_t *) &__reint_193, __p3_193)}; \
36511   __ret_193 = vcmla_rot270_f32(__s0_193, __s1_193, *(float32x2_t *) &__reint1_193); \
36512   __ret_193; \
36513 })
36514 #else
36515 #define vcmla_rot270_lane_f32(__p0_194, __p1_194, __p2_194, __p3_194) __extension__ ({ \
36516   float32x2_t __ret_194; \
36517   float32x2_t __s0_194 = __p0_194; \
36518   float32x2_t __s1_194 = __p1_194; \
36519   float32x2_t __s2_194 = __p2_194; \
36520   float32x2_t __rev0_194;  __rev0_194 = __builtin_shufflevector(__s0_194, __s0_194, 1, 0); \
36521   float32x2_t __rev1_194;  __rev1_194 = __builtin_shufflevector(__s1_194, __s1_194, 1, 0); \
36522   float32x2_t __rev2_194;  __rev2_194 = __builtin_shufflevector(__s2_194, __s2_194, 1, 0); \
36523 float32x2_t __reint_194 = __rev2_194; \
36524 uint64x1_t __reint1_194 = (uint64x1_t) {vget_lane_u64(*(uint64x1_t *) &__reint_194, __p3_194)}; \
36525   __ret_194 = __noswap_vcmla_rot270_f32(__rev0_194, __rev1_194, *(float32x2_t *) &__reint1_194); \
36526   __ret_194 = __builtin_shufflevector(__ret_194, __ret_194, 1, 0); \
36527   __ret_194; \
36528 })
36529 #endif
36530 
36531 #ifdef __LITTLE_ENDIAN__
36532 #define vcmlaq_rot270_lane_f32(__p0_195, __p1_195, __p2_195, __p3_195) __extension__ ({ \
36533   float32x4_t __ret_195; \
36534   float32x4_t __s0_195 = __p0_195; \
36535   float32x4_t __s1_195 = __p1_195; \
36536   float32x2_t __s2_195 = __p2_195; \
36537 float32x2_t __reint_195 = __s2_195; \
36538 uint64x2_t __reint1_195 = (uint64x2_t) {vget_lane_u64(*(uint64x1_t *) &__reint_195, __p3_195), vget_lane_u64(*(uint64x1_t *) &__reint_195, __p3_195)}; \
36539   __ret_195 = vcmlaq_rot270_f32(__s0_195, __s1_195, *(float32x4_t *) &__reint1_195); \
36540   __ret_195; \
36541 })
36542 #else
36543 #define vcmlaq_rot270_lane_f32(__p0_196, __p1_196, __p2_196, __p3_196) __extension__ ({ \
36544   float32x4_t __ret_196; \
36545   float32x4_t __s0_196 = __p0_196; \
36546   float32x4_t __s1_196 = __p1_196; \
36547   float32x2_t __s2_196 = __p2_196; \
36548   float32x4_t __rev0_196;  __rev0_196 = __builtin_shufflevector(__s0_196, __s0_196, 3, 2, 1, 0); \
36549   float32x4_t __rev1_196;  __rev1_196 = __builtin_shufflevector(__s1_196, __s1_196, 3, 2, 1, 0); \
36550   float32x2_t __rev2_196;  __rev2_196 = __builtin_shufflevector(__s2_196, __s2_196, 1, 0); \
36551 float32x2_t __reint_196 = __rev2_196; \
36552 uint64x2_t __reint1_196 = (uint64x2_t) {vget_lane_u64(*(uint64x1_t *) &__reint_196, __p3_196), vget_lane_u64(*(uint64x1_t *) &__reint_196, __p3_196)}; \
36553   __ret_196 = __noswap_vcmlaq_rot270_f32(__rev0_196, __rev1_196, *(float32x4_t *) &__reint1_196); \
36554   __ret_196 = __builtin_shufflevector(__ret_196, __ret_196, 3, 2, 1, 0); \
36555   __ret_196; \
36556 })
36557 #endif
36558 
36559 #ifdef __LITTLE_ENDIAN__
36560 #define vcmla_rot270_laneq_f32(__p0_197, __p1_197, __p2_197, __p3_197) __extension__ ({ \
36561   float32x2_t __ret_197; \
36562   float32x2_t __s0_197 = __p0_197; \
36563   float32x2_t __s1_197 = __p1_197; \
36564   float32x4_t __s2_197 = __p2_197; \
36565 float32x4_t __reint_197 = __s2_197; \
36566 uint64x1_t __reint1_197 = (uint64x1_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_197, __p3_197)}; \
36567   __ret_197 = vcmla_rot270_f32(__s0_197, __s1_197, *(float32x2_t *) &__reint1_197); \
36568   __ret_197; \
36569 })
36570 #else
36571 #define vcmla_rot270_laneq_f32(__p0_198, __p1_198, __p2_198, __p3_198) __extension__ ({ \
36572   float32x2_t __ret_198; \
36573   float32x2_t __s0_198 = __p0_198; \
36574   float32x2_t __s1_198 = __p1_198; \
36575   float32x4_t __s2_198 = __p2_198; \
36576   float32x2_t __rev0_198;  __rev0_198 = __builtin_shufflevector(__s0_198, __s0_198, 1, 0); \
36577   float32x2_t __rev1_198;  __rev1_198 = __builtin_shufflevector(__s1_198, __s1_198, 1, 0); \
36578   float32x4_t __rev2_198;  __rev2_198 = __builtin_shufflevector(__s2_198, __s2_198, 3, 2, 1, 0); \
36579 float32x4_t __reint_198 = __rev2_198; \
36580 uint64x1_t __reint1_198 = (uint64x1_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_198, __p3_198)}; \
36581   __ret_198 = __noswap_vcmla_rot270_f32(__rev0_198, __rev1_198, *(float32x2_t *) &__reint1_198); \
36582   __ret_198 = __builtin_shufflevector(__ret_198, __ret_198, 1, 0); \
36583   __ret_198; \
36584 })
36585 #endif
36586 
36587 #ifdef __LITTLE_ENDIAN__
36588 #define vcmlaq_rot270_laneq_f32(__p0_199, __p1_199, __p2_199, __p3_199) __extension__ ({ \
36589   float32x4_t __ret_199; \
36590   float32x4_t __s0_199 = __p0_199; \
36591   float32x4_t __s1_199 = __p1_199; \
36592   float32x4_t __s2_199 = __p2_199; \
36593 float32x4_t __reint_199 = __s2_199; \
36594 uint64x2_t __reint1_199 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_199, __p3_199), vgetq_lane_u64(*(uint64x2_t *) &__reint_199, __p3_199)}; \
36595   __ret_199 = vcmlaq_rot270_f32(__s0_199, __s1_199, *(float32x4_t *) &__reint1_199); \
36596   __ret_199; \
36597 })
36598 #else
36599 #define vcmlaq_rot270_laneq_f32(__p0_200, __p1_200, __p2_200, __p3_200) __extension__ ({ \
36600   float32x4_t __ret_200; \
36601   float32x4_t __s0_200 = __p0_200; \
36602   float32x4_t __s1_200 = __p1_200; \
36603   float32x4_t __s2_200 = __p2_200; \
36604   float32x4_t __rev0_200;  __rev0_200 = __builtin_shufflevector(__s0_200, __s0_200, 3, 2, 1, 0); \
36605   float32x4_t __rev1_200;  __rev1_200 = __builtin_shufflevector(__s1_200, __s1_200, 3, 2, 1, 0); \
36606   float32x4_t __rev2_200;  __rev2_200 = __builtin_shufflevector(__s2_200, __s2_200, 3, 2, 1, 0); \
36607 float32x4_t __reint_200 = __rev2_200; \
36608 uint64x2_t __reint1_200 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_200, __p3_200), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_200, __p3_200)}; \
36609   __ret_200 = __noswap_vcmlaq_rot270_f32(__rev0_200, __rev1_200, *(float32x4_t *) &__reint1_200); \
36610   __ret_200 = __builtin_shufflevector(__ret_200, __ret_200, 3, 2, 1, 0); \
36611   __ret_200; \
36612 })
36613 #endif
36614 
36615 #ifdef __LITTLE_ENDIAN__
vcmlaq_rot90_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36616 __ai __attribute__((target("v8.3a"))) float32x4_t vcmlaq_rot90_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36617   float32x4_t __ret;
36618   __ret = (float32x4_t) __builtin_neon_vcmlaq_rot90_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
36619   return __ret;
36620 }
36621 #else
vcmlaq_rot90_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36622 __ai __attribute__((target("v8.3a"))) float32x4_t vcmlaq_rot90_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36623   float32x4_t __ret;
36624   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
36625   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
36626   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
36627   __ret = (float32x4_t) __builtin_neon_vcmlaq_rot90_f32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
36628   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
36629   return __ret;
36630 }
__noswap_vcmlaq_rot90_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)36631 __ai __attribute__((target("v8.3a"))) float32x4_t __noswap_vcmlaq_rot90_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
36632   float32x4_t __ret;
36633   __ret = (float32x4_t) __builtin_neon_vcmlaq_rot90_f32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
36634   return __ret;
36635 }
36636 #endif
36637 
36638 #ifdef __LITTLE_ENDIAN__
vcmla_rot90_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36639 __ai __attribute__((target("v8.3a"))) float32x2_t vcmla_rot90_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36640   float32x2_t __ret;
36641   __ret = (float32x2_t) __builtin_neon_vcmla_rot90_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
36642   return __ret;
36643 }
36644 #else
vcmla_rot90_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36645 __ai __attribute__((target("v8.3a"))) float32x2_t vcmla_rot90_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36646   float32x2_t __ret;
36647   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
36648   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
36649   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
36650   __ret = (float32x2_t) __builtin_neon_vcmla_rot90_f32((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
36651   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
36652   return __ret;
36653 }
__noswap_vcmla_rot90_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)36654 __ai __attribute__((target("v8.3a"))) float32x2_t __noswap_vcmla_rot90_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
36655   float32x2_t __ret;
36656   __ret = (float32x2_t) __builtin_neon_vcmla_rot90_f32((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
36657   return __ret;
36658 }
36659 #endif
36660 
36661 #ifdef __LITTLE_ENDIAN__
36662 #define vcmla_rot90_lane_f32(__p0_201, __p1_201, __p2_201, __p3_201) __extension__ ({ \
36663   float32x2_t __ret_201; \
36664   float32x2_t __s0_201 = __p0_201; \
36665   float32x2_t __s1_201 = __p1_201; \
36666   float32x2_t __s2_201 = __p2_201; \
36667 float32x2_t __reint_201 = __s2_201; \
36668 uint64x1_t __reint1_201 = (uint64x1_t) {vget_lane_u64(*(uint64x1_t *) &__reint_201, __p3_201)}; \
36669   __ret_201 = vcmla_rot90_f32(__s0_201, __s1_201, *(float32x2_t *) &__reint1_201); \
36670   __ret_201; \
36671 })
36672 #else
36673 #define vcmla_rot90_lane_f32(__p0_202, __p1_202, __p2_202, __p3_202) __extension__ ({ \
36674   float32x2_t __ret_202; \
36675   float32x2_t __s0_202 = __p0_202; \
36676   float32x2_t __s1_202 = __p1_202; \
36677   float32x2_t __s2_202 = __p2_202; \
36678   float32x2_t __rev0_202;  __rev0_202 = __builtin_shufflevector(__s0_202, __s0_202, 1, 0); \
36679   float32x2_t __rev1_202;  __rev1_202 = __builtin_shufflevector(__s1_202, __s1_202, 1, 0); \
36680   float32x2_t __rev2_202;  __rev2_202 = __builtin_shufflevector(__s2_202, __s2_202, 1, 0); \
36681 float32x2_t __reint_202 = __rev2_202; \
36682 uint64x1_t __reint1_202 = (uint64x1_t) {vget_lane_u64(*(uint64x1_t *) &__reint_202, __p3_202)}; \
36683   __ret_202 = __noswap_vcmla_rot90_f32(__rev0_202, __rev1_202, *(float32x2_t *) &__reint1_202); \
36684   __ret_202 = __builtin_shufflevector(__ret_202, __ret_202, 1, 0); \
36685   __ret_202; \
36686 })
36687 #endif
36688 
36689 #ifdef __LITTLE_ENDIAN__
36690 #define vcmlaq_rot90_lane_f32(__p0_203, __p1_203, __p2_203, __p3_203) __extension__ ({ \
36691   float32x4_t __ret_203; \
36692   float32x4_t __s0_203 = __p0_203; \
36693   float32x4_t __s1_203 = __p1_203; \
36694   float32x2_t __s2_203 = __p2_203; \
36695 float32x2_t __reint_203 = __s2_203; \
36696 uint64x2_t __reint1_203 = (uint64x2_t) {vget_lane_u64(*(uint64x1_t *) &__reint_203, __p3_203), vget_lane_u64(*(uint64x1_t *) &__reint_203, __p3_203)}; \
36697   __ret_203 = vcmlaq_rot90_f32(__s0_203, __s1_203, *(float32x4_t *) &__reint1_203); \
36698   __ret_203; \
36699 })
36700 #else
36701 #define vcmlaq_rot90_lane_f32(__p0_204, __p1_204, __p2_204, __p3_204) __extension__ ({ \
36702   float32x4_t __ret_204; \
36703   float32x4_t __s0_204 = __p0_204; \
36704   float32x4_t __s1_204 = __p1_204; \
36705   float32x2_t __s2_204 = __p2_204; \
36706   float32x4_t __rev0_204;  __rev0_204 = __builtin_shufflevector(__s0_204, __s0_204, 3, 2, 1, 0); \
36707   float32x4_t __rev1_204;  __rev1_204 = __builtin_shufflevector(__s1_204, __s1_204, 3, 2, 1, 0); \
36708   float32x2_t __rev2_204;  __rev2_204 = __builtin_shufflevector(__s2_204, __s2_204, 1, 0); \
36709 float32x2_t __reint_204 = __rev2_204; \
36710 uint64x2_t __reint1_204 = (uint64x2_t) {vget_lane_u64(*(uint64x1_t *) &__reint_204, __p3_204), vget_lane_u64(*(uint64x1_t *) &__reint_204, __p3_204)}; \
36711   __ret_204 = __noswap_vcmlaq_rot90_f32(__rev0_204, __rev1_204, *(float32x4_t *) &__reint1_204); \
36712   __ret_204 = __builtin_shufflevector(__ret_204, __ret_204, 3, 2, 1, 0); \
36713   __ret_204; \
36714 })
36715 #endif
36716 
36717 #ifdef __LITTLE_ENDIAN__
36718 #define vcmla_rot90_laneq_f32(__p0_205, __p1_205, __p2_205, __p3_205) __extension__ ({ \
36719   float32x2_t __ret_205; \
36720   float32x2_t __s0_205 = __p0_205; \
36721   float32x2_t __s1_205 = __p1_205; \
36722   float32x4_t __s2_205 = __p2_205; \
36723 float32x4_t __reint_205 = __s2_205; \
36724 uint64x1_t __reint1_205 = (uint64x1_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_205, __p3_205)}; \
36725   __ret_205 = vcmla_rot90_f32(__s0_205, __s1_205, *(float32x2_t *) &__reint1_205); \
36726   __ret_205; \
36727 })
36728 #else
36729 #define vcmla_rot90_laneq_f32(__p0_206, __p1_206, __p2_206, __p3_206) __extension__ ({ \
36730   float32x2_t __ret_206; \
36731   float32x2_t __s0_206 = __p0_206; \
36732   float32x2_t __s1_206 = __p1_206; \
36733   float32x4_t __s2_206 = __p2_206; \
36734   float32x2_t __rev0_206;  __rev0_206 = __builtin_shufflevector(__s0_206, __s0_206, 1, 0); \
36735   float32x2_t __rev1_206;  __rev1_206 = __builtin_shufflevector(__s1_206, __s1_206, 1, 0); \
36736   float32x4_t __rev2_206;  __rev2_206 = __builtin_shufflevector(__s2_206, __s2_206, 3, 2, 1, 0); \
36737 float32x4_t __reint_206 = __rev2_206; \
36738 uint64x1_t __reint1_206 = (uint64x1_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_206, __p3_206)}; \
36739   __ret_206 = __noswap_vcmla_rot90_f32(__rev0_206, __rev1_206, *(float32x2_t *) &__reint1_206); \
36740   __ret_206 = __builtin_shufflevector(__ret_206, __ret_206, 1, 0); \
36741   __ret_206; \
36742 })
36743 #endif
36744 
36745 #ifdef __LITTLE_ENDIAN__
36746 #define vcmlaq_rot90_laneq_f32(__p0_207, __p1_207, __p2_207, __p3_207) __extension__ ({ \
36747   float32x4_t __ret_207; \
36748   float32x4_t __s0_207 = __p0_207; \
36749   float32x4_t __s1_207 = __p1_207; \
36750   float32x4_t __s2_207 = __p2_207; \
36751 float32x4_t __reint_207 = __s2_207; \
36752 uint64x2_t __reint1_207 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_207, __p3_207), vgetq_lane_u64(*(uint64x2_t *) &__reint_207, __p3_207)}; \
36753   __ret_207 = vcmlaq_rot90_f32(__s0_207, __s1_207, *(float32x4_t *) &__reint1_207); \
36754   __ret_207; \
36755 })
36756 #else
36757 #define vcmlaq_rot90_laneq_f32(__p0_208, __p1_208, __p2_208, __p3_208) __extension__ ({ \
36758   float32x4_t __ret_208; \
36759   float32x4_t __s0_208 = __p0_208; \
36760   float32x4_t __s1_208 = __p1_208; \
36761   float32x4_t __s2_208 = __p2_208; \
36762   float32x4_t __rev0_208;  __rev0_208 = __builtin_shufflevector(__s0_208, __s0_208, 3, 2, 1, 0); \
36763   float32x4_t __rev1_208;  __rev1_208 = __builtin_shufflevector(__s1_208, __s1_208, 3, 2, 1, 0); \
36764   float32x4_t __rev2_208;  __rev2_208 = __builtin_shufflevector(__s2_208, __s2_208, 3, 2, 1, 0); \
36765 float32x4_t __reint_208 = __rev2_208; \
36766 uint64x2_t __reint1_208 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_208, __p3_208), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_208, __p3_208)}; \
36767   __ret_208 = __noswap_vcmlaq_rot90_f32(__rev0_208, __rev1_208, *(float32x4_t *) &__reint1_208); \
36768   __ret_208 = __builtin_shufflevector(__ret_208, __ret_208, 3, 2, 1, 0); \
36769   __ret_208; \
36770 })
36771 #endif
36772 
36773 #ifdef __LITTLE_ENDIAN__
vcadd_rot270_f16(float16x4_t __p0,float16x4_t __p1)36774 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcadd_rot270_f16(float16x4_t __p0, float16x4_t __p1) {
36775   float16x4_t __ret;
36776   __ret = (float16x4_t) __builtin_neon_vcadd_rot270_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
36777   return __ret;
36778 }
36779 #else
vcadd_rot270_f16(float16x4_t __p0,float16x4_t __p1)36780 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcadd_rot270_f16(float16x4_t __p0, float16x4_t __p1) {
36781   float16x4_t __ret;
36782   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
36783   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
36784   __ret = (float16x4_t) __builtin_neon_vcadd_rot270_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
36785   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
36786   return __ret;
36787 }
36788 #endif
36789 
36790 #ifdef __LITTLE_ENDIAN__
vcadd_rot90_f16(float16x4_t __p0,float16x4_t __p1)36791 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcadd_rot90_f16(float16x4_t __p0, float16x4_t __p1) {
36792   float16x4_t __ret;
36793   __ret = (float16x4_t) __builtin_neon_vcadd_rot90_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
36794   return __ret;
36795 }
36796 #else
vcadd_rot90_f16(float16x4_t __p0,float16x4_t __p1)36797 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcadd_rot90_f16(float16x4_t __p0, float16x4_t __p1) {
36798   float16x4_t __ret;
36799   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
36800   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
36801   __ret = (float16x4_t) __builtin_neon_vcadd_rot90_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
36802   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
36803   return __ret;
36804 }
36805 #endif
36806 
36807 #ifdef __LITTLE_ENDIAN__
vcaddq_rot270_f16(float16x8_t __p0,float16x8_t __p1)36808 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcaddq_rot270_f16(float16x8_t __p0, float16x8_t __p1) {
36809   float16x8_t __ret;
36810   __ret = (float16x8_t) __builtin_neon_vcaddq_rot270_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
36811   return __ret;
36812 }
36813 #else
vcaddq_rot270_f16(float16x8_t __p0,float16x8_t __p1)36814 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcaddq_rot270_f16(float16x8_t __p0, float16x8_t __p1) {
36815   float16x8_t __ret;
36816   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
36817   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
36818   __ret = (float16x8_t) __builtin_neon_vcaddq_rot270_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
36819   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
36820   return __ret;
36821 }
36822 #endif
36823 
36824 #ifdef __LITTLE_ENDIAN__
vcaddq_rot90_f16(float16x8_t __p0,float16x8_t __p1)36825 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcaddq_rot90_f16(float16x8_t __p0, float16x8_t __p1) {
36826   float16x8_t __ret;
36827   __ret = (float16x8_t) __builtin_neon_vcaddq_rot90_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
36828   return __ret;
36829 }
36830 #else
vcaddq_rot90_f16(float16x8_t __p0,float16x8_t __p1)36831 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcaddq_rot90_f16(float16x8_t __p0, float16x8_t __p1) {
36832   float16x8_t __ret;
36833   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
36834   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
36835   __ret = (float16x8_t) __builtin_neon_vcaddq_rot90_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
36836   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
36837   return __ret;
36838 }
36839 #endif
36840 
36841 #ifdef __LITTLE_ENDIAN__
vcmlaq_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)36842 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcmlaq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
36843   float16x8_t __ret;
36844   __ret = (float16x8_t) __builtin_neon_vcmlaq_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
36845   return __ret;
36846 }
36847 #else
vcmlaq_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)36848 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcmlaq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
36849   float16x8_t __ret;
36850   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
36851   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
36852   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
36853   __ret = (float16x8_t) __builtin_neon_vcmlaq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 40);
36854   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
36855   return __ret;
36856 }
__noswap_vcmlaq_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)36857 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t __noswap_vcmlaq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
36858   float16x8_t __ret;
36859   __ret = (float16x8_t) __builtin_neon_vcmlaq_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
36860   return __ret;
36861 }
36862 #endif
36863 
36864 #ifdef __LITTLE_ENDIAN__
vcmla_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)36865 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcmla_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
36866   float16x4_t __ret;
36867   __ret = (float16x4_t) __builtin_neon_vcmla_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
36868   return __ret;
36869 }
36870 #else
vcmla_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)36871 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcmla_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
36872   float16x4_t __ret;
36873   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
36874   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
36875   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
36876   __ret = (float16x4_t) __builtin_neon_vcmla_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 8);
36877   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
36878   return __ret;
36879 }
__noswap_vcmla_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)36880 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t __noswap_vcmla_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
36881   float16x4_t __ret;
36882   __ret = (float16x4_t) __builtin_neon_vcmla_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
36883   return __ret;
36884 }
36885 #endif
36886 
36887 #ifdef __LITTLE_ENDIAN__
36888 #define vcmla_lane_f16(__p0_209, __p1_209, __p2_209, __p3_209) __extension__ ({ \
36889   float16x4_t __ret_209; \
36890   float16x4_t __s0_209 = __p0_209; \
36891   float16x4_t __s1_209 = __p1_209; \
36892   float16x4_t __s2_209 = __p2_209; \
36893 float16x4_t __reint_209 = __s2_209; \
36894 uint32x2_t __reint1_209 = (uint32x2_t) {vget_lane_u32(*(uint32x2_t *) &__reint_209, __p3_209), vget_lane_u32(*(uint32x2_t *) &__reint_209, __p3_209)}; \
36895   __ret_209 = vcmla_f16(__s0_209, __s1_209, *(float16x4_t *) &__reint1_209); \
36896   __ret_209; \
36897 })
36898 #else
36899 #define vcmla_lane_f16(__p0_210, __p1_210, __p2_210, __p3_210) __extension__ ({ \
36900   float16x4_t __ret_210; \
36901   float16x4_t __s0_210 = __p0_210; \
36902   float16x4_t __s1_210 = __p1_210; \
36903   float16x4_t __s2_210 = __p2_210; \
36904   float16x4_t __rev0_210;  __rev0_210 = __builtin_shufflevector(__s0_210, __s0_210, 3, 2, 1, 0); \
36905   float16x4_t __rev1_210;  __rev1_210 = __builtin_shufflevector(__s1_210, __s1_210, 3, 2, 1, 0); \
36906   float16x4_t __rev2_210;  __rev2_210 = __builtin_shufflevector(__s2_210, __s2_210, 3, 2, 1, 0); \
36907 float16x4_t __reint_210 = __rev2_210; \
36908 uint32x2_t __reint1_210 = (uint32x2_t) {__noswap_vget_lane_u32(*(uint32x2_t *) &__reint_210, __p3_210), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_210, __p3_210)}; \
36909   __ret_210 = __noswap_vcmla_f16(__rev0_210, __rev1_210, *(float16x4_t *) &__reint1_210); \
36910   __ret_210 = __builtin_shufflevector(__ret_210, __ret_210, 3, 2, 1, 0); \
36911   __ret_210; \
36912 })
36913 #endif
36914 
36915 #ifdef __LITTLE_ENDIAN__
36916 #define vcmlaq_lane_f16(__p0_211, __p1_211, __p2_211, __p3_211) __extension__ ({ \
36917   float16x8_t __ret_211; \
36918   float16x8_t __s0_211 = __p0_211; \
36919   float16x8_t __s1_211 = __p1_211; \
36920   float16x4_t __s2_211 = __p2_211; \
36921 float16x4_t __reint_211 = __s2_211; \
36922 uint32x4_t __reint1_211 = (uint32x4_t) {vget_lane_u32(*(uint32x2_t *) &__reint_211, __p3_211), vget_lane_u32(*(uint32x2_t *) &__reint_211, __p3_211), vget_lane_u32(*(uint32x2_t *) &__reint_211, __p3_211), vget_lane_u32(*(uint32x2_t *) &__reint_211, __p3_211)}; \
36923   __ret_211 = vcmlaq_f16(__s0_211, __s1_211, *(float16x8_t *) &__reint1_211); \
36924   __ret_211; \
36925 })
36926 #else
36927 #define vcmlaq_lane_f16(__p0_212, __p1_212, __p2_212, __p3_212) __extension__ ({ \
36928   float16x8_t __ret_212; \
36929   float16x8_t __s0_212 = __p0_212; \
36930   float16x8_t __s1_212 = __p1_212; \
36931   float16x4_t __s2_212 = __p2_212; \
36932   float16x8_t __rev0_212;  __rev0_212 = __builtin_shufflevector(__s0_212, __s0_212, 7, 6, 5, 4, 3, 2, 1, 0); \
36933   float16x8_t __rev1_212;  __rev1_212 = __builtin_shufflevector(__s1_212, __s1_212, 7, 6, 5, 4, 3, 2, 1, 0); \
36934   float16x4_t __rev2_212;  __rev2_212 = __builtin_shufflevector(__s2_212, __s2_212, 3, 2, 1, 0); \
36935 float16x4_t __reint_212 = __rev2_212; \
36936 uint32x4_t __reint1_212 = (uint32x4_t) {__noswap_vget_lane_u32(*(uint32x2_t *) &__reint_212, __p3_212), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_212, __p3_212), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_212, __p3_212), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_212, __p3_212)}; \
36937   __ret_212 = __noswap_vcmlaq_f16(__rev0_212, __rev1_212, *(float16x8_t *) &__reint1_212); \
36938   __ret_212 = __builtin_shufflevector(__ret_212, __ret_212, 7, 6, 5, 4, 3, 2, 1, 0); \
36939   __ret_212; \
36940 })
36941 #endif
36942 
36943 #ifdef __LITTLE_ENDIAN__
36944 #define vcmla_laneq_f16(__p0_213, __p1_213, __p2_213, __p3_213) __extension__ ({ \
36945   float16x4_t __ret_213; \
36946   float16x4_t __s0_213 = __p0_213; \
36947   float16x4_t __s1_213 = __p1_213; \
36948   float16x8_t __s2_213 = __p2_213; \
36949 float16x8_t __reint_213 = __s2_213; \
36950 uint32x2_t __reint1_213 = (uint32x2_t) {vgetq_lane_u32(*(uint32x4_t *) &__reint_213, __p3_213), vgetq_lane_u32(*(uint32x4_t *) &__reint_213, __p3_213)}; \
36951   __ret_213 = vcmla_f16(__s0_213, __s1_213, *(float16x4_t *) &__reint1_213); \
36952   __ret_213; \
36953 })
36954 #else
36955 #define vcmla_laneq_f16(__p0_214, __p1_214, __p2_214, __p3_214) __extension__ ({ \
36956   float16x4_t __ret_214; \
36957   float16x4_t __s0_214 = __p0_214; \
36958   float16x4_t __s1_214 = __p1_214; \
36959   float16x8_t __s2_214 = __p2_214; \
36960   float16x4_t __rev0_214;  __rev0_214 = __builtin_shufflevector(__s0_214, __s0_214, 3, 2, 1, 0); \
36961   float16x4_t __rev1_214;  __rev1_214 = __builtin_shufflevector(__s1_214, __s1_214, 3, 2, 1, 0); \
36962   float16x8_t __rev2_214;  __rev2_214 = __builtin_shufflevector(__s2_214, __s2_214, 7, 6, 5, 4, 3, 2, 1, 0); \
36963 float16x8_t __reint_214 = __rev2_214; \
36964 uint32x2_t __reint1_214 = (uint32x2_t) {__noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_214, __p3_214), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_214, __p3_214)}; \
36965   __ret_214 = __noswap_vcmla_f16(__rev0_214, __rev1_214, *(float16x4_t *) &__reint1_214); \
36966   __ret_214 = __builtin_shufflevector(__ret_214, __ret_214, 3, 2, 1, 0); \
36967   __ret_214; \
36968 })
36969 #endif
36970 
36971 #ifdef __LITTLE_ENDIAN__
36972 #define vcmlaq_laneq_f16(__p0_215, __p1_215, __p2_215, __p3_215) __extension__ ({ \
36973   float16x8_t __ret_215; \
36974   float16x8_t __s0_215 = __p0_215; \
36975   float16x8_t __s1_215 = __p1_215; \
36976   float16x8_t __s2_215 = __p2_215; \
36977 float16x8_t __reint_215 = __s2_215; \
36978 uint32x4_t __reint1_215 = (uint32x4_t) {vgetq_lane_u32(*(uint32x4_t *) &__reint_215, __p3_215), vgetq_lane_u32(*(uint32x4_t *) &__reint_215, __p3_215), vgetq_lane_u32(*(uint32x4_t *) &__reint_215, __p3_215), vgetq_lane_u32(*(uint32x4_t *) &__reint_215, __p3_215)}; \
36979   __ret_215 = vcmlaq_f16(__s0_215, __s1_215, *(float16x8_t *) &__reint1_215); \
36980   __ret_215; \
36981 })
36982 #else
36983 #define vcmlaq_laneq_f16(__p0_216, __p1_216, __p2_216, __p3_216) __extension__ ({ \
36984   float16x8_t __ret_216; \
36985   float16x8_t __s0_216 = __p0_216; \
36986   float16x8_t __s1_216 = __p1_216; \
36987   float16x8_t __s2_216 = __p2_216; \
36988   float16x8_t __rev0_216;  __rev0_216 = __builtin_shufflevector(__s0_216, __s0_216, 7, 6, 5, 4, 3, 2, 1, 0); \
36989   float16x8_t __rev1_216;  __rev1_216 = __builtin_shufflevector(__s1_216, __s1_216, 7, 6, 5, 4, 3, 2, 1, 0); \
36990   float16x8_t __rev2_216;  __rev2_216 = __builtin_shufflevector(__s2_216, __s2_216, 7, 6, 5, 4, 3, 2, 1, 0); \
36991 float16x8_t __reint_216 = __rev2_216; \
36992 uint32x4_t __reint1_216 = (uint32x4_t) {__noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_216, __p3_216), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_216, __p3_216), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_216, __p3_216), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_216, __p3_216)}; \
36993   __ret_216 = __noswap_vcmlaq_f16(__rev0_216, __rev1_216, *(float16x8_t *) &__reint1_216); \
36994   __ret_216 = __builtin_shufflevector(__ret_216, __ret_216, 7, 6, 5, 4, 3, 2, 1, 0); \
36995   __ret_216; \
36996 })
36997 #endif
36998 
36999 #ifdef __LITTLE_ENDIAN__
vcmlaq_rot180_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)37000 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcmlaq_rot180_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
37001   float16x8_t __ret;
37002   __ret = (float16x8_t) __builtin_neon_vcmlaq_rot180_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
37003   return __ret;
37004 }
37005 #else
vcmlaq_rot180_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)37006 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcmlaq_rot180_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
37007   float16x8_t __ret;
37008   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
37009   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
37010   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
37011   __ret = (float16x8_t) __builtin_neon_vcmlaq_rot180_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 40);
37012   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
37013   return __ret;
37014 }
__noswap_vcmlaq_rot180_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)37015 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t __noswap_vcmlaq_rot180_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
37016   float16x8_t __ret;
37017   __ret = (float16x8_t) __builtin_neon_vcmlaq_rot180_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
37018   return __ret;
37019 }
37020 #endif
37021 
37022 #ifdef __LITTLE_ENDIAN__
vcmla_rot180_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)37023 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcmla_rot180_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
37024   float16x4_t __ret;
37025   __ret = (float16x4_t) __builtin_neon_vcmla_rot180_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
37026   return __ret;
37027 }
37028 #else
vcmla_rot180_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)37029 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcmla_rot180_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
37030   float16x4_t __ret;
37031   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
37032   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
37033   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
37034   __ret = (float16x4_t) __builtin_neon_vcmla_rot180_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 8);
37035   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
37036   return __ret;
37037 }
__noswap_vcmla_rot180_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)37038 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t __noswap_vcmla_rot180_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
37039   float16x4_t __ret;
37040   __ret = (float16x4_t) __builtin_neon_vcmla_rot180_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
37041   return __ret;
37042 }
37043 #endif
37044 
37045 #ifdef __LITTLE_ENDIAN__
37046 #define vcmla_rot180_lane_f16(__p0_217, __p1_217, __p2_217, __p3_217) __extension__ ({ \
37047   float16x4_t __ret_217; \
37048   float16x4_t __s0_217 = __p0_217; \
37049   float16x4_t __s1_217 = __p1_217; \
37050   float16x4_t __s2_217 = __p2_217; \
37051 float16x4_t __reint_217 = __s2_217; \
37052 uint32x2_t __reint1_217 = (uint32x2_t) {vget_lane_u32(*(uint32x2_t *) &__reint_217, __p3_217), vget_lane_u32(*(uint32x2_t *) &__reint_217, __p3_217)}; \
37053   __ret_217 = vcmla_rot180_f16(__s0_217, __s1_217, *(float16x4_t *) &__reint1_217); \
37054   __ret_217; \
37055 })
37056 #else
37057 #define vcmla_rot180_lane_f16(__p0_218, __p1_218, __p2_218, __p3_218) __extension__ ({ \
37058   float16x4_t __ret_218; \
37059   float16x4_t __s0_218 = __p0_218; \
37060   float16x4_t __s1_218 = __p1_218; \
37061   float16x4_t __s2_218 = __p2_218; \
37062   float16x4_t __rev0_218;  __rev0_218 = __builtin_shufflevector(__s0_218, __s0_218, 3, 2, 1, 0); \
37063   float16x4_t __rev1_218;  __rev1_218 = __builtin_shufflevector(__s1_218, __s1_218, 3, 2, 1, 0); \
37064   float16x4_t __rev2_218;  __rev2_218 = __builtin_shufflevector(__s2_218, __s2_218, 3, 2, 1, 0); \
37065 float16x4_t __reint_218 = __rev2_218; \
37066 uint32x2_t __reint1_218 = (uint32x2_t) {__noswap_vget_lane_u32(*(uint32x2_t *) &__reint_218, __p3_218), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_218, __p3_218)}; \
37067   __ret_218 = __noswap_vcmla_rot180_f16(__rev0_218, __rev1_218, *(float16x4_t *) &__reint1_218); \
37068   __ret_218 = __builtin_shufflevector(__ret_218, __ret_218, 3, 2, 1, 0); \
37069   __ret_218; \
37070 })
37071 #endif
37072 
37073 #ifdef __LITTLE_ENDIAN__
37074 #define vcmlaq_rot180_lane_f16(__p0_219, __p1_219, __p2_219, __p3_219) __extension__ ({ \
37075   float16x8_t __ret_219; \
37076   float16x8_t __s0_219 = __p0_219; \
37077   float16x8_t __s1_219 = __p1_219; \
37078   float16x4_t __s2_219 = __p2_219; \
37079 float16x4_t __reint_219 = __s2_219; \
37080 uint32x4_t __reint1_219 = (uint32x4_t) {vget_lane_u32(*(uint32x2_t *) &__reint_219, __p3_219), vget_lane_u32(*(uint32x2_t *) &__reint_219, __p3_219), vget_lane_u32(*(uint32x2_t *) &__reint_219, __p3_219), vget_lane_u32(*(uint32x2_t *) &__reint_219, __p3_219)}; \
37081   __ret_219 = vcmlaq_rot180_f16(__s0_219, __s1_219, *(float16x8_t *) &__reint1_219); \
37082   __ret_219; \
37083 })
37084 #else
37085 #define vcmlaq_rot180_lane_f16(__p0_220, __p1_220, __p2_220, __p3_220) __extension__ ({ \
37086   float16x8_t __ret_220; \
37087   float16x8_t __s0_220 = __p0_220; \
37088   float16x8_t __s1_220 = __p1_220; \
37089   float16x4_t __s2_220 = __p2_220; \
37090   float16x8_t __rev0_220;  __rev0_220 = __builtin_shufflevector(__s0_220, __s0_220, 7, 6, 5, 4, 3, 2, 1, 0); \
37091   float16x8_t __rev1_220;  __rev1_220 = __builtin_shufflevector(__s1_220, __s1_220, 7, 6, 5, 4, 3, 2, 1, 0); \
37092   float16x4_t __rev2_220;  __rev2_220 = __builtin_shufflevector(__s2_220, __s2_220, 3, 2, 1, 0); \
37093 float16x4_t __reint_220 = __rev2_220; \
37094 uint32x4_t __reint1_220 = (uint32x4_t) {__noswap_vget_lane_u32(*(uint32x2_t *) &__reint_220, __p3_220), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_220, __p3_220), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_220, __p3_220), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_220, __p3_220)}; \
37095   __ret_220 = __noswap_vcmlaq_rot180_f16(__rev0_220, __rev1_220, *(float16x8_t *) &__reint1_220); \
37096   __ret_220 = __builtin_shufflevector(__ret_220, __ret_220, 7, 6, 5, 4, 3, 2, 1, 0); \
37097   __ret_220; \
37098 })
37099 #endif
37100 
37101 #ifdef __LITTLE_ENDIAN__
37102 #define vcmla_rot180_laneq_f16(__p0_221, __p1_221, __p2_221, __p3_221) __extension__ ({ \
37103   float16x4_t __ret_221; \
37104   float16x4_t __s0_221 = __p0_221; \
37105   float16x4_t __s1_221 = __p1_221; \
37106   float16x8_t __s2_221 = __p2_221; \
37107 float16x8_t __reint_221 = __s2_221; \
37108 uint32x2_t __reint1_221 = (uint32x2_t) {vgetq_lane_u32(*(uint32x4_t *) &__reint_221, __p3_221), vgetq_lane_u32(*(uint32x4_t *) &__reint_221, __p3_221)}; \
37109   __ret_221 = vcmla_rot180_f16(__s0_221, __s1_221, *(float16x4_t *) &__reint1_221); \
37110   __ret_221; \
37111 })
37112 #else
37113 #define vcmla_rot180_laneq_f16(__p0_222, __p1_222, __p2_222, __p3_222) __extension__ ({ \
37114   float16x4_t __ret_222; \
37115   float16x4_t __s0_222 = __p0_222; \
37116   float16x4_t __s1_222 = __p1_222; \
37117   float16x8_t __s2_222 = __p2_222; \
37118   float16x4_t __rev0_222;  __rev0_222 = __builtin_shufflevector(__s0_222, __s0_222, 3, 2, 1, 0); \
37119   float16x4_t __rev1_222;  __rev1_222 = __builtin_shufflevector(__s1_222, __s1_222, 3, 2, 1, 0); \
37120   float16x8_t __rev2_222;  __rev2_222 = __builtin_shufflevector(__s2_222, __s2_222, 7, 6, 5, 4, 3, 2, 1, 0); \
37121 float16x8_t __reint_222 = __rev2_222; \
37122 uint32x2_t __reint1_222 = (uint32x2_t) {__noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_222, __p3_222), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_222, __p3_222)}; \
37123   __ret_222 = __noswap_vcmla_rot180_f16(__rev0_222, __rev1_222, *(float16x4_t *) &__reint1_222); \
37124   __ret_222 = __builtin_shufflevector(__ret_222, __ret_222, 3, 2, 1, 0); \
37125   __ret_222; \
37126 })
37127 #endif
37128 
37129 #ifdef __LITTLE_ENDIAN__
37130 #define vcmlaq_rot180_laneq_f16(__p0_223, __p1_223, __p2_223, __p3_223) __extension__ ({ \
37131   float16x8_t __ret_223; \
37132   float16x8_t __s0_223 = __p0_223; \
37133   float16x8_t __s1_223 = __p1_223; \
37134   float16x8_t __s2_223 = __p2_223; \
37135 float16x8_t __reint_223 = __s2_223; \
37136 uint32x4_t __reint1_223 = (uint32x4_t) {vgetq_lane_u32(*(uint32x4_t *) &__reint_223, __p3_223), vgetq_lane_u32(*(uint32x4_t *) &__reint_223, __p3_223), vgetq_lane_u32(*(uint32x4_t *) &__reint_223, __p3_223), vgetq_lane_u32(*(uint32x4_t *) &__reint_223, __p3_223)}; \
37137   __ret_223 = vcmlaq_rot180_f16(__s0_223, __s1_223, *(float16x8_t *) &__reint1_223); \
37138   __ret_223; \
37139 })
37140 #else
37141 #define vcmlaq_rot180_laneq_f16(__p0_224, __p1_224, __p2_224, __p3_224) __extension__ ({ \
37142   float16x8_t __ret_224; \
37143   float16x8_t __s0_224 = __p0_224; \
37144   float16x8_t __s1_224 = __p1_224; \
37145   float16x8_t __s2_224 = __p2_224; \
37146   float16x8_t __rev0_224;  __rev0_224 = __builtin_shufflevector(__s0_224, __s0_224, 7, 6, 5, 4, 3, 2, 1, 0); \
37147   float16x8_t __rev1_224;  __rev1_224 = __builtin_shufflevector(__s1_224, __s1_224, 7, 6, 5, 4, 3, 2, 1, 0); \
37148   float16x8_t __rev2_224;  __rev2_224 = __builtin_shufflevector(__s2_224, __s2_224, 7, 6, 5, 4, 3, 2, 1, 0); \
37149 float16x8_t __reint_224 = __rev2_224; \
37150 uint32x4_t __reint1_224 = (uint32x4_t) {__noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_224, __p3_224), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_224, __p3_224), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_224, __p3_224), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_224, __p3_224)}; \
37151   __ret_224 = __noswap_vcmlaq_rot180_f16(__rev0_224, __rev1_224, *(float16x8_t *) &__reint1_224); \
37152   __ret_224 = __builtin_shufflevector(__ret_224, __ret_224, 7, 6, 5, 4, 3, 2, 1, 0); \
37153   __ret_224; \
37154 })
37155 #endif
37156 
37157 #ifdef __LITTLE_ENDIAN__
vcmlaq_rot270_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)37158 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcmlaq_rot270_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
37159   float16x8_t __ret;
37160   __ret = (float16x8_t) __builtin_neon_vcmlaq_rot270_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
37161   return __ret;
37162 }
37163 #else
vcmlaq_rot270_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)37164 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcmlaq_rot270_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
37165   float16x8_t __ret;
37166   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
37167   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
37168   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
37169   __ret = (float16x8_t) __builtin_neon_vcmlaq_rot270_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 40);
37170   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
37171   return __ret;
37172 }
__noswap_vcmlaq_rot270_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)37173 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t __noswap_vcmlaq_rot270_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
37174   float16x8_t __ret;
37175   __ret = (float16x8_t) __builtin_neon_vcmlaq_rot270_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
37176   return __ret;
37177 }
37178 #endif
37179 
37180 #ifdef __LITTLE_ENDIAN__
vcmla_rot270_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)37181 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcmla_rot270_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
37182   float16x4_t __ret;
37183   __ret = (float16x4_t) __builtin_neon_vcmla_rot270_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
37184   return __ret;
37185 }
37186 #else
vcmla_rot270_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)37187 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcmla_rot270_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
37188   float16x4_t __ret;
37189   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
37190   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
37191   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
37192   __ret = (float16x4_t) __builtin_neon_vcmla_rot270_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 8);
37193   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
37194   return __ret;
37195 }
__noswap_vcmla_rot270_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)37196 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t __noswap_vcmla_rot270_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
37197   float16x4_t __ret;
37198   __ret = (float16x4_t) __builtin_neon_vcmla_rot270_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
37199   return __ret;
37200 }
37201 #endif
37202 
37203 #ifdef __LITTLE_ENDIAN__
37204 #define vcmla_rot270_lane_f16(__p0_225, __p1_225, __p2_225, __p3_225) __extension__ ({ \
37205   float16x4_t __ret_225; \
37206   float16x4_t __s0_225 = __p0_225; \
37207   float16x4_t __s1_225 = __p1_225; \
37208   float16x4_t __s2_225 = __p2_225; \
37209 float16x4_t __reint_225 = __s2_225; \
37210 uint32x2_t __reint1_225 = (uint32x2_t) {vget_lane_u32(*(uint32x2_t *) &__reint_225, __p3_225), vget_lane_u32(*(uint32x2_t *) &__reint_225, __p3_225)}; \
37211   __ret_225 = vcmla_rot270_f16(__s0_225, __s1_225, *(float16x4_t *) &__reint1_225); \
37212   __ret_225; \
37213 })
37214 #else
37215 #define vcmla_rot270_lane_f16(__p0_226, __p1_226, __p2_226, __p3_226) __extension__ ({ \
37216   float16x4_t __ret_226; \
37217   float16x4_t __s0_226 = __p0_226; \
37218   float16x4_t __s1_226 = __p1_226; \
37219   float16x4_t __s2_226 = __p2_226; \
37220   float16x4_t __rev0_226;  __rev0_226 = __builtin_shufflevector(__s0_226, __s0_226, 3, 2, 1, 0); \
37221   float16x4_t __rev1_226;  __rev1_226 = __builtin_shufflevector(__s1_226, __s1_226, 3, 2, 1, 0); \
37222   float16x4_t __rev2_226;  __rev2_226 = __builtin_shufflevector(__s2_226, __s2_226, 3, 2, 1, 0); \
37223 float16x4_t __reint_226 = __rev2_226; \
37224 uint32x2_t __reint1_226 = (uint32x2_t) {__noswap_vget_lane_u32(*(uint32x2_t *) &__reint_226, __p3_226), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_226, __p3_226)}; \
37225   __ret_226 = __noswap_vcmla_rot270_f16(__rev0_226, __rev1_226, *(float16x4_t *) &__reint1_226); \
37226   __ret_226 = __builtin_shufflevector(__ret_226, __ret_226, 3, 2, 1, 0); \
37227   __ret_226; \
37228 })
37229 #endif
37230 
37231 #ifdef __LITTLE_ENDIAN__
37232 #define vcmlaq_rot270_lane_f16(__p0_227, __p1_227, __p2_227, __p3_227) __extension__ ({ \
37233   float16x8_t __ret_227; \
37234   float16x8_t __s0_227 = __p0_227; \
37235   float16x8_t __s1_227 = __p1_227; \
37236   float16x4_t __s2_227 = __p2_227; \
37237 float16x4_t __reint_227 = __s2_227; \
37238 uint32x4_t __reint1_227 = (uint32x4_t) {vget_lane_u32(*(uint32x2_t *) &__reint_227, __p3_227), vget_lane_u32(*(uint32x2_t *) &__reint_227, __p3_227), vget_lane_u32(*(uint32x2_t *) &__reint_227, __p3_227), vget_lane_u32(*(uint32x2_t *) &__reint_227, __p3_227)}; \
37239   __ret_227 = vcmlaq_rot270_f16(__s0_227, __s1_227, *(float16x8_t *) &__reint1_227); \
37240   __ret_227; \
37241 })
37242 #else
37243 #define vcmlaq_rot270_lane_f16(__p0_228, __p1_228, __p2_228, __p3_228) __extension__ ({ \
37244   float16x8_t __ret_228; \
37245   float16x8_t __s0_228 = __p0_228; \
37246   float16x8_t __s1_228 = __p1_228; \
37247   float16x4_t __s2_228 = __p2_228; \
37248   float16x8_t __rev0_228;  __rev0_228 = __builtin_shufflevector(__s0_228, __s0_228, 7, 6, 5, 4, 3, 2, 1, 0); \
37249   float16x8_t __rev1_228;  __rev1_228 = __builtin_shufflevector(__s1_228, __s1_228, 7, 6, 5, 4, 3, 2, 1, 0); \
37250   float16x4_t __rev2_228;  __rev2_228 = __builtin_shufflevector(__s2_228, __s2_228, 3, 2, 1, 0); \
37251 float16x4_t __reint_228 = __rev2_228; \
37252 uint32x4_t __reint1_228 = (uint32x4_t) {__noswap_vget_lane_u32(*(uint32x2_t *) &__reint_228, __p3_228), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_228, __p3_228), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_228, __p3_228), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_228, __p3_228)}; \
37253   __ret_228 = __noswap_vcmlaq_rot270_f16(__rev0_228, __rev1_228, *(float16x8_t *) &__reint1_228); \
37254   __ret_228 = __builtin_shufflevector(__ret_228, __ret_228, 7, 6, 5, 4, 3, 2, 1, 0); \
37255   __ret_228; \
37256 })
37257 #endif
37258 
37259 #ifdef __LITTLE_ENDIAN__
37260 #define vcmla_rot270_laneq_f16(__p0_229, __p1_229, __p2_229, __p3_229) __extension__ ({ \
37261   float16x4_t __ret_229; \
37262   float16x4_t __s0_229 = __p0_229; \
37263   float16x4_t __s1_229 = __p1_229; \
37264   float16x8_t __s2_229 = __p2_229; \
37265 float16x8_t __reint_229 = __s2_229; \
37266 uint32x2_t __reint1_229 = (uint32x2_t) {vgetq_lane_u32(*(uint32x4_t *) &__reint_229, __p3_229), vgetq_lane_u32(*(uint32x4_t *) &__reint_229, __p3_229)}; \
37267   __ret_229 = vcmla_rot270_f16(__s0_229, __s1_229, *(float16x4_t *) &__reint1_229); \
37268   __ret_229; \
37269 })
37270 #else
37271 #define vcmla_rot270_laneq_f16(__p0_230, __p1_230, __p2_230, __p3_230) __extension__ ({ \
37272   float16x4_t __ret_230; \
37273   float16x4_t __s0_230 = __p0_230; \
37274   float16x4_t __s1_230 = __p1_230; \
37275   float16x8_t __s2_230 = __p2_230; \
37276   float16x4_t __rev0_230;  __rev0_230 = __builtin_shufflevector(__s0_230, __s0_230, 3, 2, 1, 0); \
37277   float16x4_t __rev1_230;  __rev1_230 = __builtin_shufflevector(__s1_230, __s1_230, 3, 2, 1, 0); \
37278   float16x8_t __rev2_230;  __rev2_230 = __builtin_shufflevector(__s2_230, __s2_230, 7, 6, 5, 4, 3, 2, 1, 0); \
37279 float16x8_t __reint_230 = __rev2_230; \
37280 uint32x2_t __reint1_230 = (uint32x2_t) {__noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_230, __p3_230), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_230, __p3_230)}; \
37281   __ret_230 = __noswap_vcmla_rot270_f16(__rev0_230, __rev1_230, *(float16x4_t *) &__reint1_230); \
37282   __ret_230 = __builtin_shufflevector(__ret_230, __ret_230, 3, 2, 1, 0); \
37283   __ret_230; \
37284 })
37285 #endif
37286 
37287 #ifdef __LITTLE_ENDIAN__
37288 #define vcmlaq_rot270_laneq_f16(__p0_231, __p1_231, __p2_231, __p3_231) __extension__ ({ \
37289   float16x8_t __ret_231; \
37290   float16x8_t __s0_231 = __p0_231; \
37291   float16x8_t __s1_231 = __p1_231; \
37292   float16x8_t __s2_231 = __p2_231; \
37293 float16x8_t __reint_231 = __s2_231; \
37294 uint32x4_t __reint1_231 = (uint32x4_t) {vgetq_lane_u32(*(uint32x4_t *) &__reint_231, __p3_231), vgetq_lane_u32(*(uint32x4_t *) &__reint_231, __p3_231), vgetq_lane_u32(*(uint32x4_t *) &__reint_231, __p3_231), vgetq_lane_u32(*(uint32x4_t *) &__reint_231, __p3_231)}; \
37295   __ret_231 = vcmlaq_rot270_f16(__s0_231, __s1_231, *(float16x8_t *) &__reint1_231); \
37296   __ret_231; \
37297 })
37298 #else
37299 #define vcmlaq_rot270_laneq_f16(__p0_232, __p1_232, __p2_232, __p3_232) __extension__ ({ \
37300   float16x8_t __ret_232; \
37301   float16x8_t __s0_232 = __p0_232; \
37302   float16x8_t __s1_232 = __p1_232; \
37303   float16x8_t __s2_232 = __p2_232; \
37304   float16x8_t __rev0_232;  __rev0_232 = __builtin_shufflevector(__s0_232, __s0_232, 7, 6, 5, 4, 3, 2, 1, 0); \
37305   float16x8_t __rev1_232;  __rev1_232 = __builtin_shufflevector(__s1_232, __s1_232, 7, 6, 5, 4, 3, 2, 1, 0); \
37306   float16x8_t __rev2_232;  __rev2_232 = __builtin_shufflevector(__s2_232, __s2_232, 7, 6, 5, 4, 3, 2, 1, 0); \
37307 float16x8_t __reint_232 = __rev2_232; \
37308 uint32x4_t __reint1_232 = (uint32x4_t) {__noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_232, __p3_232), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_232, __p3_232), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_232, __p3_232), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_232, __p3_232)}; \
37309   __ret_232 = __noswap_vcmlaq_rot270_f16(__rev0_232, __rev1_232, *(float16x8_t *) &__reint1_232); \
37310   __ret_232 = __builtin_shufflevector(__ret_232, __ret_232, 7, 6, 5, 4, 3, 2, 1, 0); \
37311   __ret_232; \
37312 })
37313 #endif
37314 
37315 #ifdef __LITTLE_ENDIAN__
vcmlaq_rot90_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)37316 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcmlaq_rot90_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
37317   float16x8_t __ret;
37318   __ret = (float16x8_t) __builtin_neon_vcmlaq_rot90_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
37319   return __ret;
37320 }
37321 #else
vcmlaq_rot90_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)37322 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t vcmlaq_rot90_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
37323   float16x8_t __ret;
37324   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
37325   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
37326   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
37327   __ret = (float16x8_t) __builtin_neon_vcmlaq_rot90_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 40);
37328   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
37329   return __ret;
37330 }
__noswap_vcmlaq_rot90_f16(float16x8_t __p0,float16x8_t __p1,float16x8_t __p2)37331 __ai __attribute__((target("v8.3a,fullfp16"))) float16x8_t __noswap_vcmlaq_rot90_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
37332   float16x8_t __ret;
37333   __ret = (float16x8_t) __builtin_neon_vcmlaq_rot90_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
37334   return __ret;
37335 }
37336 #endif
37337 
37338 #ifdef __LITTLE_ENDIAN__
vcmla_rot90_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)37339 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcmla_rot90_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
37340   float16x4_t __ret;
37341   __ret = (float16x4_t) __builtin_neon_vcmla_rot90_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
37342   return __ret;
37343 }
37344 #else
vcmla_rot90_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)37345 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t vcmla_rot90_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
37346   float16x4_t __ret;
37347   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
37348   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
37349   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
37350   __ret = (float16x4_t) __builtin_neon_vcmla_rot90_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 8);
37351   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
37352   return __ret;
37353 }
__noswap_vcmla_rot90_f16(float16x4_t __p0,float16x4_t __p1,float16x4_t __p2)37354 __ai __attribute__((target("v8.3a,fullfp16"))) float16x4_t __noswap_vcmla_rot90_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
37355   float16x4_t __ret;
37356   __ret = (float16x4_t) __builtin_neon_vcmla_rot90_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
37357   return __ret;
37358 }
37359 #endif
37360 
37361 #ifdef __LITTLE_ENDIAN__
37362 #define vcmla_rot90_lane_f16(__p0_233, __p1_233, __p2_233, __p3_233) __extension__ ({ \
37363   float16x4_t __ret_233; \
37364   float16x4_t __s0_233 = __p0_233; \
37365   float16x4_t __s1_233 = __p1_233; \
37366   float16x4_t __s2_233 = __p2_233; \
37367 float16x4_t __reint_233 = __s2_233; \
37368 uint32x2_t __reint1_233 = (uint32x2_t) {vget_lane_u32(*(uint32x2_t *) &__reint_233, __p3_233), vget_lane_u32(*(uint32x2_t *) &__reint_233, __p3_233)}; \
37369   __ret_233 = vcmla_rot90_f16(__s0_233, __s1_233, *(float16x4_t *) &__reint1_233); \
37370   __ret_233; \
37371 })
37372 #else
37373 #define vcmla_rot90_lane_f16(__p0_234, __p1_234, __p2_234, __p3_234) __extension__ ({ \
37374   float16x4_t __ret_234; \
37375   float16x4_t __s0_234 = __p0_234; \
37376   float16x4_t __s1_234 = __p1_234; \
37377   float16x4_t __s2_234 = __p2_234; \
37378   float16x4_t __rev0_234;  __rev0_234 = __builtin_shufflevector(__s0_234, __s0_234, 3, 2, 1, 0); \
37379   float16x4_t __rev1_234;  __rev1_234 = __builtin_shufflevector(__s1_234, __s1_234, 3, 2, 1, 0); \
37380   float16x4_t __rev2_234;  __rev2_234 = __builtin_shufflevector(__s2_234, __s2_234, 3, 2, 1, 0); \
37381 float16x4_t __reint_234 = __rev2_234; \
37382 uint32x2_t __reint1_234 = (uint32x2_t) {__noswap_vget_lane_u32(*(uint32x2_t *) &__reint_234, __p3_234), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_234, __p3_234)}; \
37383   __ret_234 = __noswap_vcmla_rot90_f16(__rev0_234, __rev1_234, *(float16x4_t *) &__reint1_234); \
37384   __ret_234 = __builtin_shufflevector(__ret_234, __ret_234, 3, 2, 1, 0); \
37385   __ret_234; \
37386 })
37387 #endif
37388 
37389 #ifdef __LITTLE_ENDIAN__
37390 #define vcmlaq_rot90_lane_f16(__p0_235, __p1_235, __p2_235, __p3_235) __extension__ ({ \
37391   float16x8_t __ret_235; \
37392   float16x8_t __s0_235 = __p0_235; \
37393   float16x8_t __s1_235 = __p1_235; \
37394   float16x4_t __s2_235 = __p2_235; \
37395 float16x4_t __reint_235 = __s2_235; \
37396 uint32x4_t __reint1_235 = (uint32x4_t) {vget_lane_u32(*(uint32x2_t *) &__reint_235, __p3_235), vget_lane_u32(*(uint32x2_t *) &__reint_235, __p3_235), vget_lane_u32(*(uint32x2_t *) &__reint_235, __p3_235), vget_lane_u32(*(uint32x2_t *) &__reint_235, __p3_235)}; \
37397   __ret_235 = vcmlaq_rot90_f16(__s0_235, __s1_235, *(float16x8_t *) &__reint1_235); \
37398   __ret_235; \
37399 })
37400 #else
37401 #define vcmlaq_rot90_lane_f16(__p0_236, __p1_236, __p2_236, __p3_236) __extension__ ({ \
37402   float16x8_t __ret_236; \
37403   float16x8_t __s0_236 = __p0_236; \
37404   float16x8_t __s1_236 = __p1_236; \
37405   float16x4_t __s2_236 = __p2_236; \
37406   float16x8_t __rev0_236;  __rev0_236 = __builtin_shufflevector(__s0_236, __s0_236, 7, 6, 5, 4, 3, 2, 1, 0); \
37407   float16x8_t __rev1_236;  __rev1_236 = __builtin_shufflevector(__s1_236, __s1_236, 7, 6, 5, 4, 3, 2, 1, 0); \
37408   float16x4_t __rev2_236;  __rev2_236 = __builtin_shufflevector(__s2_236, __s2_236, 3, 2, 1, 0); \
37409 float16x4_t __reint_236 = __rev2_236; \
37410 uint32x4_t __reint1_236 = (uint32x4_t) {__noswap_vget_lane_u32(*(uint32x2_t *) &__reint_236, __p3_236), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_236, __p3_236), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_236, __p3_236), __noswap_vget_lane_u32(*(uint32x2_t *) &__reint_236, __p3_236)}; \
37411   __ret_236 = __noswap_vcmlaq_rot90_f16(__rev0_236, __rev1_236, *(float16x8_t *) &__reint1_236); \
37412   __ret_236 = __builtin_shufflevector(__ret_236, __ret_236, 7, 6, 5, 4, 3, 2, 1, 0); \
37413   __ret_236; \
37414 })
37415 #endif
37416 
37417 #ifdef __LITTLE_ENDIAN__
37418 #define vcmla_rot90_laneq_f16(__p0_237, __p1_237, __p2_237, __p3_237) __extension__ ({ \
37419   float16x4_t __ret_237; \
37420   float16x4_t __s0_237 = __p0_237; \
37421   float16x4_t __s1_237 = __p1_237; \
37422   float16x8_t __s2_237 = __p2_237; \
37423 float16x8_t __reint_237 = __s2_237; \
37424 uint32x2_t __reint1_237 = (uint32x2_t) {vgetq_lane_u32(*(uint32x4_t *) &__reint_237, __p3_237), vgetq_lane_u32(*(uint32x4_t *) &__reint_237, __p3_237)}; \
37425   __ret_237 = vcmla_rot90_f16(__s0_237, __s1_237, *(float16x4_t *) &__reint1_237); \
37426   __ret_237; \
37427 })
37428 #else
37429 #define vcmla_rot90_laneq_f16(__p0_238, __p1_238, __p2_238, __p3_238) __extension__ ({ \
37430   float16x4_t __ret_238; \
37431   float16x4_t __s0_238 = __p0_238; \
37432   float16x4_t __s1_238 = __p1_238; \
37433   float16x8_t __s2_238 = __p2_238; \
37434   float16x4_t __rev0_238;  __rev0_238 = __builtin_shufflevector(__s0_238, __s0_238, 3, 2, 1, 0); \
37435   float16x4_t __rev1_238;  __rev1_238 = __builtin_shufflevector(__s1_238, __s1_238, 3, 2, 1, 0); \
37436   float16x8_t __rev2_238;  __rev2_238 = __builtin_shufflevector(__s2_238, __s2_238, 7, 6, 5, 4, 3, 2, 1, 0); \
37437 float16x8_t __reint_238 = __rev2_238; \
37438 uint32x2_t __reint1_238 = (uint32x2_t) {__noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_238, __p3_238), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_238, __p3_238)}; \
37439   __ret_238 = __noswap_vcmla_rot90_f16(__rev0_238, __rev1_238, *(float16x4_t *) &__reint1_238); \
37440   __ret_238 = __builtin_shufflevector(__ret_238, __ret_238, 3, 2, 1, 0); \
37441   __ret_238; \
37442 })
37443 #endif
37444 
37445 #ifdef __LITTLE_ENDIAN__
37446 #define vcmlaq_rot90_laneq_f16(__p0_239, __p1_239, __p2_239, __p3_239) __extension__ ({ \
37447   float16x8_t __ret_239; \
37448   float16x8_t __s0_239 = __p0_239; \
37449   float16x8_t __s1_239 = __p1_239; \
37450   float16x8_t __s2_239 = __p2_239; \
37451 float16x8_t __reint_239 = __s2_239; \
37452 uint32x4_t __reint1_239 = (uint32x4_t) {vgetq_lane_u32(*(uint32x4_t *) &__reint_239, __p3_239), vgetq_lane_u32(*(uint32x4_t *) &__reint_239, __p3_239), vgetq_lane_u32(*(uint32x4_t *) &__reint_239, __p3_239), vgetq_lane_u32(*(uint32x4_t *) &__reint_239, __p3_239)}; \
37453   __ret_239 = vcmlaq_rot90_f16(__s0_239, __s1_239, *(float16x8_t *) &__reint1_239); \
37454   __ret_239; \
37455 })
37456 #else
37457 #define vcmlaq_rot90_laneq_f16(__p0_240, __p1_240, __p2_240, __p3_240) __extension__ ({ \
37458   float16x8_t __ret_240; \
37459   float16x8_t __s0_240 = __p0_240; \
37460   float16x8_t __s1_240 = __p1_240; \
37461   float16x8_t __s2_240 = __p2_240; \
37462   float16x8_t __rev0_240;  __rev0_240 = __builtin_shufflevector(__s0_240, __s0_240, 7, 6, 5, 4, 3, 2, 1, 0); \
37463   float16x8_t __rev1_240;  __rev1_240 = __builtin_shufflevector(__s1_240, __s1_240, 7, 6, 5, 4, 3, 2, 1, 0); \
37464   float16x8_t __rev2_240;  __rev2_240 = __builtin_shufflevector(__s2_240, __s2_240, 7, 6, 5, 4, 3, 2, 1, 0); \
37465 float16x8_t __reint_240 = __rev2_240; \
37466 uint32x4_t __reint1_240 = (uint32x4_t) {__noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_240, __p3_240), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_240, __p3_240), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_240, __p3_240), __noswap_vgetq_lane_u32(*(uint32x4_t *) &__reint_240, __p3_240)}; \
37467   __ret_240 = __noswap_vcmlaq_rot90_f16(__rev0_240, __rev1_240, *(float16x8_t *) &__reint1_240); \
37468   __ret_240 = __builtin_shufflevector(__ret_240, __ret_240, 7, 6, 5, 4, 3, 2, 1, 0); \
37469   __ret_240; \
37470 })
37471 #endif
37472 
37473 #if !defined(__aarch64__)
37474 #ifdef __LITTLE_ENDIAN__
37475 #define vqdmulhq_lane_s32(__p0_241, __p1_241, __p2_241) __extension__ ({ \
37476   int32x4_t __ret_241; \
37477   int32x4_t __s0_241 = __p0_241; \
37478   int32x2_t __s1_241 = __p1_241; \
37479   __ret_241 = vqdmulhq_s32(__s0_241, splatq_lane_s32(__s1_241, __p2_241)); \
37480   __ret_241; \
37481 })
37482 #else
37483 #define vqdmulhq_lane_s32(__p0_242, __p1_242, __p2_242) __extension__ ({ \
37484   int32x4_t __ret_242; \
37485   int32x4_t __s0_242 = __p0_242; \
37486   int32x2_t __s1_242 = __p1_242; \
37487   int32x4_t __rev0_242;  __rev0_242 = __builtin_shufflevector(__s0_242, __s0_242, 3, 2, 1, 0); \
37488   int32x2_t __rev1_242;  __rev1_242 = __builtin_shufflevector(__s1_242, __s1_242, 1, 0); \
37489   __ret_242 = __noswap_vqdmulhq_s32(__rev0_242, __noswap_splatq_lane_s32(__rev1_242, __p2_242)); \
37490   __ret_242 = __builtin_shufflevector(__ret_242, __ret_242, 3, 2, 1, 0); \
37491   __ret_242; \
37492 })
37493 #endif
37494 
37495 #ifdef __LITTLE_ENDIAN__
37496 #define vqdmulhq_lane_s16(__p0_243, __p1_243, __p2_243) __extension__ ({ \
37497   int16x8_t __ret_243; \
37498   int16x8_t __s0_243 = __p0_243; \
37499   int16x4_t __s1_243 = __p1_243; \
37500   __ret_243 = vqdmulhq_s16(__s0_243, splatq_lane_s16(__s1_243, __p2_243)); \
37501   __ret_243; \
37502 })
37503 #else
37504 #define vqdmulhq_lane_s16(__p0_244, __p1_244, __p2_244) __extension__ ({ \
37505   int16x8_t __ret_244; \
37506   int16x8_t __s0_244 = __p0_244; \
37507   int16x4_t __s1_244 = __p1_244; \
37508   int16x8_t __rev0_244;  __rev0_244 = __builtin_shufflevector(__s0_244, __s0_244, 7, 6, 5, 4, 3, 2, 1, 0); \
37509   int16x4_t __rev1_244;  __rev1_244 = __builtin_shufflevector(__s1_244, __s1_244, 3, 2, 1, 0); \
37510   __ret_244 = __noswap_vqdmulhq_s16(__rev0_244, __noswap_splatq_lane_s16(__rev1_244, __p2_244)); \
37511   __ret_244 = __builtin_shufflevector(__ret_244, __ret_244, 7, 6, 5, 4, 3, 2, 1, 0); \
37512   __ret_244; \
37513 })
37514 #endif
37515 
37516 #ifdef __LITTLE_ENDIAN__
37517 #define vqdmulh_lane_s32(__p0_245, __p1_245, __p2_245) __extension__ ({ \
37518   int32x2_t __ret_245; \
37519   int32x2_t __s0_245 = __p0_245; \
37520   int32x2_t __s1_245 = __p1_245; \
37521   __ret_245 = vqdmulh_s32(__s0_245, splat_lane_s32(__s1_245, __p2_245)); \
37522   __ret_245; \
37523 })
37524 #else
37525 #define vqdmulh_lane_s32(__p0_246, __p1_246, __p2_246) __extension__ ({ \
37526   int32x2_t __ret_246; \
37527   int32x2_t __s0_246 = __p0_246; \
37528   int32x2_t __s1_246 = __p1_246; \
37529   int32x2_t __rev0_246;  __rev0_246 = __builtin_shufflevector(__s0_246, __s0_246, 1, 0); \
37530   int32x2_t __rev1_246;  __rev1_246 = __builtin_shufflevector(__s1_246, __s1_246, 1, 0); \
37531   __ret_246 = __noswap_vqdmulh_s32(__rev0_246, __noswap_splat_lane_s32(__rev1_246, __p2_246)); \
37532   __ret_246 = __builtin_shufflevector(__ret_246, __ret_246, 1, 0); \
37533   __ret_246; \
37534 })
37535 #endif
37536 
37537 #ifdef __LITTLE_ENDIAN__
37538 #define vqdmulh_lane_s16(__p0_247, __p1_247, __p2_247) __extension__ ({ \
37539   int16x4_t __ret_247; \
37540   int16x4_t __s0_247 = __p0_247; \
37541   int16x4_t __s1_247 = __p1_247; \
37542   __ret_247 = vqdmulh_s16(__s0_247, splat_lane_s16(__s1_247, __p2_247)); \
37543   __ret_247; \
37544 })
37545 #else
37546 #define vqdmulh_lane_s16(__p0_248, __p1_248, __p2_248) __extension__ ({ \
37547   int16x4_t __ret_248; \
37548   int16x4_t __s0_248 = __p0_248; \
37549   int16x4_t __s1_248 = __p1_248; \
37550   int16x4_t __rev0_248;  __rev0_248 = __builtin_shufflevector(__s0_248, __s0_248, 3, 2, 1, 0); \
37551   int16x4_t __rev1_248;  __rev1_248 = __builtin_shufflevector(__s1_248, __s1_248, 3, 2, 1, 0); \
37552   __ret_248 = __noswap_vqdmulh_s16(__rev0_248, __noswap_splat_lane_s16(__rev1_248, __p2_248)); \
37553   __ret_248 = __builtin_shufflevector(__ret_248, __ret_248, 3, 2, 1, 0); \
37554   __ret_248; \
37555 })
37556 #endif
37557 
37558 #ifdef __LITTLE_ENDIAN__
37559 #define vqrdmulhq_lane_s32(__p0_249, __p1_249, __p2_249) __extension__ ({ \
37560   int32x4_t __ret_249; \
37561   int32x4_t __s0_249 = __p0_249; \
37562   int32x2_t __s1_249 = __p1_249; \
37563   __ret_249 = vqrdmulhq_s32(__s0_249, splatq_lane_s32(__s1_249, __p2_249)); \
37564   __ret_249; \
37565 })
37566 #else
37567 #define vqrdmulhq_lane_s32(__p0_250, __p1_250, __p2_250) __extension__ ({ \
37568   int32x4_t __ret_250; \
37569   int32x4_t __s0_250 = __p0_250; \
37570   int32x2_t __s1_250 = __p1_250; \
37571   int32x4_t __rev0_250;  __rev0_250 = __builtin_shufflevector(__s0_250, __s0_250, 3, 2, 1, 0); \
37572   int32x2_t __rev1_250;  __rev1_250 = __builtin_shufflevector(__s1_250, __s1_250, 1, 0); \
37573   __ret_250 = __noswap_vqrdmulhq_s32(__rev0_250, __noswap_splatq_lane_s32(__rev1_250, __p2_250)); \
37574   __ret_250 = __builtin_shufflevector(__ret_250, __ret_250, 3, 2, 1, 0); \
37575   __ret_250; \
37576 })
37577 #endif
37578 
37579 #ifdef __LITTLE_ENDIAN__
37580 #define vqrdmulhq_lane_s16(__p0_251, __p1_251, __p2_251) __extension__ ({ \
37581   int16x8_t __ret_251; \
37582   int16x8_t __s0_251 = __p0_251; \
37583   int16x4_t __s1_251 = __p1_251; \
37584   __ret_251 = vqrdmulhq_s16(__s0_251, splatq_lane_s16(__s1_251, __p2_251)); \
37585   __ret_251; \
37586 })
37587 #else
37588 #define vqrdmulhq_lane_s16(__p0_252, __p1_252, __p2_252) __extension__ ({ \
37589   int16x8_t __ret_252; \
37590   int16x8_t __s0_252 = __p0_252; \
37591   int16x4_t __s1_252 = __p1_252; \
37592   int16x8_t __rev0_252;  __rev0_252 = __builtin_shufflevector(__s0_252, __s0_252, 7, 6, 5, 4, 3, 2, 1, 0); \
37593   int16x4_t __rev1_252;  __rev1_252 = __builtin_shufflevector(__s1_252, __s1_252, 3, 2, 1, 0); \
37594   __ret_252 = __noswap_vqrdmulhq_s16(__rev0_252, __noswap_splatq_lane_s16(__rev1_252, __p2_252)); \
37595   __ret_252 = __builtin_shufflevector(__ret_252, __ret_252, 7, 6, 5, 4, 3, 2, 1, 0); \
37596   __ret_252; \
37597 })
37598 #endif
37599 
37600 #ifdef __LITTLE_ENDIAN__
37601 #define vqrdmulh_lane_s32(__p0_253, __p1_253, __p2_253) __extension__ ({ \
37602   int32x2_t __ret_253; \
37603   int32x2_t __s0_253 = __p0_253; \
37604   int32x2_t __s1_253 = __p1_253; \
37605   __ret_253 = vqrdmulh_s32(__s0_253, splat_lane_s32(__s1_253, __p2_253)); \
37606   __ret_253; \
37607 })
37608 #else
37609 #define vqrdmulh_lane_s32(__p0_254, __p1_254, __p2_254) __extension__ ({ \
37610   int32x2_t __ret_254; \
37611   int32x2_t __s0_254 = __p0_254; \
37612   int32x2_t __s1_254 = __p1_254; \
37613   int32x2_t __rev0_254;  __rev0_254 = __builtin_shufflevector(__s0_254, __s0_254, 1, 0); \
37614   int32x2_t __rev1_254;  __rev1_254 = __builtin_shufflevector(__s1_254, __s1_254, 1, 0); \
37615   __ret_254 = __noswap_vqrdmulh_s32(__rev0_254, __noswap_splat_lane_s32(__rev1_254, __p2_254)); \
37616   __ret_254 = __builtin_shufflevector(__ret_254, __ret_254, 1, 0); \
37617   __ret_254; \
37618 })
37619 #endif
37620 
37621 #ifdef __LITTLE_ENDIAN__
37622 #define vqrdmulh_lane_s16(__p0_255, __p1_255, __p2_255) __extension__ ({ \
37623   int16x4_t __ret_255; \
37624   int16x4_t __s0_255 = __p0_255; \
37625   int16x4_t __s1_255 = __p1_255; \
37626   __ret_255 = vqrdmulh_s16(__s0_255, splat_lane_s16(__s1_255, __p2_255)); \
37627   __ret_255; \
37628 })
37629 #else
37630 #define vqrdmulh_lane_s16(__p0_256, __p1_256, __p2_256) __extension__ ({ \
37631   int16x4_t __ret_256; \
37632   int16x4_t __s0_256 = __p0_256; \
37633   int16x4_t __s1_256 = __p1_256; \
37634   int16x4_t __rev0_256;  __rev0_256 = __builtin_shufflevector(__s0_256, __s0_256, 3, 2, 1, 0); \
37635   int16x4_t __rev1_256;  __rev1_256 = __builtin_shufflevector(__s1_256, __s1_256, 3, 2, 1, 0); \
37636   __ret_256 = __noswap_vqrdmulh_s16(__rev0_256, __noswap_splat_lane_s16(__rev1_256, __p2_256)); \
37637   __ret_256 = __builtin_shufflevector(__ret_256, __ret_256, 3, 2, 1, 0); \
37638   __ret_256; \
37639 })
37640 #endif
37641 
vreinterpret_p8_p16(poly16x4_t __p0)37642 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
37643   poly8x8_t __ret;
37644   __ret = (poly8x8_t)(__p0);
37645   return __ret;
37646 }
vreinterpret_p8_u8(uint8x8_t __p0)37647 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
37648   poly8x8_t __ret;
37649   __ret = (poly8x8_t)(__p0);
37650   return __ret;
37651 }
vreinterpret_p8_u32(uint32x2_t __p0)37652 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
37653   poly8x8_t __ret;
37654   __ret = (poly8x8_t)(__p0);
37655   return __ret;
37656 }
vreinterpret_p8_u64(uint64x1_t __p0)37657 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
37658   poly8x8_t __ret;
37659   __ret = (poly8x8_t)(__p0);
37660   return __ret;
37661 }
vreinterpret_p8_u16(uint16x4_t __p0)37662 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
37663   poly8x8_t __ret;
37664   __ret = (poly8x8_t)(__p0);
37665   return __ret;
37666 }
vreinterpret_p8_s8(int8x8_t __p0)37667 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
37668   poly8x8_t __ret;
37669   __ret = (poly8x8_t)(__p0);
37670   return __ret;
37671 }
vreinterpret_p8_f32(float32x2_t __p0)37672 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
37673   poly8x8_t __ret;
37674   __ret = (poly8x8_t)(__p0);
37675   return __ret;
37676 }
vreinterpret_p8_f16(float16x4_t __p0)37677 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
37678   poly8x8_t __ret;
37679   __ret = (poly8x8_t)(__p0);
37680   return __ret;
37681 }
vreinterpret_p8_s32(int32x2_t __p0)37682 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
37683   poly8x8_t __ret;
37684   __ret = (poly8x8_t)(__p0);
37685   return __ret;
37686 }
vreinterpret_p8_s64(int64x1_t __p0)37687 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
37688   poly8x8_t __ret;
37689   __ret = (poly8x8_t)(__p0);
37690   return __ret;
37691 }
vreinterpret_p8_s16(int16x4_t __p0)37692 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
37693   poly8x8_t __ret;
37694   __ret = (poly8x8_t)(__p0);
37695   return __ret;
37696 }
vreinterpret_p16_p8(poly8x8_t __p0)37697 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
37698   poly16x4_t __ret;
37699   __ret = (poly16x4_t)(__p0);
37700   return __ret;
37701 }
vreinterpret_p16_u8(uint8x8_t __p0)37702 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
37703   poly16x4_t __ret;
37704   __ret = (poly16x4_t)(__p0);
37705   return __ret;
37706 }
vreinterpret_p16_u32(uint32x2_t __p0)37707 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
37708   poly16x4_t __ret;
37709   __ret = (poly16x4_t)(__p0);
37710   return __ret;
37711 }
vreinterpret_p16_u64(uint64x1_t __p0)37712 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
37713   poly16x4_t __ret;
37714   __ret = (poly16x4_t)(__p0);
37715   return __ret;
37716 }
vreinterpret_p16_u16(uint16x4_t __p0)37717 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
37718   poly16x4_t __ret;
37719   __ret = (poly16x4_t)(__p0);
37720   return __ret;
37721 }
vreinterpret_p16_s8(int8x8_t __p0)37722 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
37723   poly16x4_t __ret;
37724   __ret = (poly16x4_t)(__p0);
37725   return __ret;
37726 }
vreinterpret_p16_f32(float32x2_t __p0)37727 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
37728   poly16x4_t __ret;
37729   __ret = (poly16x4_t)(__p0);
37730   return __ret;
37731 }
vreinterpret_p16_f16(float16x4_t __p0)37732 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
37733   poly16x4_t __ret;
37734   __ret = (poly16x4_t)(__p0);
37735   return __ret;
37736 }
vreinterpret_p16_s32(int32x2_t __p0)37737 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
37738   poly16x4_t __ret;
37739   __ret = (poly16x4_t)(__p0);
37740   return __ret;
37741 }
vreinterpret_p16_s64(int64x1_t __p0)37742 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
37743   poly16x4_t __ret;
37744   __ret = (poly16x4_t)(__p0);
37745   return __ret;
37746 }
vreinterpret_p16_s16(int16x4_t __p0)37747 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
37748   poly16x4_t __ret;
37749   __ret = (poly16x4_t)(__p0);
37750   return __ret;
37751 }
vreinterpretq_p8_p16(poly16x8_t __p0)37752 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
37753   poly8x16_t __ret;
37754   __ret = (poly8x16_t)(__p0);
37755   return __ret;
37756 }
vreinterpretq_p8_u8(uint8x16_t __p0)37757 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
37758   poly8x16_t __ret;
37759   __ret = (poly8x16_t)(__p0);
37760   return __ret;
37761 }
vreinterpretq_p8_u32(uint32x4_t __p0)37762 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
37763   poly8x16_t __ret;
37764   __ret = (poly8x16_t)(__p0);
37765   return __ret;
37766 }
vreinterpretq_p8_u64(uint64x2_t __p0)37767 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
37768   poly8x16_t __ret;
37769   __ret = (poly8x16_t)(__p0);
37770   return __ret;
37771 }
vreinterpretq_p8_u16(uint16x8_t __p0)37772 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
37773   poly8x16_t __ret;
37774   __ret = (poly8x16_t)(__p0);
37775   return __ret;
37776 }
vreinterpretq_p8_s8(int8x16_t __p0)37777 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
37778   poly8x16_t __ret;
37779   __ret = (poly8x16_t)(__p0);
37780   return __ret;
37781 }
vreinterpretq_p8_f32(float32x4_t __p0)37782 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
37783   poly8x16_t __ret;
37784   __ret = (poly8x16_t)(__p0);
37785   return __ret;
37786 }
vreinterpretq_p8_f16(float16x8_t __p0)37787 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
37788   poly8x16_t __ret;
37789   __ret = (poly8x16_t)(__p0);
37790   return __ret;
37791 }
vreinterpretq_p8_s32(int32x4_t __p0)37792 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
37793   poly8x16_t __ret;
37794   __ret = (poly8x16_t)(__p0);
37795   return __ret;
37796 }
vreinterpretq_p8_s64(int64x2_t __p0)37797 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
37798   poly8x16_t __ret;
37799   __ret = (poly8x16_t)(__p0);
37800   return __ret;
37801 }
vreinterpretq_p8_s16(int16x8_t __p0)37802 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
37803   poly8x16_t __ret;
37804   __ret = (poly8x16_t)(__p0);
37805   return __ret;
37806 }
vreinterpretq_p16_p8(poly8x16_t __p0)37807 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
37808   poly16x8_t __ret;
37809   __ret = (poly16x8_t)(__p0);
37810   return __ret;
37811 }
vreinterpretq_p16_u8(uint8x16_t __p0)37812 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
37813   poly16x8_t __ret;
37814   __ret = (poly16x8_t)(__p0);
37815   return __ret;
37816 }
vreinterpretq_p16_u32(uint32x4_t __p0)37817 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
37818   poly16x8_t __ret;
37819   __ret = (poly16x8_t)(__p0);
37820   return __ret;
37821 }
vreinterpretq_p16_u64(uint64x2_t __p0)37822 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
37823   poly16x8_t __ret;
37824   __ret = (poly16x8_t)(__p0);
37825   return __ret;
37826 }
vreinterpretq_p16_u16(uint16x8_t __p0)37827 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
37828   poly16x8_t __ret;
37829   __ret = (poly16x8_t)(__p0);
37830   return __ret;
37831 }
vreinterpretq_p16_s8(int8x16_t __p0)37832 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
37833   poly16x8_t __ret;
37834   __ret = (poly16x8_t)(__p0);
37835   return __ret;
37836 }
vreinterpretq_p16_f32(float32x4_t __p0)37837 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
37838   poly16x8_t __ret;
37839   __ret = (poly16x8_t)(__p0);
37840   return __ret;
37841 }
vreinterpretq_p16_f16(float16x8_t __p0)37842 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
37843   poly16x8_t __ret;
37844   __ret = (poly16x8_t)(__p0);
37845   return __ret;
37846 }
vreinterpretq_p16_s32(int32x4_t __p0)37847 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
37848   poly16x8_t __ret;
37849   __ret = (poly16x8_t)(__p0);
37850   return __ret;
37851 }
vreinterpretq_p16_s64(int64x2_t __p0)37852 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
37853   poly16x8_t __ret;
37854   __ret = (poly16x8_t)(__p0);
37855   return __ret;
37856 }
vreinterpretq_p16_s16(int16x8_t __p0)37857 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
37858   poly16x8_t __ret;
37859   __ret = (poly16x8_t)(__p0);
37860   return __ret;
37861 }
vreinterpretq_u8_p8(poly8x16_t __p0)37862 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
37863   uint8x16_t __ret;
37864   __ret = (uint8x16_t)(__p0);
37865   return __ret;
37866 }
vreinterpretq_u8_p16(poly16x8_t __p0)37867 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
37868   uint8x16_t __ret;
37869   __ret = (uint8x16_t)(__p0);
37870   return __ret;
37871 }
vreinterpretq_u8_u32(uint32x4_t __p0)37872 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
37873   uint8x16_t __ret;
37874   __ret = (uint8x16_t)(__p0);
37875   return __ret;
37876 }
vreinterpretq_u8_u64(uint64x2_t __p0)37877 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
37878   uint8x16_t __ret;
37879   __ret = (uint8x16_t)(__p0);
37880   return __ret;
37881 }
vreinterpretq_u8_u16(uint16x8_t __p0)37882 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
37883   uint8x16_t __ret;
37884   __ret = (uint8x16_t)(__p0);
37885   return __ret;
37886 }
vreinterpretq_u8_s8(int8x16_t __p0)37887 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
37888   uint8x16_t __ret;
37889   __ret = (uint8x16_t)(__p0);
37890   return __ret;
37891 }
vreinterpretq_u8_f32(float32x4_t __p0)37892 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
37893   uint8x16_t __ret;
37894   __ret = (uint8x16_t)(__p0);
37895   return __ret;
37896 }
vreinterpretq_u8_f16(float16x8_t __p0)37897 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
37898   uint8x16_t __ret;
37899   __ret = (uint8x16_t)(__p0);
37900   return __ret;
37901 }
vreinterpretq_u8_s32(int32x4_t __p0)37902 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
37903   uint8x16_t __ret;
37904   __ret = (uint8x16_t)(__p0);
37905   return __ret;
37906 }
vreinterpretq_u8_s64(int64x2_t __p0)37907 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
37908   uint8x16_t __ret;
37909   __ret = (uint8x16_t)(__p0);
37910   return __ret;
37911 }
vreinterpretq_u8_s16(int16x8_t __p0)37912 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
37913   uint8x16_t __ret;
37914   __ret = (uint8x16_t)(__p0);
37915   return __ret;
37916 }
vreinterpretq_u32_p8(poly8x16_t __p0)37917 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
37918   uint32x4_t __ret;
37919   __ret = (uint32x4_t)(__p0);
37920   return __ret;
37921 }
vreinterpretq_u32_p16(poly16x8_t __p0)37922 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
37923   uint32x4_t __ret;
37924   __ret = (uint32x4_t)(__p0);
37925   return __ret;
37926 }
vreinterpretq_u32_u8(uint8x16_t __p0)37927 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
37928   uint32x4_t __ret;
37929   __ret = (uint32x4_t)(__p0);
37930   return __ret;
37931 }
vreinterpretq_u32_u64(uint64x2_t __p0)37932 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
37933   uint32x4_t __ret;
37934   __ret = (uint32x4_t)(__p0);
37935   return __ret;
37936 }
vreinterpretq_u32_u16(uint16x8_t __p0)37937 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
37938   uint32x4_t __ret;
37939   __ret = (uint32x4_t)(__p0);
37940   return __ret;
37941 }
vreinterpretq_u32_s8(int8x16_t __p0)37942 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
37943   uint32x4_t __ret;
37944   __ret = (uint32x4_t)(__p0);
37945   return __ret;
37946 }
vreinterpretq_u32_f32(float32x4_t __p0)37947 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
37948   uint32x4_t __ret;
37949   __ret = (uint32x4_t)(__p0);
37950   return __ret;
37951 }
vreinterpretq_u32_f16(float16x8_t __p0)37952 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
37953   uint32x4_t __ret;
37954   __ret = (uint32x4_t)(__p0);
37955   return __ret;
37956 }
vreinterpretq_u32_s32(int32x4_t __p0)37957 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
37958   uint32x4_t __ret;
37959   __ret = (uint32x4_t)(__p0);
37960   return __ret;
37961 }
vreinterpretq_u32_s64(int64x2_t __p0)37962 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
37963   uint32x4_t __ret;
37964   __ret = (uint32x4_t)(__p0);
37965   return __ret;
37966 }
vreinterpretq_u32_s16(int16x8_t __p0)37967 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
37968   uint32x4_t __ret;
37969   __ret = (uint32x4_t)(__p0);
37970   return __ret;
37971 }
vreinterpretq_u64_p8(poly8x16_t __p0)37972 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
37973   uint64x2_t __ret;
37974   __ret = (uint64x2_t)(__p0);
37975   return __ret;
37976 }
vreinterpretq_u64_p16(poly16x8_t __p0)37977 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
37978   uint64x2_t __ret;
37979   __ret = (uint64x2_t)(__p0);
37980   return __ret;
37981 }
vreinterpretq_u64_u8(uint8x16_t __p0)37982 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
37983   uint64x2_t __ret;
37984   __ret = (uint64x2_t)(__p0);
37985   return __ret;
37986 }
vreinterpretq_u64_u32(uint32x4_t __p0)37987 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
37988   uint64x2_t __ret;
37989   __ret = (uint64x2_t)(__p0);
37990   return __ret;
37991 }
vreinterpretq_u64_u16(uint16x8_t __p0)37992 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
37993   uint64x2_t __ret;
37994   __ret = (uint64x2_t)(__p0);
37995   return __ret;
37996 }
vreinterpretq_u64_s8(int8x16_t __p0)37997 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
37998   uint64x2_t __ret;
37999   __ret = (uint64x2_t)(__p0);
38000   return __ret;
38001 }
vreinterpretq_u64_f32(float32x4_t __p0)38002 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
38003   uint64x2_t __ret;
38004   __ret = (uint64x2_t)(__p0);
38005   return __ret;
38006 }
vreinterpretq_u64_f16(float16x8_t __p0)38007 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
38008   uint64x2_t __ret;
38009   __ret = (uint64x2_t)(__p0);
38010   return __ret;
38011 }
vreinterpretq_u64_s32(int32x4_t __p0)38012 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
38013   uint64x2_t __ret;
38014   __ret = (uint64x2_t)(__p0);
38015   return __ret;
38016 }
vreinterpretq_u64_s64(int64x2_t __p0)38017 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
38018   uint64x2_t __ret;
38019   __ret = (uint64x2_t)(__p0);
38020   return __ret;
38021 }
vreinterpretq_u64_s16(int16x8_t __p0)38022 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
38023   uint64x2_t __ret;
38024   __ret = (uint64x2_t)(__p0);
38025   return __ret;
38026 }
vreinterpretq_u16_p8(poly8x16_t __p0)38027 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
38028   uint16x8_t __ret;
38029   __ret = (uint16x8_t)(__p0);
38030   return __ret;
38031 }
vreinterpretq_u16_p16(poly16x8_t __p0)38032 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
38033   uint16x8_t __ret;
38034   __ret = (uint16x8_t)(__p0);
38035   return __ret;
38036 }
vreinterpretq_u16_u8(uint8x16_t __p0)38037 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
38038   uint16x8_t __ret;
38039   __ret = (uint16x8_t)(__p0);
38040   return __ret;
38041 }
vreinterpretq_u16_u32(uint32x4_t __p0)38042 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
38043   uint16x8_t __ret;
38044   __ret = (uint16x8_t)(__p0);
38045   return __ret;
38046 }
vreinterpretq_u16_u64(uint64x2_t __p0)38047 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
38048   uint16x8_t __ret;
38049   __ret = (uint16x8_t)(__p0);
38050   return __ret;
38051 }
vreinterpretq_u16_s8(int8x16_t __p0)38052 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
38053   uint16x8_t __ret;
38054   __ret = (uint16x8_t)(__p0);
38055   return __ret;
38056 }
vreinterpretq_u16_f32(float32x4_t __p0)38057 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
38058   uint16x8_t __ret;
38059   __ret = (uint16x8_t)(__p0);
38060   return __ret;
38061 }
vreinterpretq_u16_f16(float16x8_t __p0)38062 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
38063   uint16x8_t __ret;
38064   __ret = (uint16x8_t)(__p0);
38065   return __ret;
38066 }
vreinterpretq_u16_s32(int32x4_t __p0)38067 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
38068   uint16x8_t __ret;
38069   __ret = (uint16x8_t)(__p0);
38070   return __ret;
38071 }
vreinterpretq_u16_s64(int64x2_t __p0)38072 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
38073   uint16x8_t __ret;
38074   __ret = (uint16x8_t)(__p0);
38075   return __ret;
38076 }
vreinterpretq_u16_s16(int16x8_t __p0)38077 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
38078   uint16x8_t __ret;
38079   __ret = (uint16x8_t)(__p0);
38080   return __ret;
38081 }
vreinterpretq_s8_p8(poly8x16_t __p0)38082 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
38083   int8x16_t __ret;
38084   __ret = (int8x16_t)(__p0);
38085   return __ret;
38086 }
vreinterpretq_s8_p16(poly16x8_t __p0)38087 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
38088   int8x16_t __ret;
38089   __ret = (int8x16_t)(__p0);
38090   return __ret;
38091 }
vreinterpretq_s8_u8(uint8x16_t __p0)38092 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
38093   int8x16_t __ret;
38094   __ret = (int8x16_t)(__p0);
38095   return __ret;
38096 }
vreinterpretq_s8_u32(uint32x4_t __p0)38097 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
38098   int8x16_t __ret;
38099   __ret = (int8x16_t)(__p0);
38100   return __ret;
38101 }
vreinterpretq_s8_u64(uint64x2_t __p0)38102 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
38103   int8x16_t __ret;
38104   __ret = (int8x16_t)(__p0);
38105   return __ret;
38106 }
vreinterpretq_s8_u16(uint16x8_t __p0)38107 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
38108   int8x16_t __ret;
38109   __ret = (int8x16_t)(__p0);
38110   return __ret;
38111 }
vreinterpretq_s8_f32(float32x4_t __p0)38112 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
38113   int8x16_t __ret;
38114   __ret = (int8x16_t)(__p0);
38115   return __ret;
38116 }
vreinterpretq_s8_f16(float16x8_t __p0)38117 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
38118   int8x16_t __ret;
38119   __ret = (int8x16_t)(__p0);
38120   return __ret;
38121 }
vreinterpretq_s8_s32(int32x4_t __p0)38122 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
38123   int8x16_t __ret;
38124   __ret = (int8x16_t)(__p0);
38125   return __ret;
38126 }
vreinterpretq_s8_s64(int64x2_t __p0)38127 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
38128   int8x16_t __ret;
38129   __ret = (int8x16_t)(__p0);
38130   return __ret;
38131 }
vreinterpretq_s8_s16(int16x8_t __p0)38132 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
38133   int8x16_t __ret;
38134   __ret = (int8x16_t)(__p0);
38135   return __ret;
38136 }
vreinterpretq_f32_p8(poly8x16_t __p0)38137 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
38138   float32x4_t __ret;
38139   __ret = (float32x4_t)(__p0);
38140   return __ret;
38141 }
vreinterpretq_f32_p16(poly16x8_t __p0)38142 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
38143   float32x4_t __ret;
38144   __ret = (float32x4_t)(__p0);
38145   return __ret;
38146 }
vreinterpretq_f32_u8(uint8x16_t __p0)38147 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
38148   float32x4_t __ret;
38149   __ret = (float32x4_t)(__p0);
38150   return __ret;
38151 }
vreinterpretq_f32_u32(uint32x4_t __p0)38152 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
38153   float32x4_t __ret;
38154   __ret = (float32x4_t)(__p0);
38155   return __ret;
38156 }
vreinterpretq_f32_u64(uint64x2_t __p0)38157 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
38158   float32x4_t __ret;
38159   __ret = (float32x4_t)(__p0);
38160   return __ret;
38161 }
vreinterpretq_f32_u16(uint16x8_t __p0)38162 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
38163   float32x4_t __ret;
38164   __ret = (float32x4_t)(__p0);
38165   return __ret;
38166 }
vreinterpretq_f32_s8(int8x16_t __p0)38167 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
38168   float32x4_t __ret;
38169   __ret = (float32x4_t)(__p0);
38170   return __ret;
38171 }
vreinterpretq_f32_f16(float16x8_t __p0)38172 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
38173   float32x4_t __ret;
38174   __ret = (float32x4_t)(__p0);
38175   return __ret;
38176 }
vreinterpretq_f32_s32(int32x4_t __p0)38177 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
38178   float32x4_t __ret;
38179   __ret = (float32x4_t)(__p0);
38180   return __ret;
38181 }
vreinterpretq_f32_s64(int64x2_t __p0)38182 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
38183   float32x4_t __ret;
38184   __ret = (float32x4_t)(__p0);
38185   return __ret;
38186 }
vreinterpretq_f32_s16(int16x8_t __p0)38187 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
38188   float32x4_t __ret;
38189   __ret = (float32x4_t)(__p0);
38190   return __ret;
38191 }
vreinterpretq_f16_p8(poly8x16_t __p0)38192 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
38193   float16x8_t __ret;
38194   __ret = (float16x8_t)(__p0);
38195   return __ret;
38196 }
vreinterpretq_f16_p16(poly16x8_t __p0)38197 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
38198   float16x8_t __ret;
38199   __ret = (float16x8_t)(__p0);
38200   return __ret;
38201 }
vreinterpretq_f16_u8(uint8x16_t __p0)38202 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
38203   float16x8_t __ret;
38204   __ret = (float16x8_t)(__p0);
38205   return __ret;
38206 }
vreinterpretq_f16_u32(uint32x4_t __p0)38207 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
38208   float16x8_t __ret;
38209   __ret = (float16x8_t)(__p0);
38210   return __ret;
38211 }
vreinterpretq_f16_u64(uint64x2_t __p0)38212 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
38213   float16x8_t __ret;
38214   __ret = (float16x8_t)(__p0);
38215   return __ret;
38216 }
vreinterpretq_f16_u16(uint16x8_t __p0)38217 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
38218   float16x8_t __ret;
38219   __ret = (float16x8_t)(__p0);
38220   return __ret;
38221 }
vreinterpretq_f16_s8(int8x16_t __p0)38222 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
38223   float16x8_t __ret;
38224   __ret = (float16x8_t)(__p0);
38225   return __ret;
38226 }
vreinterpretq_f16_f32(float32x4_t __p0)38227 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
38228   float16x8_t __ret;
38229   __ret = (float16x8_t)(__p0);
38230   return __ret;
38231 }
vreinterpretq_f16_s32(int32x4_t __p0)38232 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
38233   float16x8_t __ret;
38234   __ret = (float16x8_t)(__p0);
38235   return __ret;
38236 }
vreinterpretq_f16_s64(int64x2_t __p0)38237 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
38238   float16x8_t __ret;
38239   __ret = (float16x8_t)(__p0);
38240   return __ret;
38241 }
vreinterpretq_f16_s16(int16x8_t __p0)38242 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
38243   float16x8_t __ret;
38244   __ret = (float16x8_t)(__p0);
38245   return __ret;
38246 }
vreinterpretq_s32_p8(poly8x16_t __p0)38247 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
38248   int32x4_t __ret;
38249   __ret = (int32x4_t)(__p0);
38250   return __ret;
38251 }
vreinterpretq_s32_p16(poly16x8_t __p0)38252 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
38253   int32x4_t __ret;
38254   __ret = (int32x4_t)(__p0);
38255   return __ret;
38256 }
vreinterpretq_s32_u8(uint8x16_t __p0)38257 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
38258   int32x4_t __ret;
38259   __ret = (int32x4_t)(__p0);
38260   return __ret;
38261 }
vreinterpretq_s32_u32(uint32x4_t __p0)38262 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
38263   int32x4_t __ret;
38264   __ret = (int32x4_t)(__p0);
38265   return __ret;
38266 }
vreinterpretq_s32_u64(uint64x2_t __p0)38267 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
38268   int32x4_t __ret;
38269   __ret = (int32x4_t)(__p0);
38270   return __ret;
38271 }
vreinterpretq_s32_u16(uint16x8_t __p0)38272 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
38273   int32x4_t __ret;
38274   __ret = (int32x4_t)(__p0);
38275   return __ret;
38276 }
vreinterpretq_s32_s8(int8x16_t __p0)38277 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
38278   int32x4_t __ret;
38279   __ret = (int32x4_t)(__p0);
38280   return __ret;
38281 }
vreinterpretq_s32_f32(float32x4_t __p0)38282 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
38283   int32x4_t __ret;
38284   __ret = (int32x4_t)(__p0);
38285   return __ret;
38286 }
vreinterpretq_s32_f16(float16x8_t __p0)38287 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
38288   int32x4_t __ret;
38289   __ret = (int32x4_t)(__p0);
38290   return __ret;
38291 }
vreinterpretq_s32_s64(int64x2_t __p0)38292 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
38293   int32x4_t __ret;
38294   __ret = (int32x4_t)(__p0);
38295   return __ret;
38296 }
vreinterpretq_s32_s16(int16x8_t __p0)38297 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
38298   int32x4_t __ret;
38299   __ret = (int32x4_t)(__p0);
38300   return __ret;
38301 }
vreinterpretq_s64_p8(poly8x16_t __p0)38302 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
38303   int64x2_t __ret;
38304   __ret = (int64x2_t)(__p0);
38305   return __ret;
38306 }
vreinterpretq_s64_p16(poly16x8_t __p0)38307 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
38308   int64x2_t __ret;
38309   __ret = (int64x2_t)(__p0);
38310   return __ret;
38311 }
vreinterpretq_s64_u8(uint8x16_t __p0)38312 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
38313   int64x2_t __ret;
38314   __ret = (int64x2_t)(__p0);
38315   return __ret;
38316 }
vreinterpretq_s64_u32(uint32x4_t __p0)38317 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
38318   int64x2_t __ret;
38319   __ret = (int64x2_t)(__p0);
38320   return __ret;
38321 }
vreinterpretq_s64_u64(uint64x2_t __p0)38322 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
38323   int64x2_t __ret;
38324   __ret = (int64x2_t)(__p0);
38325   return __ret;
38326 }
vreinterpretq_s64_u16(uint16x8_t __p0)38327 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
38328   int64x2_t __ret;
38329   __ret = (int64x2_t)(__p0);
38330   return __ret;
38331 }
vreinterpretq_s64_s8(int8x16_t __p0)38332 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
38333   int64x2_t __ret;
38334   __ret = (int64x2_t)(__p0);
38335   return __ret;
38336 }
vreinterpretq_s64_f32(float32x4_t __p0)38337 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
38338   int64x2_t __ret;
38339   __ret = (int64x2_t)(__p0);
38340   return __ret;
38341 }
vreinterpretq_s64_f16(float16x8_t __p0)38342 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
38343   int64x2_t __ret;
38344   __ret = (int64x2_t)(__p0);
38345   return __ret;
38346 }
vreinterpretq_s64_s32(int32x4_t __p0)38347 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
38348   int64x2_t __ret;
38349   __ret = (int64x2_t)(__p0);
38350   return __ret;
38351 }
vreinterpretq_s64_s16(int16x8_t __p0)38352 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
38353   int64x2_t __ret;
38354   __ret = (int64x2_t)(__p0);
38355   return __ret;
38356 }
vreinterpretq_s16_p8(poly8x16_t __p0)38357 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
38358   int16x8_t __ret;
38359   __ret = (int16x8_t)(__p0);
38360   return __ret;
38361 }
vreinterpretq_s16_p16(poly16x8_t __p0)38362 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
38363   int16x8_t __ret;
38364   __ret = (int16x8_t)(__p0);
38365   return __ret;
38366 }
vreinterpretq_s16_u8(uint8x16_t __p0)38367 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
38368   int16x8_t __ret;
38369   __ret = (int16x8_t)(__p0);
38370   return __ret;
38371 }
vreinterpretq_s16_u32(uint32x4_t __p0)38372 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
38373   int16x8_t __ret;
38374   __ret = (int16x8_t)(__p0);
38375   return __ret;
38376 }
vreinterpretq_s16_u64(uint64x2_t __p0)38377 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
38378   int16x8_t __ret;
38379   __ret = (int16x8_t)(__p0);
38380   return __ret;
38381 }
vreinterpretq_s16_u16(uint16x8_t __p0)38382 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
38383   int16x8_t __ret;
38384   __ret = (int16x8_t)(__p0);
38385   return __ret;
38386 }
vreinterpretq_s16_s8(int8x16_t __p0)38387 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
38388   int16x8_t __ret;
38389   __ret = (int16x8_t)(__p0);
38390   return __ret;
38391 }
vreinterpretq_s16_f32(float32x4_t __p0)38392 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
38393   int16x8_t __ret;
38394   __ret = (int16x8_t)(__p0);
38395   return __ret;
38396 }
vreinterpretq_s16_f16(float16x8_t __p0)38397 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
38398   int16x8_t __ret;
38399   __ret = (int16x8_t)(__p0);
38400   return __ret;
38401 }
vreinterpretq_s16_s32(int32x4_t __p0)38402 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
38403   int16x8_t __ret;
38404   __ret = (int16x8_t)(__p0);
38405   return __ret;
38406 }
vreinterpretq_s16_s64(int64x2_t __p0)38407 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
38408   int16x8_t __ret;
38409   __ret = (int16x8_t)(__p0);
38410   return __ret;
38411 }
vreinterpret_u8_p8(poly8x8_t __p0)38412 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
38413   uint8x8_t __ret;
38414   __ret = (uint8x8_t)(__p0);
38415   return __ret;
38416 }
vreinterpret_u8_p16(poly16x4_t __p0)38417 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
38418   uint8x8_t __ret;
38419   __ret = (uint8x8_t)(__p0);
38420   return __ret;
38421 }
vreinterpret_u8_u32(uint32x2_t __p0)38422 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
38423   uint8x8_t __ret;
38424   __ret = (uint8x8_t)(__p0);
38425   return __ret;
38426 }
vreinterpret_u8_u64(uint64x1_t __p0)38427 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
38428   uint8x8_t __ret;
38429   __ret = (uint8x8_t)(__p0);
38430   return __ret;
38431 }
vreinterpret_u8_u16(uint16x4_t __p0)38432 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
38433   uint8x8_t __ret;
38434   __ret = (uint8x8_t)(__p0);
38435   return __ret;
38436 }
vreinterpret_u8_s8(int8x8_t __p0)38437 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
38438   uint8x8_t __ret;
38439   __ret = (uint8x8_t)(__p0);
38440   return __ret;
38441 }
vreinterpret_u8_f32(float32x2_t __p0)38442 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
38443   uint8x8_t __ret;
38444   __ret = (uint8x8_t)(__p0);
38445   return __ret;
38446 }
vreinterpret_u8_f16(float16x4_t __p0)38447 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
38448   uint8x8_t __ret;
38449   __ret = (uint8x8_t)(__p0);
38450   return __ret;
38451 }
vreinterpret_u8_s32(int32x2_t __p0)38452 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
38453   uint8x8_t __ret;
38454   __ret = (uint8x8_t)(__p0);
38455   return __ret;
38456 }
vreinterpret_u8_s64(int64x1_t __p0)38457 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
38458   uint8x8_t __ret;
38459   __ret = (uint8x8_t)(__p0);
38460   return __ret;
38461 }
vreinterpret_u8_s16(int16x4_t __p0)38462 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
38463   uint8x8_t __ret;
38464   __ret = (uint8x8_t)(__p0);
38465   return __ret;
38466 }
vreinterpret_u32_p8(poly8x8_t __p0)38467 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
38468   uint32x2_t __ret;
38469   __ret = (uint32x2_t)(__p0);
38470   return __ret;
38471 }
vreinterpret_u32_p16(poly16x4_t __p0)38472 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
38473   uint32x2_t __ret;
38474   __ret = (uint32x2_t)(__p0);
38475   return __ret;
38476 }
vreinterpret_u32_u8(uint8x8_t __p0)38477 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
38478   uint32x2_t __ret;
38479   __ret = (uint32x2_t)(__p0);
38480   return __ret;
38481 }
vreinterpret_u32_u64(uint64x1_t __p0)38482 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
38483   uint32x2_t __ret;
38484   __ret = (uint32x2_t)(__p0);
38485   return __ret;
38486 }
vreinterpret_u32_u16(uint16x4_t __p0)38487 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
38488   uint32x2_t __ret;
38489   __ret = (uint32x2_t)(__p0);
38490   return __ret;
38491 }
vreinterpret_u32_s8(int8x8_t __p0)38492 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
38493   uint32x2_t __ret;
38494   __ret = (uint32x2_t)(__p0);
38495   return __ret;
38496 }
vreinterpret_u32_f32(float32x2_t __p0)38497 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
38498   uint32x2_t __ret;
38499   __ret = (uint32x2_t)(__p0);
38500   return __ret;
38501 }
vreinterpret_u32_f16(float16x4_t __p0)38502 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
38503   uint32x2_t __ret;
38504   __ret = (uint32x2_t)(__p0);
38505   return __ret;
38506 }
vreinterpret_u32_s32(int32x2_t __p0)38507 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
38508   uint32x2_t __ret;
38509   __ret = (uint32x2_t)(__p0);
38510   return __ret;
38511 }
vreinterpret_u32_s64(int64x1_t __p0)38512 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
38513   uint32x2_t __ret;
38514   __ret = (uint32x2_t)(__p0);
38515   return __ret;
38516 }
vreinterpret_u32_s16(int16x4_t __p0)38517 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
38518   uint32x2_t __ret;
38519   __ret = (uint32x2_t)(__p0);
38520   return __ret;
38521 }
vreinterpret_u64_p8(poly8x8_t __p0)38522 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
38523   uint64x1_t __ret;
38524   __ret = (uint64x1_t)(__p0);
38525   return __ret;
38526 }
vreinterpret_u64_p16(poly16x4_t __p0)38527 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
38528   uint64x1_t __ret;
38529   __ret = (uint64x1_t)(__p0);
38530   return __ret;
38531 }
vreinterpret_u64_u8(uint8x8_t __p0)38532 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
38533   uint64x1_t __ret;
38534   __ret = (uint64x1_t)(__p0);
38535   return __ret;
38536 }
vreinterpret_u64_u32(uint32x2_t __p0)38537 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
38538   uint64x1_t __ret;
38539   __ret = (uint64x1_t)(__p0);
38540   return __ret;
38541 }
vreinterpret_u64_u16(uint16x4_t __p0)38542 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
38543   uint64x1_t __ret;
38544   __ret = (uint64x1_t)(__p0);
38545   return __ret;
38546 }
vreinterpret_u64_s8(int8x8_t __p0)38547 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
38548   uint64x1_t __ret;
38549   __ret = (uint64x1_t)(__p0);
38550   return __ret;
38551 }
vreinterpret_u64_f32(float32x2_t __p0)38552 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
38553   uint64x1_t __ret;
38554   __ret = (uint64x1_t)(__p0);
38555   return __ret;
38556 }
vreinterpret_u64_f16(float16x4_t __p0)38557 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
38558   uint64x1_t __ret;
38559   __ret = (uint64x1_t)(__p0);
38560   return __ret;
38561 }
vreinterpret_u64_s32(int32x2_t __p0)38562 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
38563   uint64x1_t __ret;
38564   __ret = (uint64x1_t)(__p0);
38565   return __ret;
38566 }
vreinterpret_u64_s64(int64x1_t __p0)38567 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
38568   uint64x1_t __ret;
38569   __ret = (uint64x1_t)(__p0);
38570   return __ret;
38571 }
vreinterpret_u64_s16(int16x4_t __p0)38572 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
38573   uint64x1_t __ret;
38574   __ret = (uint64x1_t)(__p0);
38575   return __ret;
38576 }
vreinterpret_u16_p8(poly8x8_t __p0)38577 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
38578   uint16x4_t __ret;
38579   __ret = (uint16x4_t)(__p0);
38580   return __ret;
38581 }
vreinterpret_u16_p16(poly16x4_t __p0)38582 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
38583   uint16x4_t __ret;
38584   __ret = (uint16x4_t)(__p0);
38585   return __ret;
38586 }
vreinterpret_u16_u8(uint8x8_t __p0)38587 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
38588   uint16x4_t __ret;
38589   __ret = (uint16x4_t)(__p0);
38590   return __ret;
38591 }
vreinterpret_u16_u32(uint32x2_t __p0)38592 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
38593   uint16x4_t __ret;
38594   __ret = (uint16x4_t)(__p0);
38595   return __ret;
38596 }
vreinterpret_u16_u64(uint64x1_t __p0)38597 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
38598   uint16x4_t __ret;
38599   __ret = (uint16x4_t)(__p0);
38600   return __ret;
38601 }
vreinterpret_u16_s8(int8x8_t __p0)38602 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
38603   uint16x4_t __ret;
38604   __ret = (uint16x4_t)(__p0);
38605   return __ret;
38606 }
vreinterpret_u16_f32(float32x2_t __p0)38607 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
38608   uint16x4_t __ret;
38609   __ret = (uint16x4_t)(__p0);
38610   return __ret;
38611 }
vreinterpret_u16_f16(float16x4_t __p0)38612 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
38613   uint16x4_t __ret;
38614   __ret = (uint16x4_t)(__p0);
38615   return __ret;
38616 }
vreinterpret_u16_s32(int32x2_t __p0)38617 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
38618   uint16x4_t __ret;
38619   __ret = (uint16x4_t)(__p0);
38620   return __ret;
38621 }
vreinterpret_u16_s64(int64x1_t __p0)38622 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
38623   uint16x4_t __ret;
38624   __ret = (uint16x4_t)(__p0);
38625   return __ret;
38626 }
vreinterpret_u16_s16(int16x4_t __p0)38627 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
38628   uint16x4_t __ret;
38629   __ret = (uint16x4_t)(__p0);
38630   return __ret;
38631 }
vreinterpret_s8_p8(poly8x8_t __p0)38632 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
38633   int8x8_t __ret;
38634   __ret = (int8x8_t)(__p0);
38635   return __ret;
38636 }
vreinterpret_s8_p16(poly16x4_t __p0)38637 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
38638   int8x8_t __ret;
38639   __ret = (int8x8_t)(__p0);
38640   return __ret;
38641 }
vreinterpret_s8_u8(uint8x8_t __p0)38642 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
38643   int8x8_t __ret;
38644   __ret = (int8x8_t)(__p0);
38645   return __ret;
38646 }
vreinterpret_s8_u32(uint32x2_t __p0)38647 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
38648   int8x8_t __ret;
38649   __ret = (int8x8_t)(__p0);
38650   return __ret;
38651 }
vreinterpret_s8_u64(uint64x1_t __p0)38652 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
38653   int8x8_t __ret;
38654   __ret = (int8x8_t)(__p0);
38655   return __ret;
38656 }
vreinterpret_s8_u16(uint16x4_t __p0)38657 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
38658   int8x8_t __ret;
38659   __ret = (int8x8_t)(__p0);
38660   return __ret;
38661 }
vreinterpret_s8_f32(float32x2_t __p0)38662 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
38663   int8x8_t __ret;
38664   __ret = (int8x8_t)(__p0);
38665   return __ret;
38666 }
vreinterpret_s8_f16(float16x4_t __p0)38667 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
38668   int8x8_t __ret;
38669   __ret = (int8x8_t)(__p0);
38670   return __ret;
38671 }
vreinterpret_s8_s32(int32x2_t __p0)38672 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
38673   int8x8_t __ret;
38674   __ret = (int8x8_t)(__p0);
38675   return __ret;
38676 }
vreinterpret_s8_s64(int64x1_t __p0)38677 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
38678   int8x8_t __ret;
38679   __ret = (int8x8_t)(__p0);
38680   return __ret;
38681 }
vreinterpret_s8_s16(int16x4_t __p0)38682 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
38683   int8x8_t __ret;
38684   __ret = (int8x8_t)(__p0);
38685   return __ret;
38686 }
vreinterpret_f32_p8(poly8x8_t __p0)38687 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
38688   float32x2_t __ret;
38689   __ret = (float32x2_t)(__p0);
38690   return __ret;
38691 }
vreinterpret_f32_p16(poly16x4_t __p0)38692 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
38693   float32x2_t __ret;
38694   __ret = (float32x2_t)(__p0);
38695   return __ret;
38696 }
vreinterpret_f32_u8(uint8x8_t __p0)38697 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
38698   float32x2_t __ret;
38699   __ret = (float32x2_t)(__p0);
38700   return __ret;
38701 }
vreinterpret_f32_u32(uint32x2_t __p0)38702 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
38703   float32x2_t __ret;
38704   __ret = (float32x2_t)(__p0);
38705   return __ret;
38706 }
vreinterpret_f32_u64(uint64x1_t __p0)38707 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
38708   float32x2_t __ret;
38709   __ret = (float32x2_t)(__p0);
38710   return __ret;
38711 }
vreinterpret_f32_u16(uint16x4_t __p0)38712 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
38713   float32x2_t __ret;
38714   __ret = (float32x2_t)(__p0);
38715   return __ret;
38716 }
vreinterpret_f32_s8(int8x8_t __p0)38717 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
38718   float32x2_t __ret;
38719   __ret = (float32x2_t)(__p0);
38720   return __ret;
38721 }
vreinterpret_f32_f16(float16x4_t __p0)38722 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
38723   float32x2_t __ret;
38724   __ret = (float32x2_t)(__p0);
38725   return __ret;
38726 }
vreinterpret_f32_s32(int32x2_t __p0)38727 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
38728   float32x2_t __ret;
38729   __ret = (float32x2_t)(__p0);
38730   return __ret;
38731 }
vreinterpret_f32_s64(int64x1_t __p0)38732 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
38733   float32x2_t __ret;
38734   __ret = (float32x2_t)(__p0);
38735   return __ret;
38736 }
vreinterpret_f32_s16(int16x4_t __p0)38737 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
38738   float32x2_t __ret;
38739   __ret = (float32x2_t)(__p0);
38740   return __ret;
38741 }
vreinterpret_f16_p8(poly8x8_t __p0)38742 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
38743   float16x4_t __ret;
38744   __ret = (float16x4_t)(__p0);
38745   return __ret;
38746 }
vreinterpret_f16_p16(poly16x4_t __p0)38747 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
38748   float16x4_t __ret;
38749   __ret = (float16x4_t)(__p0);
38750   return __ret;
38751 }
vreinterpret_f16_u8(uint8x8_t __p0)38752 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
38753   float16x4_t __ret;
38754   __ret = (float16x4_t)(__p0);
38755   return __ret;
38756 }
vreinterpret_f16_u32(uint32x2_t __p0)38757 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
38758   float16x4_t __ret;
38759   __ret = (float16x4_t)(__p0);
38760   return __ret;
38761 }
vreinterpret_f16_u64(uint64x1_t __p0)38762 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
38763   float16x4_t __ret;
38764   __ret = (float16x4_t)(__p0);
38765   return __ret;
38766 }
vreinterpret_f16_u16(uint16x4_t __p0)38767 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
38768   float16x4_t __ret;
38769   __ret = (float16x4_t)(__p0);
38770   return __ret;
38771 }
vreinterpret_f16_s8(int8x8_t __p0)38772 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
38773   float16x4_t __ret;
38774   __ret = (float16x4_t)(__p0);
38775   return __ret;
38776 }
vreinterpret_f16_f32(float32x2_t __p0)38777 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
38778   float16x4_t __ret;
38779   __ret = (float16x4_t)(__p0);
38780   return __ret;
38781 }
vreinterpret_f16_s32(int32x2_t __p0)38782 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
38783   float16x4_t __ret;
38784   __ret = (float16x4_t)(__p0);
38785   return __ret;
38786 }
vreinterpret_f16_s64(int64x1_t __p0)38787 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
38788   float16x4_t __ret;
38789   __ret = (float16x4_t)(__p0);
38790   return __ret;
38791 }
vreinterpret_f16_s16(int16x4_t __p0)38792 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
38793   float16x4_t __ret;
38794   __ret = (float16x4_t)(__p0);
38795   return __ret;
38796 }
vreinterpret_s32_p8(poly8x8_t __p0)38797 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
38798   int32x2_t __ret;
38799   __ret = (int32x2_t)(__p0);
38800   return __ret;
38801 }
vreinterpret_s32_p16(poly16x4_t __p0)38802 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
38803   int32x2_t __ret;
38804   __ret = (int32x2_t)(__p0);
38805   return __ret;
38806 }
vreinterpret_s32_u8(uint8x8_t __p0)38807 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
38808   int32x2_t __ret;
38809   __ret = (int32x2_t)(__p0);
38810   return __ret;
38811 }
vreinterpret_s32_u32(uint32x2_t __p0)38812 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
38813   int32x2_t __ret;
38814   __ret = (int32x2_t)(__p0);
38815   return __ret;
38816 }
vreinterpret_s32_u64(uint64x1_t __p0)38817 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
38818   int32x2_t __ret;
38819   __ret = (int32x2_t)(__p0);
38820   return __ret;
38821 }
vreinterpret_s32_u16(uint16x4_t __p0)38822 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
38823   int32x2_t __ret;
38824   __ret = (int32x2_t)(__p0);
38825   return __ret;
38826 }
vreinterpret_s32_s8(int8x8_t __p0)38827 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
38828   int32x2_t __ret;
38829   __ret = (int32x2_t)(__p0);
38830   return __ret;
38831 }
vreinterpret_s32_f32(float32x2_t __p0)38832 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
38833   int32x2_t __ret;
38834   __ret = (int32x2_t)(__p0);
38835   return __ret;
38836 }
vreinterpret_s32_f16(float16x4_t __p0)38837 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
38838   int32x2_t __ret;
38839   __ret = (int32x2_t)(__p0);
38840   return __ret;
38841 }
vreinterpret_s32_s64(int64x1_t __p0)38842 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
38843   int32x2_t __ret;
38844   __ret = (int32x2_t)(__p0);
38845   return __ret;
38846 }
vreinterpret_s32_s16(int16x4_t __p0)38847 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
38848   int32x2_t __ret;
38849   __ret = (int32x2_t)(__p0);
38850   return __ret;
38851 }
vreinterpret_s64_p8(poly8x8_t __p0)38852 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
38853   int64x1_t __ret;
38854   __ret = (int64x1_t)(__p0);
38855   return __ret;
38856 }
vreinterpret_s64_p16(poly16x4_t __p0)38857 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
38858   int64x1_t __ret;
38859   __ret = (int64x1_t)(__p0);
38860   return __ret;
38861 }
vreinterpret_s64_u8(uint8x8_t __p0)38862 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
38863   int64x1_t __ret;
38864   __ret = (int64x1_t)(__p0);
38865   return __ret;
38866 }
vreinterpret_s64_u32(uint32x2_t __p0)38867 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
38868   int64x1_t __ret;
38869   __ret = (int64x1_t)(__p0);
38870   return __ret;
38871 }
vreinterpret_s64_u64(uint64x1_t __p0)38872 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
38873   int64x1_t __ret;
38874   __ret = (int64x1_t)(__p0);
38875   return __ret;
38876 }
vreinterpret_s64_u16(uint16x4_t __p0)38877 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
38878   int64x1_t __ret;
38879   __ret = (int64x1_t)(__p0);
38880   return __ret;
38881 }
vreinterpret_s64_s8(int8x8_t __p0)38882 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
38883   int64x1_t __ret;
38884   __ret = (int64x1_t)(__p0);
38885   return __ret;
38886 }
vreinterpret_s64_f32(float32x2_t __p0)38887 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
38888   int64x1_t __ret;
38889   __ret = (int64x1_t)(__p0);
38890   return __ret;
38891 }
vreinterpret_s64_f16(float16x4_t __p0)38892 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
38893   int64x1_t __ret;
38894   __ret = (int64x1_t)(__p0);
38895   return __ret;
38896 }
vreinterpret_s64_s32(int32x2_t __p0)38897 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
38898   int64x1_t __ret;
38899   __ret = (int64x1_t)(__p0);
38900   return __ret;
38901 }
vreinterpret_s64_s16(int16x4_t __p0)38902 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
38903   int64x1_t __ret;
38904   __ret = (int64x1_t)(__p0);
38905   return __ret;
38906 }
vreinterpret_s16_p8(poly8x8_t __p0)38907 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
38908   int16x4_t __ret;
38909   __ret = (int16x4_t)(__p0);
38910   return __ret;
38911 }
vreinterpret_s16_p16(poly16x4_t __p0)38912 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
38913   int16x4_t __ret;
38914   __ret = (int16x4_t)(__p0);
38915   return __ret;
38916 }
vreinterpret_s16_u8(uint8x8_t __p0)38917 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
38918   int16x4_t __ret;
38919   __ret = (int16x4_t)(__p0);
38920   return __ret;
38921 }
vreinterpret_s16_u32(uint32x2_t __p0)38922 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
38923   int16x4_t __ret;
38924   __ret = (int16x4_t)(__p0);
38925   return __ret;
38926 }
vreinterpret_s16_u64(uint64x1_t __p0)38927 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
38928   int16x4_t __ret;
38929   __ret = (int16x4_t)(__p0);
38930   return __ret;
38931 }
vreinterpret_s16_u16(uint16x4_t __p0)38932 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
38933   int16x4_t __ret;
38934   __ret = (int16x4_t)(__p0);
38935   return __ret;
38936 }
vreinterpret_s16_s8(int8x8_t __p0)38937 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
38938   int16x4_t __ret;
38939   __ret = (int16x4_t)(__p0);
38940   return __ret;
38941 }
vreinterpret_s16_f32(float32x2_t __p0)38942 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
38943   int16x4_t __ret;
38944   __ret = (int16x4_t)(__p0);
38945   return __ret;
38946 }
vreinterpret_s16_f16(float16x4_t __p0)38947 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
38948   int16x4_t __ret;
38949   __ret = (int16x4_t)(__p0);
38950   return __ret;
38951 }
vreinterpret_s16_s32(int32x2_t __p0)38952 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
38953   int16x4_t __ret;
38954   __ret = (int16x4_t)(__p0);
38955   return __ret;
38956 }
vreinterpret_s16_s64(int64x1_t __p0)38957 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
38958   int16x4_t __ret;
38959   __ret = (int16x4_t)(__p0);
38960   return __ret;
38961 }
38962 #ifdef __LITTLE_ENDIAN__
__a32_vcvt_bf16_f32(float32x4_t __p0)38963 __ai __attribute__((target("bf16"))) bfloat16x4_t __a32_vcvt_bf16_f32(float32x4_t __p0) {
38964   bfloat16x4_t __ret;
38965   __ret = (bfloat16x4_t) __builtin_neon___a32_vcvt_bf16_f32((int8x16_t)__p0, 11);
38966   return __ret;
38967 }
38968 #else
__a32_vcvt_bf16_f32(float32x4_t __p0)38969 __ai __attribute__((target("bf16"))) bfloat16x4_t __a32_vcvt_bf16_f32(float32x4_t __p0) {
38970   bfloat16x4_t __ret;
38971   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
38972   __ret = (bfloat16x4_t) __builtin_neon___a32_vcvt_bf16_f32((int8x16_t)__rev0, 11);
38973   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
38974   return __ret;
38975 }
__noswap___a32_vcvt_bf16_f32(float32x4_t __p0)38976 __ai __attribute__((target("bf16"))) bfloat16x4_t __noswap___a32_vcvt_bf16_f32(float32x4_t __p0) {
38977   bfloat16x4_t __ret;
38978   __ret = (bfloat16x4_t) __builtin_neon___a32_vcvt_bf16_f32((int8x16_t)__p0, 11);
38979   return __ret;
38980 }
38981 #endif
38982 
38983 #ifdef __LITTLE_ENDIAN__
vcvt_bf16_f32(float32x4_t __p0)38984 __ai __attribute__((target("bf16"))) bfloat16x4_t vcvt_bf16_f32(float32x4_t __p0) {
38985   bfloat16x4_t __ret;
38986   __ret = __a32_vcvt_bf16_f32(__p0);
38987   return __ret;
38988 }
38989 #else
vcvt_bf16_f32(float32x4_t __p0)38990 __ai __attribute__((target("bf16"))) bfloat16x4_t vcvt_bf16_f32(float32x4_t __p0) {
38991   bfloat16x4_t __ret;
38992   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
38993   __ret = __noswap___a32_vcvt_bf16_f32(__rev0);
38994   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
38995   return __ret;
38996 }
38997 #endif
38998 
38999 #ifdef __LITTLE_ENDIAN__
vcvtq_high_bf16_f32(bfloat16x8_t __p0,float32x4_t __p1)39000 __ai __attribute__((target("bf16"))) bfloat16x8_t vcvtq_high_bf16_f32(bfloat16x8_t __p0, float32x4_t __p1) {
39001   bfloat16x8_t __ret;
39002   __ret = vcombine_bf16(__a32_vcvt_bf16_f32(__p1), vget_low_bf16(__p0));
39003   return __ret;
39004 }
39005 #else
vcvtq_high_bf16_f32(bfloat16x8_t __p0,float32x4_t __p1)39006 __ai __attribute__((target("bf16"))) bfloat16x8_t vcvtq_high_bf16_f32(bfloat16x8_t __p0, float32x4_t __p1) {
39007   bfloat16x8_t __ret;
39008   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
39009   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
39010   __ret = __noswap_vcombine_bf16(__noswap___a32_vcvt_bf16_f32(__rev1), __noswap_vget_low_bf16(__rev0));
39011   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
39012   return __ret;
39013 }
39014 #endif
39015 
39016 #ifdef __LITTLE_ENDIAN__
vcvtq_low_bf16_f32(float32x4_t __p0)39017 __ai __attribute__((target("bf16"))) bfloat16x8_t vcvtq_low_bf16_f32(float32x4_t __p0) {
39018   bfloat16x8_t __ret;
39019   __ret = vcombine_bf16((bfloat16x4_t)(0ULL), __a32_vcvt_bf16_f32(__p0));
39020   return __ret;
39021 }
39022 #else
vcvtq_low_bf16_f32(float32x4_t __p0)39023 __ai __attribute__((target("bf16"))) bfloat16x8_t vcvtq_low_bf16_f32(float32x4_t __p0) {
39024   bfloat16x8_t __ret;
39025   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39026   __ret = __noswap_vcombine_bf16((bfloat16x4_t)(0ULL), __noswap___a32_vcvt_bf16_f32(__rev0));
39027   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
39028   return __ret;
39029 }
39030 #endif
39031 
vreinterpret_p8_bf16(bfloat16x4_t __p0)39032 __ai __attribute__((target("bf16"))) poly8x8_t vreinterpret_p8_bf16(bfloat16x4_t __p0) {
39033   poly8x8_t __ret;
39034   __ret = (poly8x8_t)(__p0);
39035   return __ret;
39036 }
vreinterpret_p64_bf16(bfloat16x4_t __p0)39037 __ai __attribute__((target("bf16"))) poly64x1_t vreinterpret_p64_bf16(bfloat16x4_t __p0) {
39038   poly64x1_t __ret;
39039   __ret = (poly64x1_t)(__p0);
39040   return __ret;
39041 }
vreinterpret_p16_bf16(bfloat16x4_t __p0)39042 __ai __attribute__((target("bf16"))) poly16x4_t vreinterpret_p16_bf16(bfloat16x4_t __p0) {
39043   poly16x4_t __ret;
39044   __ret = (poly16x4_t)(__p0);
39045   return __ret;
39046 }
vreinterpretq_p8_bf16(bfloat16x8_t __p0)39047 __ai __attribute__((target("bf16"))) poly8x16_t vreinterpretq_p8_bf16(bfloat16x8_t __p0) {
39048   poly8x16_t __ret;
39049   __ret = (poly8x16_t)(__p0);
39050   return __ret;
39051 }
vreinterpretq_p64_bf16(bfloat16x8_t __p0)39052 __ai __attribute__((target("bf16"))) poly64x2_t vreinterpretq_p64_bf16(bfloat16x8_t __p0) {
39053   poly64x2_t __ret;
39054   __ret = (poly64x2_t)(__p0);
39055   return __ret;
39056 }
vreinterpretq_p16_bf16(bfloat16x8_t __p0)39057 __ai __attribute__((target("bf16"))) poly16x8_t vreinterpretq_p16_bf16(bfloat16x8_t __p0) {
39058   poly16x8_t __ret;
39059   __ret = (poly16x8_t)(__p0);
39060   return __ret;
39061 }
vreinterpretq_u8_bf16(bfloat16x8_t __p0)39062 __ai __attribute__((target("bf16"))) uint8x16_t vreinterpretq_u8_bf16(bfloat16x8_t __p0) {
39063   uint8x16_t __ret;
39064   __ret = (uint8x16_t)(__p0);
39065   return __ret;
39066 }
vreinterpretq_u32_bf16(bfloat16x8_t __p0)39067 __ai __attribute__((target("bf16"))) uint32x4_t vreinterpretq_u32_bf16(bfloat16x8_t __p0) {
39068   uint32x4_t __ret;
39069   __ret = (uint32x4_t)(__p0);
39070   return __ret;
39071 }
vreinterpretq_u64_bf16(bfloat16x8_t __p0)39072 __ai __attribute__((target("bf16"))) uint64x2_t vreinterpretq_u64_bf16(bfloat16x8_t __p0) {
39073   uint64x2_t __ret;
39074   __ret = (uint64x2_t)(__p0);
39075   return __ret;
39076 }
vreinterpretq_u16_bf16(bfloat16x8_t __p0)39077 __ai __attribute__((target("bf16"))) uint16x8_t vreinterpretq_u16_bf16(bfloat16x8_t __p0) {
39078   uint16x8_t __ret;
39079   __ret = (uint16x8_t)(__p0);
39080   return __ret;
39081 }
vreinterpretq_s8_bf16(bfloat16x8_t __p0)39082 __ai __attribute__((target("bf16"))) int8x16_t vreinterpretq_s8_bf16(bfloat16x8_t __p0) {
39083   int8x16_t __ret;
39084   __ret = (int8x16_t)(__p0);
39085   return __ret;
39086 }
vreinterpretq_f32_bf16(bfloat16x8_t __p0)39087 __ai __attribute__((target("bf16"))) float32x4_t vreinterpretq_f32_bf16(bfloat16x8_t __p0) {
39088   float32x4_t __ret;
39089   __ret = (float32x4_t)(__p0);
39090   return __ret;
39091 }
vreinterpretq_f16_bf16(bfloat16x8_t __p0)39092 __ai __attribute__((target("bf16"))) float16x8_t vreinterpretq_f16_bf16(bfloat16x8_t __p0) {
39093   float16x8_t __ret;
39094   __ret = (float16x8_t)(__p0);
39095   return __ret;
39096 }
vreinterpretq_s32_bf16(bfloat16x8_t __p0)39097 __ai __attribute__((target("bf16"))) int32x4_t vreinterpretq_s32_bf16(bfloat16x8_t __p0) {
39098   int32x4_t __ret;
39099   __ret = (int32x4_t)(__p0);
39100   return __ret;
39101 }
vreinterpretq_s64_bf16(bfloat16x8_t __p0)39102 __ai __attribute__((target("bf16"))) int64x2_t vreinterpretq_s64_bf16(bfloat16x8_t __p0) {
39103   int64x2_t __ret;
39104   __ret = (int64x2_t)(__p0);
39105   return __ret;
39106 }
vreinterpretq_s16_bf16(bfloat16x8_t __p0)39107 __ai __attribute__((target("bf16"))) int16x8_t vreinterpretq_s16_bf16(bfloat16x8_t __p0) {
39108   int16x8_t __ret;
39109   __ret = (int16x8_t)(__p0);
39110   return __ret;
39111 }
vreinterpret_u8_bf16(bfloat16x4_t __p0)39112 __ai __attribute__((target("bf16"))) uint8x8_t vreinterpret_u8_bf16(bfloat16x4_t __p0) {
39113   uint8x8_t __ret;
39114   __ret = (uint8x8_t)(__p0);
39115   return __ret;
39116 }
vreinterpret_u32_bf16(bfloat16x4_t __p0)39117 __ai __attribute__((target("bf16"))) uint32x2_t vreinterpret_u32_bf16(bfloat16x4_t __p0) {
39118   uint32x2_t __ret;
39119   __ret = (uint32x2_t)(__p0);
39120   return __ret;
39121 }
vreinterpret_u64_bf16(bfloat16x4_t __p0)39122 __ai __attribute__((target("bf16"))) uint64x1_t vreinterpret_u64_bf16(bfloat16x4_t __p0) {
39123   uint64x1_t __ret;
39124   __ret = (uint64x1_t)(__p0);
39125   return __ret;
39126 }
vreinterpret_u16_bf16(bfloat16x4_t __p0)39127 __ai __attribute__((target("bf16"))) uint16x4_t vreinterpret_u16_bf16(bfloat16x4_t __p0) {
39128   uint16x4_t __ret;
39129   __ret = (uint16x4_t)(__p0);
39130   return __ret;
39131 }
vreinterpret_s8_bf16(bfloat16x4_t __p0)39132 __ai __attribute__((target("bf16"))) int8x8_t vreinterpret_s8_bf16(bfloat16x4_t __p0) {
39133   int8x8_t __ret;
39134   __ret = (int8x8_t)(__p0);
39135   return __ret;
39136 }
vreinterpret_f32_bf16(bfloat16x4_t __p0)39137 __ai __attribute__((target("bf16"))) float32x2_t vreinterpret_f32_bf16(bfloat16x4_t __p0) {
39138   float32x2_t __ret;
39139   __ret = (float32x2_t)(__p0);
39140   return __ret;
39141 }
vreinterpret_f16_bf16(bfloat16x4_t __p0)39142 __ai __attribute__((target("bf16"))) float16x4_t vreinterpret_f16_bf16(bfloat16x4_t __p0) {
39143   float16x4_t __ret;
39144   __ret = (float16x4_t)(__p0);
39145   return __ret;
39146 }
vreinterpret_s32_bf16(bfloat16x4_t __p0)39147 __ai __attribute__((target("bf16"))) int32x2_t vreinterpret_s32_bf16(bfloat16x4_t __p0) {
39148   int32x2_t __ret;
39149   __ret = (int32x2_t)(__p0);
39150   return __ret;
39151 }
vreinterpret_s64_bf16(bfloat16x4_t __p0)39152 __ai __attribute__((target("bf16"))) int64x1_t vreinterpret_s64_bf16(bfloat16x4_t __p0) {
39153   int64x1_t __ret;
39154   __ret = (int64x1_t)(__p0);
39155   return __ret;
39156 }
vreinterpret_s16_bf16(bfloat16x4_t __p0)39157 __ai __attribute__((target("bf16"))) int16x4_t vreinterpret_s16_bf16(bfloat16x4_t __p0) {
39158   int16x4_t __ret;
39159   __ret = (int16x4_t)(__p0);
39160   return __ret;
39161 }
vreinterpretq_bf16_p8(poly8x16_t __p0)39162 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_p8(poly8x16_t __p0) {
39163   bfloat16x8_t __ret;
39164   __ret = (bfloat16x8_t)(__p0);
39165   return __ret;
39166 }
vreinterpretq_bf16_p64(poly64x2_t __p0)39167 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_p64(poly64x2_t __p0) {
39168   bfloat16x8_t __ret;
39169   __ret = (bfloat16x8_t)(__p0);
39170   return __ret;
39171 }
vreinterpretq_bf16_p16(poly16x8_t __p0)39172 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_p16(poly16x8_t __p0) {
39173   bfloat16x8_t __ret;
39174   __ret = (bfloat16x8_t)(__p0);
39175   return __ret;
39176 }
vreinterpretq_bf16_u8(uint8x16_t __p0)39177 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_u8(uint8x16_t __p0) {
39178   bfloat16x8_t __ret;
39179   __ret = (bfloat16x8_t)(__p0);
39180   return __ret;
39181 }
vreinterpretq_bf16_u32(uint32x4_t __p0)39182 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_u32(uint32x4_t __p0) {
39183   bfloat16x8_t __ret;
39184   __ret = (bfloat16x8_t)(__p0);
39185   return __ret;
39186 }
vreinterpretq_bf16_u64(uint64x2_t __p0)39187 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_u64(uint64x2_t __p0) {
39188   bfloat16x8_t __ret;
39189   __ret = (bfloat16x8_t)(__p0);
39190   return __ret;
39191 }
vreinterpretq_bf16_u16(uint16x8_t __p0)39192 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_u16(uint16x8_t __p0) {
39193   bfloat16x8_t __ret;
39194   __ret = (bfloat16x8_t)(__p0);
39195   return __ret;
39196 }
vreinterpretq_bf16_s8(int8x16_t __p0)39197 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_s8(int8x16_t __p0) {
39198   bfloat16x8_t __ret;
39199   __ret = (bfloat16x8_t)(__p0);
39200   return __ret;
39201 }
vreinterpretq_bf16_f32(float32x4_t __p0)39202 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_f32(float32x4_t __p0) {
39203   bfloat16x8_t __ret;
39204   __ret = (bfloat16x8_t)(__p0);
39205   return __ret;
39206 }
vreinterpretq_bf16_f16(float16x8_t __p0)39207 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_f16(float16x8_t __p0) {
39208   bfloat16x8_t __ret;
39209   __ret = (bfloat16x8_t)(__p0);
39210   return __ret;
39211 }
vreinterpretq_bf16_s32(int32x4_t __p0)39212 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_s32(int32x4_t __p0) {
39213   bfloat16x8_t __ret;
39214   __ret = (bfloat16x8_t)(__p0);
39215   return __ret;
39216 }
vreinterpretq_bf16_s64(int64x2_t __p0)39217 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_s64(int64x2_t __p0) {
39218   bfloat16x8_t __ret;
39219   __ret = (bfloat16x8_t)(__p0);
39220   return __ret;
39221 }
vreinterpretq_bf16_s16(int16x8_t __p0)39222 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_s16(int16x8_t __p0) {
39223   bfloat16x8_t __ret;
39224   __ret = (bfloat16x8_t)(__p0);
39225   return __ret;
39226 }
vreinterpret_bf16_p8(poly8x8_t __p0)39227 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_p8(poly8x8_t __p0) {
39228   bfloat16x4_t __ret;
39229   __ret = (bfloat16x4_t)(__p0);
39230   return __ret;
39231 }
vreinterpret_bf16_p64(poly64x1_t __p0)39232 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_p64(poly64x1_t __p0) {
39233   bfloat16x4_t __ret;
39234   __ret = (bfloat16x4_t)(__p0);
39235   return __ret;
39236 }
vreinterpret_bf16_p16(poly16x4_t __p0)39237 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_p16(poly16x4_t __p0) {
39238   bfloat16x4_t __ret;
39239   __ret = (bfloat16x4_t)(__p0);
39240   return __ret;
39241 }
vreinterpret_bf16_u8(uint8x8_t __p0)39242 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_u8(uint8x8_t __p0) {
39243   bfloat16x4_t __ret;
39244   __ret = (bfloat16x4_t)(__p0);
39245   return __ret;
39246 }
vreinterpret_bf16_u32(uint32x2_t __p0)39247 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_u32(uint32x2_t __p0) {
39248   bfloat16x4_t __ret;
39249   __ret = (bfloat16x4_t)(__p0);
39250   return __ret;
39251 }
vreinterpret_bf16_u64(uint64x1_t __p0)39252 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_u64(uint64x1_t __p0) {
39253   bfloat16x4_t __ret;
39254   __ret = (bfloat16x4_t)(__p0);
39255   return __ret;
39256 }
vreinterpret_bf16_u16(uint16x4_t __p0)39257 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_u16(uint16x4_t __p0) {
39258   bfloat16x4_t __ret;
39259   __ret = (bfloat16x4_t)(__p0);
39260   return __ret;
39261 }
vreinterpret_bf16_s8(int8x8_t __p0)39262 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_s8(int8x8_t __p0) {
39263   bfloat16x4_t __ret;
39264   __ret = (bfloat16x4_t)(__p0);
39265   return __ret;
39266 }
vreinterpret_bf16_f32(float32x2_t __p0)39267 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_f32(float32x2_t __p0) {
39268   bfloat16x4_t __ret;
39269   __ret = (bfloat16x4_t)(__p0);
39270   return __ret;
39271 }
vreinterpret_bf16_f16(float16x4_t __p0)39272 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_f16(float16x4_t __p0) {
39273   bfloat16x4_t __ret;
39274   __ret = (bfloat16x4_t)(__p0);
39275   return __ret;
39276 }
vreinterpret_bf16_s32(int32x2_t __p0)39277 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_s32(int32x2_t __p0) {
39278   bfloat16x4_t __ret;
39279   __ret = (bfloat16x4_t)(__p0);
39280   return __ret;
39281 }
vreinterpret_bf16_s64(int64x1_t __p0)39282 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_s64(int64x1_t __p0) {
39283   bfloat16x4_t __ret;
39284   __ret = (bfloat16x4_t)(__p0);
39285   return __ret;
39286 }
vreinterpret_bf16_s16(int16x4_t __p0)39287 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_s16(int16x4_t __p0) {
39288   bfloat16x4_t __ret;
39289   __ret = (bfloat16x4_t)(__p0);
39290   return __ret;
39291 }
39292 #endif
39293 #if (__ARM_FP & 2)
39294 #ifdef __LITTLE_ENDIAN__
vcvt_f16_f32(float32x4_t __p0)39295 __ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
39296   float16x4_t __ret;
39297   __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 41);
39298   return __ret;
39299 }
39300 #else
vcvt_f16_f32(float32x4_t __p0)39301 __ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
39302   float16x4_t __ret;
39303   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39304   __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__rev0, 41);
39305   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39306   return __ret;
39307 }
__noswap_vcvt_f16_f32(float32x4_t __p0)39308 __ai float16x4_t __noswap_vcvt_f16_f32(float32x4_t __p0) {
39309   float16x4_t __ret;
39310   __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 41);
39311   return __ret;
39312 }
39313 #endif
39314 
39315 #ifdef __LITTLE_ENDIAN__
vcvt_f32_f16(float16x4_t __p0)39316 __ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
39317   float32x4_t __ret;
39318   __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 8);
39319   return __ret;
39320 }
39321 #else
vcvt_f32_f16(float16x4_t __p0)39322 __ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
39323   float32x4_t __ret;
39324   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39325   __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__rev0, 8);
39326   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39327   return __ret;
39328 }
__noswap_vcvt_f32_f16(float16x4_t __p0)39329 __ai float32x4_t __noswap_vcvt_f32_f16(float16x4_t __p0) {
39330   float32x4_t __ret;
39331   __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 8);
39332   return __ret;
39333 }
39334 #endif
39335 
39336 #ifdef __LITTLE_ENDIAN__
39337 #define vld1q_f16(__p0) __extension__ ({ \
39338   float16x8_t __ret; \
39339   __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
39340   __ret; \
39341 })
39342 #else
39343 #define vld1q_f16(__p0) __extension__ ({ \
39344   float16x8_t __ret; \
39345   __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
39346   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
39347   __ret; \
39348 })
39349 #endif
39350 
39351 #ifdef __LITTLE_ENDIAN__
39352 #define vld1_f16(__p0) __extension__ ({ \
39353   float16x4_t __ret; \
39354   __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
39355   __ret; \
39356 })
39357 #else
39358 #define vld1_f16(__p0) __extension__ ({ \
39359   float16x4_t __ret; \
39360   __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
39361   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
39362   __ret; \
39363 })
39364 #endif
39365 
39366 #ifdef __LITTLE_ENDIAN__
39367 #define vld1q_dup_f16(__p0) __extension__ ({ \
39368   float16x8_t __ret; \
39369   __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
39370   __ret; \
39371 })
39372 #else
39373 #define vld1q_dup_f16(__p0) __extension__ ({ \
39374   float16x8_t __ret; \
39375   __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
39376   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
39377   __ret; \
39378 })
39379 #endif
39380 
39381 #ifdef __LITTLE_ENDIAN__
39382 #define vld1_dup_f16(__p0) __extension__ ({ \
39383   float16x4_t __ret; \
39384   __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
39385   __ret; \
39386 })
39387 #else
39388 #define vld1_dup_f16(__p0) __extension__ ({ \
39389   float16x4_t __ret; \
39390   __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
39391   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
39392   __ret; \
39393 })
39394 #endif
39395 
39396 #ifdef __LITTLE_ENDIAN__
39397 #define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39398   float16x8_t __ret; \
39399   float16x8_t __s1 = __p1; \
39400   __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
39401   __ret; \
39402 })
39403 #else
39404 #define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39405   float16x8_t __ret; \
39406   float16x8_t __s1 = __p1; \
39407   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
39408   __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
39409   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
39410   __ret; \
39411 })
39412 #endif
39413 
39414 #ifdef __LITTLE_ENDIAN__
39415 #define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39416   float16x4_t __ret; \
39417   float16x4_t __s1 = __p1; \
39418   __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
39419   __ret; \
39420 })
39421 #else
39422 #define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39423   float16x4_t __ret; \
39424   float16x4_t __s1 = __p1; \
39425   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
39426   __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
39427   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
39428   __ret; \
39429 })
39430 #endif
39431 
39432 #ifdef __LITTLE_ENDIAN__
39433 #define vld1q_f16_x2(__p0) __extension__ ({ \
39434   float16x8x2_t __ret; \
39435   __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
39436   __ret; \
39437 })
39438 #else
39439 #define vld1q_f16_x2(__p0) __extension__ ({ \
39440   float16x8x2_t __ret; \
39441   __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
39442  \
39443   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39444   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39445   __ret; \
39446 })
39447 #endif
39448 
39449 #ifdef __LITTLE_ENDIAN__
39450 #define vld1_f16_x2(__p0) __extension__ ({ \
39451   float16x4x2_t __ret; \
39452   __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
39453   __ret; \
39454 })
39455 #else
39456 #define vld1_f16_x2(__p0) __extension__ ({ \
39457   float16x4x2_t __ret; \
39458   __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
39459  \
39460   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39461   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39462   __ret; \
39463 })
39464 #endif
39465 
39466 #ifdef __LITTLE_ENDIAN__
39467 #define vld1q_f16_x3(__p0) __extension__ ({ \
39468   float16x8x3_t __ret; \
39469   __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
39470   __ret; \
39471 })
39472 #else
39473 #define vld1q_f16_x3(__p0) __extension__ ({ \
39474   float16x8x3_t __ret; \
39475   __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
39476  \
39477   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39478   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39479   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39480   __ret; \
39481 })
39482 #endif
39483 
39484 #ifdef __LITTLE_ENDIAN__
39485 #define vld1_f16_x3(__p0) __extension__ ({ \
39486   float16x4x3_t __ret; \
39487   __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
39488   __ret; \
39489 })
39490 #else
39491 #define vld1_f16_x3(__p0) __extension__ ({ \
39492   float16x4x3_t __ret; \
39493   __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
39494  \
39495   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39496   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39497   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39498   __ret; \
39499 })
39500 #endif
39501 
39502 #ifdef __LITTLE_ENDIAN__
39503 #define vld1q_f16_x4(__p0) __extension__ ({ \
39504   float16x8x4_t __ret; \
39505   __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
39506   __ret; \
39507 })
39508 #else
39509 #define vld1q_f16_x4(__p0) __extension__ ({ \
39510   float16x8x4_t __ret; \
39511   __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
39512  \
39513   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39514   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39515   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39516   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39517   __ret; \
39518 })
39519 #endif
39520 
39521 #ifdef __LITTLE_ENDIAN__
39522 #define vld1_f16_x4(__p0) __extension__ ({ \
39523   float16x4x4_t __ret; \
39524   __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
39525   __ret; \
39526 })
39527 #else
39528 #define vld1_f16_x4(__p0) __extension__ ({ \
39529   float16x4x4_t __ret; \
39530   __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
39531  \
39532   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39533   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39534   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39535   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
39536   __ret; \
39537 })
39538 #endif
39539 
39540 #ifdef __LITTLE_ENDIAN__
39541 #define vld2q_f16(__p0) __extension__ ({ \
39542   float16x8x2_t __ret; \
39543   __builtin_neon_vld2q_v(&__ret, __p0, 40); \
39544   __ret; \
39545 })
39546 #else
39547 #define vld2q_f16(__p0) __extension__ ({ \
39548   float16x8x2_t __ret; \
39549   __builtin_neon_vld2q_v(&__ret, __p0, 40); \
39550  \
39551   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39552   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39553   __ret; \
39554 })
39555 #endif
39556 
39557 #ifdef __LITTLE_ENDIAN__
39558 #define vld2_f16(__p0) __extension__ ({ \
39559   float16x4x2_t __ret; \
39560   __builtin_neon_vld2_v(&__ret, __p0, 8); \
39561   __ret; \
39562 })
39563 #else
39564 #define vld2_f16(__p0) __extension__ ({ \
39565   float16x4x2_t __ret; \
39566   __builtin_neon_vld2_v(&__ret, __p0, 8); \
39567  \
39568   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39569   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39570   __ret; \
39571 })
39572 #endif
39573 
39574 #ifdef __LITTLE_ENDIAN__
39575 #define vld2q_dup_f16(__p0) __extension__ ({ \
39576   float16x8x2_t __ret; \
39577   __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
39578   __ret; \
39579 })
39580 #else
39581 #define vld2q_dup_f16(__p0) __extension__ ({ \
39582   float16x8x2_t __ret; \
39583   __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
39584  \
39585   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39586   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39587   __ret; \
39588 })
39589 #endif
39590 
39591 #ifdef __LITTLE_ENDIAN__
39592 #define vld2_dup_f16(__p0) __extension__ ({ \
39593   float16x4x2_t __ret; \
39594   __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
39595   __ret; \
39596 })
39597 #else
39598 #define vld2_dup_f16(__p0) __extension__ ({ \
39599   float16x4x2_t __ret; \
39600   __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
39601  \
39602   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39603   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39604   __ret; \
39605 })
39606 #endif
39607 
39608 #ifdef __LITTLE_ENDIAN__
39609 #define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39610   float16x8x2_t __ret; \
39611   float16x8x2_t __s1 = __p1; \
39612   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 40); \
39613   __ret; \
39614 })
39615 #else
39616 #define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39617   float16x8x2_t __ret; \
39618   float16x8x2_t __s1 = __p1; \
39619   float16x8x2_t __rev1; \
39620   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39621   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39622   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 40); \
39623  \
39624   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39625   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39626   __ret; \
39627 })
39628 #endif
39629 
39630 #ifdef __LITTLE_ENDIAN__
39631 #define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39632   float16x4x2_t __ret; \
39633   float16x4x2_t __s1 = __p1; \
39634   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 8); \
39635   __ret; \
39636 })
39637 #else
39638 #define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39639   float16x4x2_t __ret; \
39640   float16x4x2_t __s1 = __p1; \
39641   float16x4x2_t __rev1; \
39642   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39643   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39644   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 8); \
39645  \
39646   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39647   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39648   __ret; \
39649 })
39650 #endif
39651 
39652 #ifdef __LITTLE_ENDIAN__
39653 #define vld3q_f16(__p0) __extension__ ({ \
39654   float16x8x3_t __ret; \
39655   __builtin_neon_vld3q_v(&__ret, __p0, 40); \
39656   __ret; \
39657 })
39658 #else
39659 #define vld3q_f16(__p0) __extension__ ({ \
39660   float16x8x3_t __ret; \
39661   __builtin_neon_vld3q_v(&__ret, __p0, 40); \
39662  \
39663   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39664   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39665   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39666   __ret; \
39667 })
39668 #endif
39669 
39670 #ifdef __LITTLE_ENDIAN__
39671 #define vld3_f16(__p0) __extension__ ({ \
39672   float16x4x3_t __ret; \
39673   __builtin_neon_vld3_v(&__ret, __p0, 8); \
39674   __ret; \
39675 })
39676 #else
39677 #define vld3_f16(__p0) __extension__ ({ \
39678   float16x4x3_t __ret; \
39679   __builtin_neon_vld3_v(&__ret, __p0, 8); \
39680  \
39681   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39682   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39683   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39684   __ret; \
39685 })
39686 #endif
39687 
39688 #ifdef __LITTLE_ENDIAN__
39689 #define vld3q_dup_f16(__p0) __extension__ ({ \
39690   float16x8x3_t __ret; \
39691   __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
39692   __ret; \
39693 })
39694 #else
39695 #define vld3q_dup_f16(__p0) __extension__ ({ \
39696   float16x8x3_t __ret; \
39697   __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
39698  \
39699   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39700   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39701   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39702   __ret; \
39703 })
39704 #endif
39705 
39706 #ifdef __LITTLE_ENDIAN__
39707 #define vld3_dup_f16(__p0) __extension__ ({ \
39708   float16x4x3_t __ret; \
39709   __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
39710   __ret; \
39711 })
39712 #else
39713 #define vld3_dup_f16(__p0) __extension__ ({ \
39714   float16x4x3_t __ret; \
39715   __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
39716  \
39717   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39718   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39719   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39720   __ret; \
39721 })
39722 #endif
39723 
39724 #ifdef __LITTLE_ENDIAN__
39725 #define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39726   float16x8x3_t __ret; \
39727   float16x8x3_t __s1 = __p1; \
39728   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 40); \
39729   __ret; \
39730 })
39731 #else
39732 #define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39733   float16x8x3_t __ret; \
39734   float16x8x3_t __s1 = __p1; \
39735   float16x8x3_t __rev1; \
39736   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39737   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39738   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39739   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 40); \
39740  \
39741   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39742   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39743   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39744   __ret; \
39745 })
39746 #endif
39747 
39748 #ifdef __LITTLE_ENDIAN__
39749 #define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39750   float16x4x3_t __ret; \
39751   float16x4x3_t __s1 = __p1; \
39752   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 8); \
39753   __ret; \
39754 })
39755 #else
39756 #define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39757   float16x4x3_t __ret; \
39758   float16x4x3_t __s1 = __p1; \
39759   float16x4x3_t __rev1; \
39760   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39761   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39762   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
39763   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 8); \
39764  \
39765   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39766   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39767   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39768   __ret; \
39769 })
39770 #endif
39771 
39772 #ifdef __LITTLE_ENDIAN__
39773 #define vld4q_f16(__p0) __extension__ ({ \
39774   float16x8x4_t __ret; \
39775   __builtin_neon_vld4q_v(&__ret, __p0, 40); \
39776   __ret; \
39777 })
39778 #else
39779 #define vld4q_f16(__p0) __extension__ ({ \
39780   float16x8x4_t __ret; \
39781   __builtin_neon_vld4q_v(&__ret, __p0, 40); \
39782  \
39783   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39784   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39785   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39786   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39787   __ret; \
39788 })
39789 #endif
39790 
39791 #ifdef __LITTLE_ENDIAN__
39792 #define vld4_f16(__p0) __extension__ ({ \
39793   float16x4x4_t __ret; \
39794   __builtin_neon_vld4_v(&__ret, __p0, 8); \
39795   __ret; \
39796 })
39797 #else
39798 #define vld4_f16(__p0) __extension__ ({ \
39799   float16x4x4_t __ret; \
39800   __builtin_neon_vld4_v(&__ret, __p0, 8); \
39801  \
39802   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39803   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39804   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39805   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
39806   __ret; \
39807 })
39808 #endif
39809 
39810 #ifdef __LITTLE_ENDIAN__
39811 #define vld4q_dup_f16(__p0) __extension__ ({ \
39812   float16x8x4_t __ret; \
39813   __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
39814   __ret; \
39815 })
39816 #else
39817 #define vld4q_dup_f16(__p0) __extension__ ({ \
39818   float16x8x4_t __ret; \
39819   __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
39820  \
39821   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39822   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39823   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39824   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39825   __ret; \
39826 })
39827 #endif
39828 
39829 #ifdef __LITTLE_ENDIAN__
39830 #define vld4_dup_f16(__p0) __extension__ ({ \
39831   float16x4x4_t __ret; \
39832   __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
39833   __ret; \
39834 })
39835 #else
39836 #define vld4_dup_f16(__p0) __extension__ ({ \
39837   float16x4x4_t __ret; \
39838   __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
39839  \
39840   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39841   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39842   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39843   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
39844   __ret; \
39845 })
39846 #endif
39847 
39848 #ifdef __LITTLE_ENDIAN__
39849 #define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39850   float16x8x4_t __ret; \
39851   float16x8x4_t __s1 = __p1; \
39852   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 40); \
39853   __ret; \
39854 })
39855 #else
39856 #define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39857   float16x8x4_t __ret; \
39858   float16x8x4_t __s1 = __p1; \
39859   float16x8x4_t __rev1; \
39860   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39861   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39862   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39863   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39864   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 40); \
39865  \
39866   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39867   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39868   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39869   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39870   __ret; \
39871 })
39872 #endif
39873 
39874 #ifdef __LITTLE_ENDIAN__
39875 #define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39876   float16x4x4_t __ret; \
39877   float16x4x4_t __s1 = __p1; \
39878   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 8); \
39879   __ret; \
39880 })
39881 #else
39882 #define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39883   float16x4x4_t __ret; \
39884   float16x4x4_t __s1 = __p1; \
39885   float16x4x4_t __rev1; \
39886   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39887   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39888   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
39889   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
39890   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 8); \
39891  \
39892   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39893   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39894   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39895   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
39896   __ret; \
39897 })
39898 #endif
39899 
39900 #ifdef __LITTLE_ENDIAN__
39901 #define vst1q_f16(__p0, __p1) __extension__ ({ \
39902   float16x8_t __s1 = __p1; \
39903   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 40); \
39904 })
39905 #else
39906 #define vst1q_f16(__p0, __p1) __extension__ ({ \
39907   float16x8_t __s1 = __p1; \
39908   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
39909   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 40); \
39910 })
39911 #endif
39912 
39913 #ifdef __LITTLE_ENDIAN__
39914 #define vst1_f16(__p0, __p1) __extension__ ({ \
39915   float16x4_t __s1 = __p1; \
39916   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 8); \
39917 })
39918 #else
39919 #define vst1_f16(__p0, __p1) __extension__ ({ \
39920   float16x4_t __s1 = __p1; \
39921   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
39922   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 8); \
39923 })
39924 #endif
39925 
39926 #ifdef __LITTLE_ENDIAN__
39927 #define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39928   float16x8_t __s1 = __p1; \
39929   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
39930 })
39931 #else
39932 #define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39933   float16x8_t __s1 = __p1; \
39934   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
39935   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
39936 })
39937 #endif
39938 
39939 #ifdef __LITTLE_ENDIAN__
39940 #define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39941   float16x4_t __s1 = __p1; \
39942   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
39943 })
39944 #else
39945 #define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
39946   float16x4_t __s1 = __p1; \
39947   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
39948   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
39949 })
39950 #endif
39951 
39952 #ifdef __LITTLE_ENDIAN__
39953 #define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
39954   float16x8x2_t __s1 = __p1; \
39955   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 40); \
39956 })
39957 #else
39958 #define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
39959   float16x8x2_t __s1 = __p1; \
39960   float16x8x2_t __rev1; \
39961   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39962   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39963   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 40); \
39964 })
39965 #endif
39966 
39967 #ifdef __LITTLE_ENDIAN__
39968 #define vst1_f16_x2(__p0, __p1) __extension__ ({ \
39969   float16x4x2_t __s1 = __p1; \
39970   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 8); \
39971 })
39972 #else
39973 #define vst1_f16_x2(__p0, __p1) __extension__ ({ \
39974   float16x4x2_t __s1 = __p1; \
39975   float16x4x2_t __rev1; \
39976   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39977   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39978   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 8); \
39979 })
39980 #endif
39981 
39982 #ifdef __LITTLE_ENDIAN__
39983 #define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
39984   float16x8x3_t __s1 = __p1; \
39985   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 40); \
39986 })
39987 #else
39988 #define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
39989   float16x8x3_t __s1 = __p1; \
39990   float16x8x3_t __rev1; \
39991   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39992   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39993   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39994   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 40); \
39995 })
39996 #endif
39997 
39998 #ifdef __LITTLE_ENDIAN__
39999 #define vst1_f16_x3(__p0, __p1) __extension__ ({ \
40000   float16x4x3_t __s1 = __p1; \
40001   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 8); \
40002 })
40003 #else
40004 #define vst1_f16_x3(__p0, __p1) __extension__ ({ \
40005   float16x4x3_t __s1 = __p1; \
40006   float16x4x3_t __rev1; \
40007   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40008   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40009   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40010   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 8); \
40011 })
40012 #endif
40013 
40014 #ifdef __LITTLE_ENDIAN__
40015 #define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
40016   float16x8x4_t __s1 = __p1; \
40017   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 40); \
40018 })
40019 #else
40020 #define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
40021   float16x8x4_t __s1 = __p1; \
40022   float16x8x4_t __rev1; \
40023   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40024   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40025   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
40026   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
40027   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 40); \
40028 })
40029 #endif
40030 
40031 #ifdef __LITTLE_ENDIAN__
40032 #define vst1_f16_x4(__p0, __p1) __extension__ ({ \
40033   float16x4x4_t __s1 = __p1; \
40034   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 8); \
40035 })
40036 #else
40037 #define vst1_f16_x4(__p0, __p1) __extension__ ({ \
40038   float16x4x4_t __s1 = __p1; \
40039   float16x4x4_t __rev1; \
40040   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40041   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40042   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40043   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
40044   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 8); \
40045 })
40046 #endif
40047 
40048 #ifdef __LITTLE_ENDIAN__
40049 #define vst2q_f16(__p0, __p1) __extension__ ({ \
40050   float16x8x2_t __s1 = __p1; \
40051   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 40); \
40052 })
40053 #else
40054 #define vst2q_f16(__p0, __p1) __extension__ ({ \
40055   float16x8x2_t __s1 = __p1; \
40056   float16x8x2_t __rev1; \
40057   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40058   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40059   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 40); \
40060 })
40061 #endif
40062 
40063 #ifdef __LITTLE_ENDIAN__
40064 #define vst2_f16(__p0, __p1) __extension__ ({ \
40065   float16x4x2_t __s1 = __p1; \
40066   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 8); \
40067 })
40068 #else
40069 #define vst2_f16(__p0, __p1) __extension__ ({ \
40070   float16x4x2_t __s1 = __p1; \
40071   float16x4x2_t __rev1; \
40072   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40073   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40074   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 8); \
40075 })
40076 #endif
40077 
40078 #ifdef __LITTLE_ENDIAN__
40079 #define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40080   float16x8x2_t __s1 = __p1; \
40081   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 40); \
40082 })
40083 #else
40084 #define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40085   float16x8x2_t __s1 = __p1; \
40086   float16x8x2_t __rev1; \
40087   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40088   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40089   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 40); \
40090 })
40091 #endif
40092 
40093 #ifdef __LITTLE_ENDIAN__
40094 #define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40095   float16x4x2_t __s1 = __p1; \
40096   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 8); \
40097 })
40098 #else
40099 #define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40100   float16x4x2_t __s1 = __p1; \
40101   float16x4x2_t __rev1; \
40102   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40103   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40104   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 8); \
40105 })
40106 #endif
40107 
40108 #ifdef __LITTLE_ENDIAN__
40109 #define vst3q_f16(__p0, __p1) __extension__ ({ \
40110   float16x8x3_t __s1 = __p1; \
40111   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 40); \
40112 })
40113 #else
40114 #define vst3q_f16(__p0, __p1) __extension__ ({ \
40115   float16x8x3_t __s1 = __p1; \
40116   float16x8x3_t __rev1; \
40117   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40118   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40119   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
40120   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 40); \
40121 })
40122 #endif
40123 
40124 #ifdef __LITTLE_ENDIAN__
40125 #define vst3_f16(__p0, __p1) __extension__ ({ \
40126   float16x4x3_t __s1 = __p1; \
40127   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 8); \
40128 })
40129 #else
40130 #define vst3_f16(__p0, __p1) __extension__ ({ \
40131   float16x4x3_t __s1 = __p1; \
40132   float16x4x3_t __rev1; \
40133   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40134   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40135   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40136   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 8); \
40137 })
40138 #endif
40139 
40140 #ifdef __LITTLE_ENDIAN__
40141 #define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40142   float16x8x3_t __s1 = __p1; \
40143   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 40); \
40144 })
40145 #else
40146 #define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40147   float16x8x3_t __s1 = __p1; \
40148   float16x8x3_t __rev1; \
40149   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40150   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40151   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
40152   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 40); \
40153 })
40154 #endif
40155 
40156 #ifdef __LITTLE_ENDIAN__
40157 #define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40158   float16x4x3_t __s1 = __p1; \
40159   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 8); \
40160 })
40161 #else
40162 #define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40163   float16x4x3_t __s1 = __p1; \
40164   float16x4x3_t __rev1; \
40165   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40166   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40167   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40168   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 8); \
40169 })
40170 #endif
40171 
40172 #ifdef __LITTLE_ENDIAN__
40173 #define vst4q_f16(__p0, __p1) __extension__ ({ \
40174   float16x8x4_t __s1 = __p1; \
40175   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 40); \
40176 })
40177 #else
40178 #define vst4q_f16(__p0, __p1) __extension__ ({ \
40179   float16x8x4_t __s1 = __p1; \
40180   float16x8x4_t __rev1; \
40181   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40182   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40183   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
40184   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
40185   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 40); \
40186 })
40187 #endif
40188 
40189 #ifdef __LITTLE_ENDIAN__
40190 #define vst4_f16(__p0, __p1) __extension__ ({ \
40191   float16x4x4_t __s1 = __p1; \
40192   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 8); \
40193 })
40194 #else
40195 #define vst4_f16(__p0, __p1) __extension__ ({ \
40196   float16x4x4_t __s1 = __p1; \
40197   float16x4x4_t __rev1; \
40198   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40199   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40200   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40201   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
40202   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 8); \
40203 })
40204 #endif
40205 
40206 #ifdef __LITTLE_ENDIAN__
40207 #define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40208   float16x8x4_t __s1 = __p1; \
40209   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 40); \
40210 })
40211 #else
40212 #define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40213   float16x8x4_t __s1 = __p1; \
40214   float16x8x4_t __rev1; \
40215   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40216   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40217   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
40218   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
40219   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 40); \
40220 })
40221 #endif
40222 
40223 #ifdef __LITTLE_ENDIAN__
40224 #define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40225   float16x4x4_t __s1 = __p1; \
40226   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 8); \
40227 })
40228 #else
40229 #define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
40230   float16x4x4_t __s1 = __p1; \
40231   float16x4x4_t __rev1; \
40232   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40233   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40234   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40235   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
40236   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 8); \
40237 })
40238 #endif
40239 
40240 #endif
40241 #if __ARM_ARCH >= 8
40242 #ifdef __LITTLE_ENDIAN__
vcvtaq_s32_f32(float32x4_t __p0)40243 __ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
40244   int32x4_t __ret;
40245   __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__p0, 34);
40246   return __ret;
40247 }
40248 #else
vcvtaq_s32_f32(float32x4_t __p0)40249 __ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
40250   int32x4_t __ret;
40251   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40252   __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__rev0, 34);
40253   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40254   return __ret;
40255 }
40256 #endif
40257 
40258 #ifdef __LITTLE_ENDIAN__
vcvta_s32_f32(float32x2_t __p0)40259 __ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
40260   int32x2_t __ret;
40261   __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__p0, 2);
40262   return __ret;
40263 }
40264 #else
vcvta_s32_f32(float32x2_t __p0)40265 __ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
40266   int32x2_t __ret;
40267   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40268   __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__rev0, 2);
40269   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40270   return __ret;
40271 }
40272 #endif
40273 
40274 #ifdef __LITTLE_ENDIAN__
vcvtaq_u32_f32(float32x4_t __p0)40275 __ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
40276   uint32x4_t __ret;
40277   __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__p0, 50);
40278   return __ret;
40279 }
40280 #else
vcvtaq_u32_f32(float32x4_t __p0)40281 __ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
40282   uint32x4_t __ret;
40283   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40284   __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__rev0, 50);
40285   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40286   return __ret;
40287 }
40288 #endif
40289 
40290 #ifdef __LITTLE_ENDIAN__
vcvta_u32_f32(float32x2_t __p0)40291 __ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
40292   uint32x2_t __ret;
40293   __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__p0, 18);
40294   return __ret;
40295 }
40296 #else
vcvta_u32_f32(float32x2_t __p0)40297 __ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
40298   uint32x2_t __ret;
40299   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40300   __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__rev0, 18);
40301   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40302   return __ret;
40303 }
40304 #endif
40305 
40306 #ifdef __LITTLE_ENDIAN__
vcvtmq_s32_f32(float32x4_t __p0)40307 __ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
40308   int32x4_t __ret;
40309   __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__p0, 34);
40310   return __ret;
40311 }
40312 #else
vcvtmq_s32_f32(float32x4_t __p0)40313 __ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
40314   int32x4_t __ret;
40315   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40316   __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__rev0, 34);
40317   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40318   return __ret;
40319 }
40320 #endif
40321 
40322 #ifdef __LITTLE_ENDIAN__
vcvtm_s32_f32(float32x2_t __p0)40323 __ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
40324   int32x2_t __ret;
40325   __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__p0, 2);
40326   return __ret;
40327 }
40328 #else
vcvtm_s32_f32(float32x2_t __p0)40329 __ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
40330   int32x2_t __ret;
40331   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40332   __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__rev0, 2);
40333   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40334   return __ret;
40335 }
40336 #endif
40337 
40338 #ifdef __LITTLE_ENDIAN__
vcvtmq_u32_f32(float32x4_t __p0)40339 __ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
40340   uint32x4_t __ret;
40341   __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__p0, 50);
40342   return __ret;
40343 }
40344 #else
vcvtmq_u32_f32(float32x4_t __p0)40345 __ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
40346   uint32x4_t __ret;
40347   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40348   __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__rev0, 50);
40349   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40350   return __ret;
40351 }
40352 #endif
40353 
40354 #ifdef __LITTLE_ENDIAN__
vcvtm_u32_f32(float32x2_t __p0)40355 __ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
40356   uint32x2_t __ret;
40357   __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__p0, 18);
40358   return __ret;
40359 }
40360 #else
vcvtm_u32_f32(float32x2_t __p0)40361 __ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
40362   uint32x2_t __ret;
40363   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40364   __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__rev0, 18);
40365   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40366   return __ret;
40367 }
40368 #endif
40369 
40370 #ifdef __LITTLE_ENDIAN__
vcvtnq_s32_f32(float32x4_t __p0)40371 __ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
40372   int32x4_t __ret;
40373   __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__p0, 34);
40374   return __ret;
40375 }
40376 #else
vcvtnq_s32_f32(float32x4_t __p0)40377 __ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
40378   int32x4_t __ret;
40379   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40380   __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__rev0, 34);
40381   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40382   return __ret;
40383 }
40384 #endif
40385 
40386 #ifdef __LITTLE_ENDIAN__
vcvtn_s32_f32(float32x2_t __p0)40387 __ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
40388   int32x2_t __ret;
40389   __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__p0, 2);
40390   return __ret;
40391 }
40392 #else
vcvtn_s32_f32(float32x2_t __p0)40393 __ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
40394   int32x2_t __ret;
40395   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40396   __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__rev0, 2);
40397   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40398   return __ret;
40399 }
40400 #endif
40401 
40402 #ifdef __LITTLE_ENDIAN__
vcvtnq_u32_f32(float32x4_t __p0)40403 __ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
40404   uint32x4_t __ret;
40405   __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__p0, 50);
40406   return __ret;
40407 }
40408 #else
vcvtnq_u32_f32(float32x4_t __p0)40409 __ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
40410   uint32x4_t __ret;
40411   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40412   __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__rev0, 50);
40413   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40414   return __ret;
40415 }
40416 #endif
40417 
40418 #ifdef __LITTLE_ENDIAN__
vcvtn_u32_f32(float32x2_t __p0)40419 __ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
40420   uint32x2_t __ret;
40421   __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__p0, 18);
40422   return __ret;
40423 }
40424 #else
vcvtn_u32_f32(float32x2_t __p0)40425 __ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
40426   uint32x2_t __ret;
40427   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40428   __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__rev0, 18);
40429   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40430   return __ret;
40431 }
40432 #endif
40433 
40434 #ifdef __LITTLE_ENDIAN__
vcvtpq_s32_f32(float32x4_t __p0)40435 __ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
40436   int32x4_t __ret;
40437   __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__p0, 34);
40438   return __ret;
40439 }
40440 #else
vcvtpq_s32_f32(float32x4_t __p0)40441 __ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
40442   int32x4_t __ret;
40443   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40444   __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__rev0, 34);
40445   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40446   return __ret;
40447 }
40448 #endif
40449 
40450 #ifdef __LITTLE_ENDIAN__
vcvtp_s32_f32(float32x2_t __p0)40451 __ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
40452   int32x2_t __ret;
40453   __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__p0, 2);
40454   return __ret;
40455 }
40456 #else
vcvtp_s32_f32(float32x2_t __p0)40457 __ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
40458   int32x2_t __ret;
40459   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40460   __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__rev0, 2);
40461   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40462   return __ret;
40463 }
40464 #endif
40465 
40466 #ifdef __LITTLE_ENDIAN__
vcvtpq_u32_f32(float32x4_t __p0)40467 __ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
40468   uint32x4_t __ret;
40469   __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__p0, 50);
40470   return __ret;
40471 }
40472 #else
vcvtpq_u32_f32(float32x4_t __p0)40473 __ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
40474   uint32x4_t __ret;
40475   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40476   __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__rev0, 50);
40477   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40478   return __ret;
40479 }
40480 #endif
40481 
40482 #ifdef __LITTLE_ENDIAN__
vcvtp_u32_f32(float32x2_t __p0)40483 __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
40484   uint32x2_t __ret;
40485   __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__p0, 18);
40486   return __ret;
40487 }
40488 #else
vcvtp_u32_f32(float32x2_t __p0)40489 __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
40490   uint32x2_t __ret;
40491   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40492   __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__rev0, 18);
40493   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40494   return __ret;
40495 }
40496 #endif
40497 
40498 #ifdef __LITTLE_ENDIAN__
vaesdq_u8(uint8x16_t __p0,uint8x16_t __p1)40499 __ai __attribute__((target("aes"))) uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40500   uint8x16_t __ret;
40501   __ret = (uint8x16_t) __builtin_neon_vaesdq_u8((int8x16_t)__p0, (int8x16_t)__p1, 48);
40502   return __ret;
40503 }
40504 #else
vaesdq_u8(uint8x16_t __p0,uint8x16_t __p1)40505 __ai __attribute__((target("aes"))) uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40506   uint8x16_t __ret;
40507   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40508   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40509   __ret = (uint8x16_t) __builtin_neon_vaesdq_u8((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
40510   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40511   return __ret;
40512 }
40513 #endif
40514 
40515 #ifdef __LITTLE_ENDIAN__
vaeseq_u8(uint8x16_t __p0,uint8x16_t __p1)40516 __ai __attribute__((target("aes"))) uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40517   uint8x16_t __ret;
40518   __ret = (uint8x16_t) __builtin_neon_vaeseq_u8((int8x16_t)__p0, (int8x16_t)__p1, 48);
40519   return __ret;
40520 }
40521 #else
vaeseq_u8(uint8x16_t __p0,uint8x16_t __p1)40522 __ai __attribute__((target("aes"))) uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40523   uint8x16_t __ret;
40524   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40525   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40526   __ret = (uint8x16_t) __builtin_neon_vaeseq_u8((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
40527   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40528   return __ret;
40529 }
40530 #endif
40531 
40532 #ifdef __LITTLE_ENDIAN__
vaesimcq_u8(uint8x16_t __p0)40533 __ai __attribute__((target("aes"))) uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
40534   uint8x16_t __ret;
40535   __ret = (uint8x16_t) __builtin_neon_vaesimcq_u8((int8x16_t)__p0, 48);
40536   return __ret;
40537 }
40538 #else
vaesimcq_u8(uint8x16_t __p0)40539 __ai __attribute__((target("aes"))) uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
40540   uint8x16_t __ret;
40541   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40542   __ret = (uint8x16_t) __builtin_neon_vaesimcq_u8((int8x16_t)__rev0, 48);
40543   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40544   return __ret;
40545 }
40546 #endif
40547 
40548 #ifdef __LITTLE_ENDIAN__
vaesmcq_u8(uint8x16_t __p0)40549 __ai __attribute__((target("aes"))) uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
40550   uint8x16_t __ret;
40551   __ret = (uint8x16_t) __builtin_neon_vaesmcq_u8((int8x16_t)__p0, 48);
40552   return __ret;
40553 }
40554 #else
vaesmcq_u8(uint8x16_t __p0)40555 __ai __attribute__((target("aes"))) uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
40556   uint8x16_t __ret;
40557   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40558   __ret = (uint8x16_t) __builtin_neon_vaesmcq_u8((int8x16_t)__rev0, 48);
40559   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40560   return __ret;
40561 }
40562 #endif
40563 
40564 #ifdef __LITTLE_ENDIAN__
vsha1cq_u32(uint32x4_t __p0,uint32_t __p1,uint32x4_t __p2)40565 __ai __attribute__((target("sha2"))) uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40566   uint32x4_t __ret;
40567   __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__p0, __p1, __p2);
40568   return __ret;
40569 }
40570 #else
vsha1cq_u32(uint32x4_t __p0,uint32_t __p1,uint32x4_t __p2)40571 __ai __attribute__((target("sha2"))) uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40572   uint32x4_t __ret;
40573   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40574   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40575   __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__rev0, __p1, __rev2);
40576   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40577   return __ret;
40578 }
40579 #endif
40580 
vsha1h_u32(uint32_t __p0)40581 __ai __attribute__((target("sha2"))) uint32_t vsha1h_u32(uint32_t __p0) {
40582   uint32_t __ret;
40583   __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
40584   return __ret;
40585 }
40586 #ifdef __LITTLE_ENDIAN__
vsha1mq_u32(uint32x4_t __p0,uint32_t __p1,uint32x4_t __p2)40587 __ai __attribute__((target("sha2"))) uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40588   uint32x4_t __ret;
40589   __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__p0, __p1, __p2);
40590   return __ret;
40591 }
40592 #else
vsha1mq_u32(uint32x4_t __p0,uint32_t __p1,uint32x4_t __p2)40593 __ai __attribute__((target("sha2"))) uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40594   uint32x4_t __ret;
40595   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40596   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40597   __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__rev0, __p1, __rev2);
40598   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40599   return __ret;
40600 }
40601 #endif
40602 
40603 #ifdef __LITTLE_ENDIAN__
vsha1pq_u32(uint32x4_t __p0,uint32_t __p1,uint32x4_t __p2)40604 __ai __attribute__((target("sha2"))) uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40605   uint32x4_t __ret;
40606   __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__p0, __p1, __p2);
40607   return __ret;
40608 }
40609 #else
vsha1pq_u32(uint32x4_t __p0,uint32_t __p1,uint32x4_t __p2)40610 __ai __attribute__((target("sha2"))) uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40611   uint32x4_t __ret;
40612   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40613   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40614   __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__rev0, __p1, __rev2);
40615   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40616   return __ret;
40617 }
40618 #endif
40619 
40620 #ifdef __LITTLE_ENDIAN__
vsha1su0q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)40621 __ai __attribute__((target("sha2"))) uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40622   uint32x4_t __ret;
40623   __ret = (uint32x4_t) __builtin_neon_vsha1su0q_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40624   return __ret;
40625 }
40626 #else
vsha1su0q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)40627 __ai __attribute__((target("sha2"))) uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40628   uint32x4_t __ret;
40629   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40630   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40631   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40632   __ret = (uint32x4_t) __builtin_neon_vsha1su0q_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40633   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40634   return __ret;
40635 }
40636 #endif
40637 
40638 #ifdef __LITTLE_ENDIAN__
vsha1su1q_u32(uint32x4_t __p0,uint32x4_t __p1)40639 __ai __attribute__((target("sha2"))) uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40640   uint32x4_t __ret;
40641   __ret = (uint32x4_t) __builtin_neon_vsha1su1q_u32((int8x16_t)__p0, (int8x16_t)__p1, 50);
40642   return __ret;
40643 }
40644 #else
vsha1su1q_u32(uint32x4_t __p0,uint32x4_t __p1)40645 __ai __attribute__((target("sha2"))) uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40646   uint32x4_t __ret;
40647   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40648   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40649   __ret = (uint32x4_t) __builtin_neon_vsha1su1q_u32((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
40650   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40651   return __ret;
40652 }
40653 #endif
40654 
40655 #ifdef __LITTLE_ENDIAN__
vsha256hq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)40656 __ai __attribute__((target("sha2"))) uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40657   uint32x4_t __ret;
40658   __ret = (uint32x4_t) __builtin_neon_vsha256hq_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40659   return __ret;
40660 }
40661 #else
vsha256hq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)40662 __ai __attribute__((target("sha2"))) uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40663   uint32x4_t __ret;
40664   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40665   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40666   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40667   __ret = (uint32x4_t) __builtin_neon_vsha256hq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40668   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40669   return __ret;
40670 }
40671 #endif
40672 
40673 #ifdef __LITTLE_ENDIAN__
vsha256h2q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)40674 __ai __attribute__((target("sha2"))) uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40675   uint32x4_t __ret;
40676   __ret = (uint32x4_t) __builtin_neon_vsha256h2q_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40677   return __ret;
40678 }
40679 #else
vsha256h2q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)40680 __ai __attribute__((target("sha2"))) uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40681   uint32x4_t __ret;
40682   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40683   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40684   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40685   __ret = (uint32x4_t) __builtin_neon_vsha256h2q_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40686   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40687   return __ret;
40688 }
40689 #endif
40690 
40691 #ifdef __LITTLE_ENDIAN__
vsha256su0q_u32(uint32x4_t __p0,uint32x4_t __p1)40692 __ai __attribute__((target("sha2"))) uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40693   uint32x4_t __ret;
40694   __ret = (uint32x4_t) __builtin_neon_vsha256su0q_u32((int8x16_t)__p0, (int8x16_t)__p1, 50);
40695   return __ret;
40696 }
40697 #else
vsha256su0q_u32(uint32x4_t __p0,uint32x4_t __p1)40698 __ai __attribute__((target("sha2"))) uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40699   uint32x4_t __ret;
40700   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40701   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40702   __ret = (uint32x4_t) __builtin_neon_vsha256su0q_u32((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
40703   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40704   return __ret;
40705 }
40706 #endif
40707 
40708 #ifdef __LITTLE_ENDIAN__
vsha256su1q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)40709 __ai __attribute__((target("sha2"))) uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40710   uint32x4_t __ret;
40711   __ret = (uint32x4_t) __builtin_neon_vsha256su1q_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40712   return __ret;
40713 }
40714 #else
vsha256su1q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)40715 __ai __attribute__((target("sha2"))) uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40716   uint32x4_t __ret;
40717   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40718   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40719   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40720   __ret = (uint32x4_t) __builtin_neon_vsha256su1q_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40721   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40722   return __ret;
40723 }
40724 #endif
40725 
40726 #endif
40727 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
40728 #ifdef __LITTLE_ENDIAN__
vrndq_f32(float32x4_t __p0)40729 __ai float32x4_t vrndq_f32(float32x4_t __p0) {
40730   float32x4_t __ret;
40731   __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 41);
40732   return __ret;
40733 }
40734 #else
vrndq_f32(float32x4_t __p0)40735 __ai float32x4_t vrndq_f32(float32x4_t __p0) {
40736   float32x4_t __ret;
40737   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40738   __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 41);
40739   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40740   return __ret;
40741 }
40742 #endif
40743 
40744 #ifdef __LITTLE_ENDIAN__
vrnd_f32(float32x2_t __p0)40745 __ai float32x2_t vrnd_f32(float32x2_t __p0) {
40746   float32x2_t __ret;
40747   __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 9);
40748   return __ret;
40749 }
40750 #else
vrnd_f32(float32x2_t __p0)40751 __ai float32x2_t vrnd_f32(float32x2_t __p0) {
40752   float32x2_t __ret;
40753   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40754   __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__rev0, 9);
40755   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40756   return __ret;
40757 }
40758 #endif
40759 
40760 #ifdef __LITTLE_ENDIAN__
vrndaq_f32(float32x4_t __p0)40761 __ai float32x4_t vrndaq_f32(float32x4_t __p0) {
40762   float32x4_t __ret;
40763   __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 41);
40764   return __ret;
40765 }
40766 #else
vrndaq_f32(float32x4_t __p0)40767 __ai float32x4_t vrndaq_f32(float32x4_t __p0) {
40768   float32x4_t __ret;
40769   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40770   __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 41);
40771   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40772   return __ret;
40773 }
40774 #endif
40775 
40776 #ifdef __LITTLE_ENDIAN__
vrnda_f32(float32x2_t __p0)40777 __ai float32x2_t vrnda_f32(float32x2_t __p0) {
40778   float32x2_t __ret;
40779   __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 9);
40780   return __ret;
40781 }
40782 #else
vrnda_f32(float32x2_t __p0)40783 __ai float32x2_t vrnda_f32(float32x2_t __p0) {
40784   float32x2_t __ret;
40785   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40786   __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__rev0, 9);
40787   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40788   return __ret;
40789 }
40790 #endif
40791 
40792 #ifdef __LITTLE_ENDIAN__
vrndiq_f32(float32x4_t __p0)40793 __ai float32x4_t vrndiq_f32(float32x4_t __p0) {
40794   float32x4_t __ret;
40795   __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 41);
40796   return __ret;
40797 }
40798 #else
vrndiq_f32(float32x4_t __p0)40799 __ai float32x4_t vrndiq_f32(float32x4_t __p0) {
40800   float32x4_t __ret;
40801   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40802   __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 41);
40803   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40804   return __ret;
40805 }
40806 #endif
40807 
40808 #ifdef __LITTLE_ENDIAN__
vrndi_f32(float32x2_t __p0)40809 __ai float32x2_t vrndi_f32(float32x2_t __p0) {
40810   float32x2_t __ret;
40811   __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 9);
40812   return __ret;
40813 }
40814 #else
vrndi_f32(float32x2_t __p0)40815 __ai float32x2_t vrndi_f32(float32x2_t __p0) {
40816   float32x2_t __ret;
40817   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40818   __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__rev0, 9);
40819   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40820   return __ret;
40821 }
40822 #endif
40823 
40824 #ifdef __LITTLE_ENDIAN__
vrndmq_f32(float32x4_t __p0)40825 __ai float32x4_t vrndmq_f32(float32x4_t __p0) {
40826   float32x4_t __ret;
40827   __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 41);
40828   return __ret;
40829 }
40830 #else
vrndmq_f32(float32x4_t __p0)40831 __ai float32x4_t vrndmq_f32(float32x4_t __p0) {
40832   float32x4_t __ret;
40833   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40834   __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 41);
40835   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40836   return __ret;
40837 }
40838 #endif
40839 
40840 #ifdef __LITTLE_ENDIAN__
vrndm_f32(float32x2_t __p0)40841 __ai float32x2_t vrndm_f32(float32x2_t __p0) {
40842   float32x2_t __ret;
40843   __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 9);
40844   return __ret;
40845 }
40846 #else
vrndm_f32(float32x2_t __p0)40847 __ai float32x2_t vrndm_f32(float32x2_t __p0) {
40848   float32x2_t __ret;
40849   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40850   __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__rev0, 9);
40851   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40852   return __ret;
40853 }
40854 #endif
40855 
40856 #ifdef __LITTLE_ENDIAN__
vrndnq_f32(float32x4_t __p0)40857 __ai float32x4_t vrndnq_f32(float32x4_t __p0) {
40858   float32x4_t __ret;
40859   __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 41);
40860   return __ret;
40861 }
40862 #else
vrndnq_f32(float32x4_t __p0)40863 __ai float32x4_t vrndnq_f32(float32x4_t __p0) {
40864   float32x4_t __ret;
40865   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40866   __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 41);
40867   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40868   return __ret;
40869 }
40870 #endif
40871 
40872 #ifdef __LITTLE_ENDIAN__
vrndn_f32(float32x2_t __p0)40873 __ai float32x2_t vrndn_f32(float32x2_t __p0) {
40874   float32x2_t __ret;
40875   __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 9);
40876   return __ret;
40877 }
40878 #else
vrndn_f32(float32x2_t __p0)40879 __ai float32x2_t vrndn_f32(float32x2_t __p0) {
40880   float32x2_t __ret;
40881   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40882   __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__rev0, 9);
40883   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40884   return __ret;
40885 }
40886 #endif
40887 
vrndns_f32(float32_t __p0)40888 __ai float32_t vrndns_f32(float32_t __p0) {
40889   float32_t __ret;
40890   __ret = (float32_t) __builtin_neon_vrndns_f32(__p0);
40891   return __ret;
40892 }
40893 #ifdef __LITTLE_ENDIAN__
vrndpq_f32(float32x4_t __p0)40894 __ai float32x4_t vrndpq_f32(float32x4_t __p0) {
40895   float32x4_t __ret;
40896   __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 41);
40897   return __ret;
40898 }
40899 #else
vrndpq_f32(float32x4_t __p0)40900 __ai float32x4_t vrndpq_f32(float32x4_t __p0) {
40901   float32x4_t __ret;
40902   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40903   __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 41);
40904   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40905   return __ret;
40906 }
40907 #endif
40908 
40909 #ifdef __LITTLE_ENDIAN__
vrndp_f32(float32x2_t __p0)40910 __ai float32x2_t vrndp_f32(float32x2_t __p0) {
40911   float32x2_t __ret;
40912   __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 9);
40913   return __ret;
40914 }
40915 #else
vrndp_f32(float32x2_t __p0)40916 __ai float32x2_t vrndp_f32(float32x2_t __p0) {
40917   float32x2_t __ret;
40918   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40919   __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__rev0, 9);
40920   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40921   return __ret;
40922 }
40923 #endif
40924 
40925 #ifdef __LITTLE_ENDIAN__
vrndxq_f32(float32x4_t __p0)40926 __ai float32x4_t vrndxq_f32(float32x4_t __p0) {
40927   float32x4_t __ret;
40928   __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 41);
40929   return __ret;
40930 }
40931 #else
vrndxq_f32(float32x4_t __p0)40932 __ai float32x4_t vrndxq_f32(float32x4_t __p0) {
40933   float32x4_t __ret;
40934   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40935   __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 41);
40936   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40937   return __ret;
40938 }
40939 #endif
40940 
40941 #ifdef __LITTLE_ENDIAN__
vrndx_f32(float32x2_t __p0)40942 __ai float32x2_t vrndx_f32(float32x2_t __p0) {
40943   float32x2_t __ret;
40944   __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 9);
40945   return __ret;
40946 }
40947 #else
vrndx_f32(float32x2_t __p0)40948 __ai float32x2_t vrndx_f32(float32x2_t __p0) {
40949   float32x2_t __ret;
40950   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40951   __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__rev0, 9);
40952   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40953   return __ret;
40954 }
40955 #endif
40956 
40957 #ifdef __LITTLE_ENDIAN__
vrndq_f16(float16x8_t __p0)40958 __ai __attribute__((target("fullfp16"))) float16x8_t vrndq_f16(float16x8_t __p0) {
40959   float16x8_t __ret;
40960   __ret = (float16x8_t) __builtin_neon_vrndq_f16((int8x16_t)__p0, 40);
40961   return __ret;
40962 }
40963 #else
vrndq_f16(float16x8_t __p0)40964 __ai __attribute__((target("fullfp16"))) float16x8_t vrndq_f16(float16x8_t __p0) {
40965   float16x8_t __ret;
40966   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40967   __ret = (float16x8_t) __builtin_neon_vrndq_f16((int8x16_t)__rev0, 40);
40968   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40969   return __ret;
40970 }
40971 #endif
40972 
40973 #ifdef __LITTLE_ENDIAN__
vrnd_f16(float16x4_t __p0)40974 __ai __attribute__((target("fullfp16"))) float16x4_t vrnd_f16(float16x4_t __p0) {
40975   float16x4_t __ret;
40976   __ret = (float16x4_t) __builtin_neon_vrnd_f16((int8x8_t)__p0, 8);
40977   return __ret;
40978 }
40979 #else
vrnd_f16(float16x4_t __p0)40980 __ai __attribute__((target("fullfp16"))) float16x4_t vrnd_f16(float16x4_t __p0) {
40981   float16x4_t __ret;
40982   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40983   __ret = (float16x4_t) __builtin_neon_vrnd_f16((int8x8_t)__rev0, 8);
40984   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40985   return __ret;
40986 }
40987 #endif
40988 
40989 #ifdef __LITTLE_ENDIAN__
vrndaq_f16(float16x8_t __p0)40990 __ai __attribute__((target("fullfp16"))) float16x8_t vrndaq_f16(float16x8_t __p0) {
40991   float16x8_t __ret;
40992   __ret = (float16x8_t) __builtin_neon_vrndaq_f16((int8x16_t)__p0, 40);
40993   return __ret;
40994 }
40995 #else
vrndaq_f16(float16x8_t __p0)40996 __ai __attribute__((target("fullfp16"))) float16x8_t vrndaq_f16(float16x8_t __p0) {
40997   float16x8_t __ret;
40998   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40999   __ret = (float16x8_t) __builtin_neon_vrndaq_f16((int8x16_t)__rev0, 40);
41000   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41001   return __ret;
41002 }
41003 #endif
41004 
41005 #ifdef __LITTLE_ENDIAN__
vrnda_f16(float16x4_t __p0)41006 __ai __attribute__((target("fullfp16"))) float16x4_t vrnda_f16(float16x4_t __p0) {
41007   float16x4_t __ret;
41008   __ret = (float16x4_t) __builtin_neon_vrnda_f16((int8x8_t)__p0, 8);
41009   return __ret;
41010 }
41011 #else
vrnda_f16(float16x4_t __p0)41012 __ai __attribute__((target("fullfp16"))) float16x4_t vrnda_f16(float16x4_t __p0) {
41013   float16x4_t __ret;
41014   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41015   __ret = (float16x4_t) __builtin_neon_vrnda_f16((int8x8_t)__rev0, 8);
41016   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41017   return __ret;
41018 }
41019 #endif
41020 
41021 #ifdef __LITTLE_ENDIAN__
vrndmq_f16(float16x8_t __p0)41022 __ai __attribute__((target("fullfp16"))) float16x8_t vrndmq_f16(float16x8_t __p0) {
41023   float16x8_t __ret;
41024   __ret = (float16x8_t) __builtin_neon_vrndmq_f16((int8x16_t)__p0, 40);
41025   return __ret;
41026 }
41027 #else
vrndmq_f16(float16x8_t __p0)41028 __ai __attribute__((target("fullfp16"))) float16x8_t vrndmq_f16(float16x8_t __p0) {
41029   float16x8_t __ret;
41030   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41031   __ret = (float16x8_t) __builtin_neon_vrndmq_f16((int8x16_t)__rev0, 40);
41032   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41033   return __ret;
41034 }
41035 #endif
41036 
41037 #ifdef __LITTLE_ENDIAN__
vrndm_f16(float16x4_t __p0)41038 __ai __attribute__((target("fullfp16"))) float16x4_t vrndm_f16(float16x4_t __p0) {
41039   float16x4_t __ret;
41040   __ret = (float16x4_t) __builtin_neon_vrndm_f16((int8x8_t)__p0, 8);
41041   return __ret;
41042 }
41043 #else
vrndm_f16(float16x4_t __p0)41044 __ai __attribute__((target("fullfp16"))) float16x4_t vrndm_f16(float16x4_t __p0) {
41045   float16x4_t __ret;
41046   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41047   __ret = (float16x4_t) __builtin_neon_vrndm_f16((int8x8_t)__rev0, 8);
41048   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41049   return __ret;
41050 }
41051 #endif
41052 
41053 #ifdef __LITTLE_ENDIAN__
vrndnq_f16(float16x8_t __p0)41054 __ai __attribute__((target("fullfp16"))) float16x8_t vrndnq_f16(float16x8_t __p0) {
41055   float16x8_t __ret;
41056   __ret = (float16x8_t) __builtin_neon_vrndnq_f16((int8x16_t)__p0, 40);
41057   return __ret;
41058 }
41059 #else
vrndnq_f16(float16x8_t __p0)41060 __ai __attribute__((target("fullfp16"))) float16x8_t vrndnq_f16(float16x8_t __p0) {
41061   float16x8_t __ret;
41062   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41063   __ret = (float16x8_t) __builtin_neon_vrndnq_f16((int8x16_t)__rev0, 40);
41064   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41065   return __ret;
41066 }
41067 #endif
41068 
41069 #ifdef __LITTLE_ENDIAN__
vrndn_f16(float16x4_t __p0)41070 __ai __attribute__((target("fullfp16"))) float16x4_t vrndn_f16(float16x4_t __p0) {
41071   float16x4_t __ret;
41072   __ret = (float16x4_t) __builtin_neon_vrndn_f16((int8x8_t)__p0, 8);
41073   return __ret;
41074 }
41075 #else
vrndn_f16(float16x4_t __p0)41076 __ai __attribute__((target("fullfp16"))) float16x4_t vrndn_f16(float16x4_t __p0) {
41077   float16x4_t __ret;
41078   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41079   __ret = (float16x4_t) __builtin_neon_vrndn_f16((int8x8_t)__rev0, 8);
41080   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41081   return __ret;
41082 }
41083 #endif
41084 
41085 #ifdef __LITTLE_ENDIAN__
vrndpq_f16(float16x8_t __p0)41086 __ai __attribute__((target("fullfp16"))) float16x8_t vrndpq_f16(float16x8_t __p0) {
41087   float16x8_t __ret;
41088   __ret = (float16x8_t) __builtin_neon_vrndpq_f16((int8x16_t)__p0, 40);
41089   return __ret;
41090 }
41091 #else
vrndpq_f16(float16x8_t __p0)41092 __ai __attribute__((target("fullfp16"))) float16x8_t vrndpq_f16(float16x8_t __p0) {
41093   float16x8_t __ret;
41094   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41095   __ret = (float16x8_t) __builtin_neon_vrndpq_f16((int8x16_t)__rev0, 40);
41096   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41097   return __ret;
41098 }
41099 #endif
41100 
41101 #ifdef __LITTLE_ENDIAN__
vrndp_f16(float16x4_t __p0)41102 __ai __attribute__((target("fullfp16"))) float16x4_t vrndp_f16(float16x4_t __p0) {
41103   float16x4_t __ret;
41104   __ret = (float16x4_t) __builtin_neon_vrndp_f16((int8x8_t)__p0, 8);
41105   return __ret;
41106 }
41107 #else
vrndp_f16(float16x4_t __p0)41108 __ai __attribute__((target("fullfp16"))) float16x4_t vrndp_f16(float16x4_t __p0) {
41109   float16x4_t __ret;
41110   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41111   __ret = (float16x4_t) __builtin_neon_vrndp_f16((int8x8_t)__rev0, 8);
41112   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41113   return __ret;
41114 }
41115 #endif
41116 
41117 #ifdef __LITTLE_ENDIAN__
vrndxq_f16(float16x8_t __p0)41118 __ai __attribute__((target("fullfp16"))) float16x8_t vrndxq_f16(float16x8_t __p0) {
41119   float16x8_t __ret;
41120   __ret = (float16x8_t) __builtin_neon_vrndxq_f16((int8x16_t)__p0, 40);
41121   return __ret;
41122 }
41123 #else
vrndxq_f16(float16x8_t __p0)41124 __ai __attribute__((target("fullfp16"))) float16x8_t vrndxq_f16(float16x8_t __p0) {
41125   float16x8_t __ret;
41126   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41127   __ret = (float16x8_t) __builtin_neon_vrndxq_f16((int8x16_t)__rev0, 40);
41128   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41129   return __ret;
41130 }
41131 #endif
41132 
41133 #ifdef __LITTLE_ENDIAN__
vrndx_f16(float16x4_t __p0)41134 __ai __attribute__((target("fullfp16"))) float16x4_t vrndx_f16(float16x4_t __p0) {
41135   float16x4_t __ret;
41136   __ret = (float16x4_t) __builtin_neon_vrndx_f16((int8x8_t)__p0, 8);
41137   return __ret;
41138 }
41139 #else
vrndx_f16(float16x4_t __p0)41140 __ai __attribute__((target("fullfp16"))) float16x4_t vrndx_f16(float16x4_t __p0) {
41141   float16x4_t __ret;
41142   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41143   __ret = (float16x4_t) __builtin_neon_vrndx_f16((int8x8_t)__rev0, 8);
41144   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41145   return __ret;
41146 }
41147 #endif
41148 
41149 #endif
41150 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
41151 #ifdef __LITTLE_ENDIAN__
vmaxnmq_f32(float32x4_t __p0,float32x4_t __p1)41152 __ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
41153   float32x4_t __ret;
41154   __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
41155   return __ret;
41156 }
41157 #else
vmaxnmq_f32(float32x4_t __p0,float32x4_t __p1)41158 __ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
41159   float32x4_t __ret;
41160   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41161   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41162   __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
41163   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41164   return __ret;
41165 }
41166 #endif
41167 
41168 #ifdef __LITTLE_ENDIAN__
vmaxnm_f32(float32x2_t __p0,float32x2_t __p1)41169 __ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
41170   float32x2_t __ret;
41171   __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
41172   return __ret;
41173 }
41174 #else
vmaxnm_f32(float32x2_t __p0,float32x2_t __p1)41175 __ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
41176   float32x2_t __ret;
41177   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41178   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41179   __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
41180   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41181   return __ret;
41182 }
41183 #endif
41184 
41185 #ifdef __LITTLE_ENDIAN__
vminnmq_f32(float32x4_t __p0,float32x4_t __p1)41186 __ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
41187   float32x4_t __ret;
41188   __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
41189   return __ret;
41190 }
41191 #else
vminnmq_f32(float32x4_t __p0,float32x4_t __p1)41192 __ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
41193   float32x4_t __ret;
41194   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41195   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41196   __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
41197   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41198   return __ret;
41199 }
41200 #endif
41201 
41202 #ifdef __LITTLE_ENDIAN__
vminnm_f32(float32x2_t __p0,float32x2_t __p1)41203 __ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
41204   float32x2_t __ret;
41205   __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
41206   return __ret;
41207 }
41208 #else
vminnm_f32(float32x2_t __p0,float32x2_t __p1)41209 __ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
41210   float32x2_t __ret;
41211   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41212   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41213   __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
41214   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41215   return __ret;
41216 }
41217 #endif
41218 
41219 #ifdef __LITTLE_ENDIAN__
vmaxnmq_f16(float16x8_t __p0,float16x8_t __p1)41220 __ai __attribute__((target("fullfp16"))) float16x8_t vmaxnmq_f16(float16x8_t __p0, float16x8_t __p1) {
41221   float16x8_t __ret;
41222   __ret = (float16x8_t) __builtin_neon_vmaxnmq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
41223   return __ret;
41224 }
41225 #else
vmaxnmq_f16(float16x8_t __p0,float16x8_t __p1)41226 __ai __attribute__((target("fullfp16"))) float16x8_t vmaxnmq_f16(float16x8_t __p0, float16x8_t __p1) {
41227   float16x8_t __ret;
41228   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41229   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41230   __ret = (float16x8_t) __builtin_neon_vmaxnmq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
41231   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41232   return __ret;
41233 }
41234 #endif
41235 
41236 #ifdef __LITTLE_ENDIAN__
vmaxnm_f16(float16x4_t __p0,float16x4_t __p1)41237 __ai __attribute__((target("fullfp16"))) float16x4_t vmaxnm_f16(float16x4_t __p0, float16x4_t __p1) {
41238   float16x4_t __ret;
41239   __ret = (float16x4_t) __builtin_neon_vmaxnm_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
41240   return __ret;
41241 }
41242 #else
vmaxnm_f16(float16x4_t __p0,float16x4_t __p1)41243 __ai __attribute__((target("fullfp16"))) float16x4_t vmaxnm_f16(float16x4_t __p0, float16x4_t __p1) {
41244   float16x4_t __ret;
41245   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41246   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41247   __ret = (float16x4_t) __builtin_neon_vmaxnm_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
41248   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41249   return __ret;
41250 }
41251 #endif
41252 
41253 #ifdef __LITTLE_ENDIAN__
vminnmq_f16(float16x8_t __p0,float16x8_t __p1)41254 __ai __attribute__((target("fullfp16"))) float16x8_t vminnmq_f16(float16x8_t __p0, float16x8_t __p1) {
41255   float16x8_t __ret;
41256   __ret = (float16x8_t) __builtin_neon_vminnmq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
41257   return __ret;
41258 }
41259 #else
vminnmq_f16(float16x8_t __p0,float16x8_t __p1)41260 __ai __attribute__((target("fullfp16"))) float16x8_t vminnmq_f16(float16x8_t __p0, float16x8_t __p1) {
41261   float16x8_t __ret;
41262   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41263   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41264   __ret = (float16x8_t) __builtin_neon_vminnmq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
41265   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41266   return __ret;
41267 }
41268 #endif
41269 
41270 #ifdef __LITTLE_ENDIAN__
vminnm_f16(float16x4_t __p0,float16x4_t __p1)41271 __ai __attribute__((target("fullfp16"))) float16x4_t vminnm_f16(float16x4_t __p0, float16x4_t __p1) {
41272   float16x4_t __ret;
41273   __ret = (float16x4_t) __builtin_neon_vminnm_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
41274   return __ret;
41275 }
41276 #else
vminnm_f16(float16x4_t __p0,float16x4_t __p1)41277 __ai __attribute__((target("fullfp16"))) float16x4_t vminnm_f16(float16x4_t __p0, float16x4_t __p1) {
41278   float16x4_t __ret;
41279   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41280   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41281   __ret = (float16x4_t) __builtin_neon_vminnm_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
41282   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41283   return __ret;
41284 }
41285 #endif
41286 
41287 #endif
41288 #if defined(__ARM_FEATURE_FMA)
41289 #ifdef __LITTLE_ENDIAN__
vfmaq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)41290 __ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
41291   float32x4_t __ret;
41292   __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
41293   return __ret;
41294 }
41295 #else
vfmaq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)41296 __ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
41297   float32x4_t __ret;
41298   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41299   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41300   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41301   __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
41302   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41303   return __ret;
41304 }
__noswap_vfmaq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)41305 __ai float32x4_t __noswap_vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
41306   float32x4_t __ret;
41307   __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
41308   return __ret;
41309 }
41310 #endif
41311 
41312 #ifdef __LITTLE_ENDIAN__
vfma_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)41313 __ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
41314   float32x2_t __ret;
41315   __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41316   return __ret;
41317 }
41318 #else
vfma_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)41319 __ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
41320   float32x2_t __ret;
41321   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41322   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41323   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41324   __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
41325   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41326   return __ret;
41327 }
__noswap_vfma_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)41328 __ai float32x2_t __noswap_vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
41329   float32x2_t __ret;
41330   __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41331   return __ret;
41332 }
41333 #endif
41334 
41335 #ifdef __LITTLE_ENDIAN__
vfmaq_n_f32(float32x4_t __p0,float32x4_t __p1,float32_t __p2)41336 __ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
41337   float32x4_t __ret;
41338   __ret = vfmaq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
41339   return __ret;
41340 }
41341 #else
vfmaq_n_f32(float32x4_t __p0,float32x4_t __p1,float32_t __p2)41342 __ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
41343   float32x4_t __ret;
41344   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41345   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41346   __ret = __noswap_vfmaq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
41347   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41348   return __ret;
41349 }
41350 #endif
41351 
41352 #ifdef __LITTLE_ENDIAN__
vfma_n_f32(float32x2_t __p0,float32x2_t __p1,float32_t __p2)41353 __ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
41354   float32x2_t __ret;
41355   __ret = vfma_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
41356   return __ret;
41357 }
41358 #else
vfma_n_f32(float32x2_t __p0,float32x2_t __p1,float32_t __p2)41359 __ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
41360   float32x2_t __ret;
41361   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41362   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41363   __ret = __noswap_vfma_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
41364   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41365   return __ret;
41366 }
41367 #endif
41368 
41369 #ifdef __LITTLE_ENDIAN__
vfmsq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)41370 __ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
41371   float32x4_t __ret;
41372   __ret = vfmaq_f32(__p0, -__p1, __p2);
41373   return __ret;
41374 }
41375 #else
vfmsq_f32(float32x4_t __p0,float32x4_t __p1,float32x4_t __p2)41376 __ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
41377   float32x4_t __ret;
41378   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41379   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41380   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41381   __ret = __noswap_vfmaq_f32(__rev0, -__rev1, __rev2);
41382   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41383   return __ret;
41384 }
41385 #endif
41386 
41387 #ifdef __LITTLE_ENDIAN__
vfms_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)41388 __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
41389   float32x2_t __ret;
41390   __ret = vfma_f32(__p0, -__p1, __p2);
41391   return __ret;
41392 }
41393 #else
vfms_f32(float32x2_t __p0,float32x2_t __p1,float32x2_t __p2)41394 __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
41395   float32x2_t __ret;
41396   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41397   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41398   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41399   __ret = __noswap_vfma_f32(__rev0, -__rev1, __rev2);
41400   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41401   return __ret;
41402 }
41403 #endif
41404 
41405 #endif
41406 #if defined(__aarch64__)
41407 #ifdef __LITTLE_ENDIAN__
vabdq_f64(float64x2_t __p0,float64x2_t __p1)41408 __ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
41409   float64x2_t __ret;
41410   __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
41411   return __ret;
41412 }
41413 #else
vabdq_f64(float64x2_t __p0,float64x2_t __p1)41414 __ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
41415   float64x2_t __ret;
41416   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41417   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41418   __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
41419   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41420   return __ret;
41421 }
41422 #endif
41423 
vabd_f64(float64x1_t __p0,float64x1_t __p1)41424 __ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
41425   float64x1_t __ret;
41426   __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
41427   return __ret;
41428 }
vabdd_f64(float64_t __p0,float64_t __p1)41429 __ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
41430   float64_t __ret;
41431   __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
41432   return __ret;
41433 }
vabds_f32(float32_t __p0,float32_t __p1)41434 __ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
41435   float32_t __ret;
41436   __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
41437   return __ret;
41438 }
41439 #ifdef __LITTLE_ENDIAN__
vabsq_f64(float64x2_t __p0)41440 __ai float64x2_t vabsq_f64(float64x2_t __p0) {
41441   float64x2_t __ret;
41442   __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 42);
41443   return __ret;
41444 }
41445 #else
vabsq_f64(float64x2_t __p0)41446 __ai float64x2_t vabsq_f64(float64x2_t __p0) {
41447   float64x2_t __ret;
41448   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41449   __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 42);
41450   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41451   return __ret;
41452 }
41453 #endif
41454 
41455 #ifdef __LITTLE_ENDIAN__
vabsq_s64(int64x2_t __p0)41456 __ai int64x2_t vabsq_s64(int64x2_t __p0) {
41457   int64x2_t __ret;
41458   __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 35);
41459   return __ret;
41460 }
41461 #else
vabsq_s64(int64x2_t __p0)41462 __ai int64x2_t vabsq_s64(int64x2_t __p0) {
41463   int64x2_t __ret;
41464   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41465   __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 35);
41466   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41467   return __ret;
41468 }
41469 #endif
41470 
vabs_f64(float64x1_t __p0)41471 __ai float64x1_t vabs_f64(float64x1_t __p0) {
41472   float64x1_t __ret;
41473   __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
41474   return __ret;
41475 }
vabs_s64(int64x1_t __p0)41476 __ai int64x1_t vabs_s64(int64x1_t __p0) {
41477   int64x1_t __ret;
41478   __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
41479   return __ret;
41480 }
vabsd_s64(int64_t __p0)41481 __ai int64_t vabsd_s64(int64_t __p0) {
41482   int64_t __ret;
41483   __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
41484   return __ret;
41485 }
41486 #ifdef __LITTLE_ENDIAN__
vaddq_f64(float64x2_t __p0,float64x2_t __p1)41487 __ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
41488   float64x2_t __ret;
41489   __ret = __p0 + __p1;
41490   return __ret;
41491 }
41492 #else
vaddq_f64(float64x2_t __p0,float64x2_t __p1)41493 __ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
41494   float64x2_t __ret;
41495   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41496   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41497   __ret = __rev0 + __rev1;
41498   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41499   return __ret;
41500 }
41501 #endif
41502 
vadd_f64(float64x1_t __p0,float64x1_t __p1)41503 __ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
41504   float64x1_t __ret;
41505   __ret = __p0 + __p1;
41506   return __ret;
41507 }
vaddd_u64(uint64_t __p0,uint64_t __p1)41508 __ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
41509   uint64_t __ret;
41510   __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
41511   return __ret;
41512 }
vaddd_s64(int64_t __p0,int64_t __p1)41513 __ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
41514   int64_t __ret;
41515   __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
41516   return __ret;
41517 }
vaddq_p128(poly128_t __p0,poly128_t __p1)41518 __ai poly128_t vaddq_p128(poly128_t __p0, poly128_t __p1) {
41519   poly128_t __ret;
41520   __ret = (poly128_t) __builtin_neon_vaddq_p128(__p0, __p1);
41521   return __ret;
41522 }
41523 #ifdef __LITTLE_ENDIAN__
vaddhn_high_u32(uint16x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)41524 __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
41525   uint16x8_t __ret;
41526   __ret = vcombine_u16(__p0, vaddhn_u32(__p1, __p2));
41527   return __ret;
41528 }
41529 #else
vaddhn_high_u32(uint16x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)41530 __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
41531   uint16x8_t __ret;
41532   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41533   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41534   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41535   __ret = __noswap_vcombine_u16(__rev0, __noswap_vaddhn_u32(__rev1, __rev2));
41536   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41537   return __ret;
41538 }
41539 #endif
41540 
41541 #ifdef __LITTLE_ENDIAN__
vaddhn_high_u64(uint32x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)41542 __ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
41543   uint32x4_t __ret;
41544   __ret = vcombine_u32(__p0, vaddhn_u64(__p1, __p2));
41545   return __ret;
41546 }
41547 #else
vaddhn_high_u64(uint32x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)41548 __ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
41549   uint32x4_t __ret;
41550   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41551   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41552   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41553   __ret = __noswap_vcombine_u32(__rev0, __noswap_vaddhn_u64(__rev1, __rev2));
41554   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41555   return __ret;
41556 }
41557 #endif
41558 
41559 #ifdef __LITTLE_ENDIAN__
vaddhn_high_u16(uint8x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)41560 __ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
41561   uint8x16_t __ret;
41562   __ret = vcombine_u8(__p0, vaddhn_u16(__p1, __p2));
41563   return __ret;
41564 }
41565 #else
vaddhn_high_u16(uint8x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)41566 __ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
41567   uint8x16_t __ret;
41568   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41569   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41570   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41571   __ret = __noswap_vcombine_u8(__rev0, __noswap_vaddhn_u16(__rev1, __rev2));
41572   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41573   return __ret;
41574 }
41575 #endif
41576 
41577 #ifdef __LITTLE_ENDIAN__
vaddhn_high_s32(int16x4_t __p0,int32x4_t __p1,int32x4_t __p2)41578 __ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
41579   int16x8_t __ret;
41580   __ret = vcombine_s16(__p0, vaddhn_s32(__p1, __p2));
41581   return __ret;
41582 }
41583 #else
vaddhn_high_s32(int16x4_t __p0,int32x4_t __p1,int32x4_t __p2)41584 __ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
41585   int16x8_t __ret;
41586   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41587   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41588   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41589   __ret = __noswap_vcombine_s16(__rev0, __noswap_vaddhn_s32(__rev1, __rev2));
41590   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41591   return __ret;
41592 }
41593 #endif
41594 
41595 #ifdef __LITTLE_ENDIAN__
vaddhn_high_s64(int32x2_t __p0,int64x2_t __p1,int64x2_t __p2)41596 __ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
41597   int32x4_t __ret;
41598   __ret = vcombine_s32(__p0, vaddhn_s64(__p1, __p2));
41599   return __ret;
41600 }
41601 #else
vaddhn_high_s64(int32x2_t __p0,int64x2_t __p1,int64x2_t __p2)41602 __ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
41603   int32x4_t __ret;
41604   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41605   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41606   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41607   __ret = __noswap_vcombine_s32(__rev0, __noswap_vaddhn_s64(__rev1, __rev2));
41608   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41609   return __ret;
41610 }
41611 #endif
41612 
41613 #ifdef __LITTLE_ENDIAN__
vaddhn_high_s16(int8x8_t __p0,int16x8_t __p1,int16x8_t __p2)41614 __ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
41615   int8x16_t __ret;
41616   __ret = vcombine_s8(__p0, vaddhn_s16(__p1, __p2));
41617   return __ret;
41618 }
41619 #else
vaddhn_high_s16(int8x8_t __p0,int16x8_t __p1,int16x8_t __p2)41620 __ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
41621   int8x16_t __ret;
41622   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41623   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41624   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41625   __ret = __noswap_vcombine_s8(__rev0, __noswap_vaddhn_s16(__rev1, __rev2));
41626   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41627   return __ret;
41628 }
41629 #endif
41630 
41631 #ifdef __LITTLE_ENDIAN__
vaddlvq_u8(uint8x16_t __p0)41632 __ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
41633   uint16_t __ret;
41634   __ret = (uint16_t) __builtin_neon_vaddlvq_u8(__p0);
41635   return __ret;
41636 }
41637 #else
vaddlvq_u8(uint8x16_t __p0)41638 __ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
41639   uint16_t __ret;
41640   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41641   __ret = (uint16_t) __builtin_neon_vaddlvq_u8(__rev0);
41642   return __ret;
41643 }
41644 #endif
41645 
41646 #ifdef __LITTLE_ENDIAN__
vaddlvq_u32(uint32x4_t __p0)41647 __ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
41648   uint64_t __ret;
41649   __ret = (uint64_t) __builtin_neon_vaddlvq_u32(__p0);
41650   return __ret;
41651 }
41652 #else
vaddlvq_u32(uint32x4_t __p0)41653 __ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
41654   uint64_t __ret;
41655   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41656   __ret = (uint64_t) __builtin_neon_vaddlvq_u32(__rev0);
41657   return __ret;
41658 }
41659 #endif
41660 
41661 #ifdef __LITTLE_ENDIAN__
vaddlvq_u16(uint16x8_t __p0)41662 __ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
41663   uint32_t __ret;
41664   __ret = (uint32_t) __builtin_neon_vaddlvq_u16(__p0);
41665   return __ret;
41666 }
41667 #else
vaddlvq_u16(uint16x8_t __p0)41668 __ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
41669   uint32_t __ret;
41670   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41671   __ret = (uint32_t) __builtin_neon_vaddlvq_u16(__rev0);
41672   return __ret;
41673 }
41674 #endif
41675 
41676 #ifdef __LITTLE_ENDIAN__
vaddlvq_s8(int8x16_t __p0)41677 __ai int16_t vaddlvq_s8(int8x16_t __p0) {
41678   int16_t __ret;
41679   __ret = (int16_t) __builtin_neon_vaddlvq_s8(__p0);
41680   return __ret;
41681 }
41682 #else
vaddlvq_s8(int8x16_t __p0)41683 __ai int16_t vaddlvq_s8(int8x16_t __p0) {
41684   int16_t __ret;
41685   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41686   __ret = (int16_t) __builtin_neon_vaddlvq_s8(__rev0);
41687   return __ret;
41688 }
41689 #endif
41690 
41691 #ifdef __LITTLE_ENDIAN__
vaddlvq_s32(int32x4_t __p0)41692 __ai int64_t vaddlvq_s32(int32x4_t __p0) {
41693   int64_t __ret;
41694   __ret = (int64_t) __builtin_neon_vaddlvq_s32(__p0);
41695   return __ret;
41696 }
41697 #else
vaddlvq_s32(int32x4_t __p0)41698 __ai int64_t vaddlvq_s32(int32x4_t __p0) {
41699   int64_t __ret;
41700   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41701   __ret = (int64_t) __builtin_neon_vaddlvq_s32(__rev0);
41702   return __ret;
41703 }
41704 #endif
41705 
41706 #ifdef __LITTLE_ENDIAN__
vaddlvq_s16(int16x8_t __p0)41707 __ai int32_t vaddlvq_s16(int16x8_t __p0) {
41708   int32_t __ret;
41709   __ret = (int32_t) __builtin_neon_vaddlvq_s16(__p0);
41710   return __ret;
41711 }
41712 #else
vaddlvq_s16(int16x8_t __p0)41713 __ai int32_t vaddlvq_s16(int16x8_t __p0) {
41714   int32_t __ret;
41715   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41716   __ret = (int32_t) __builtin_neon_vaddlvq_s16(__rev0);
41717   return __ret;
41718 }
41719 #endif
41720 
41721 #ifdef __LITTLE_ENDIAN__
vaddlv_u8(uint8x8_t __p0)41722 __ai uint16_t vaddlv_u8(uint8x8_t __p0) {
41723   uint16_t __ret;
41724   __ret = (uint16_t) __builtin_neon_vaddlv_u8(__p0);
41725   return __ret;
41726 }
41727 #else
vaddlv_u8(uint8x8_t __p0)41728 __ai uint16_t vaddlv_u8(uint8x8_t __p0) {
41729   uint16_t __ret;
41730   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41731   __ret = (uint16_t) __builtin_neon_vaddlv_u8(__rev0);
41732   return __ret;
41733 }
41734 #endif
41735 
41736 #ifdef __LITTLE_ENDIAN__
vaddlv_u32(uint32x2_t __p0)41737 __ai uint64_t vaddlv_u32(uint32x2_t __p0) {
41738   uint64_t __ret;
41739   __ret = (uint64_t) __builtin_neon_vaddlv_u32(__p0);
41740   return __ret;
41741 }
41742 #else
vaddlv_u32(uint32x2_t __p0)41743 __ai uint64_t vaddlv_u32(uint32x2_t __p0) {
41744   uint64_t __ret;
41745   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41746   __ret = (uint64_t) __builtin_neon_vaddlv_u32(__rev0);
41747   return __ret;
41748 }
41749 #endif
41750 
41751 #ifdef __LITTLE_ENDIAN__
vaddlv_u16(uint16x4_t __p0)41752 __ai uint32_t vaddlv_u16(uint16x4_t __p0) {
41753   uint32_t __ret;
41754   __ret = (uint32_t) __builtin_neon_vaddlv_u16(__p0);
41755   return __ret;
41756 }
41757 #else
vaddlv_u16(uint16x4_t __p0)41758 __ai uint32_t vaddlv_u16(uint16x4_t __p0) {
41759   uint32_t __ret;
41760   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41761   __ret = (uint32_t) __builtin_neon_vaddlv_u16(__rev0);
41762   return __ret;
41763 }
41764 #endif
41765 
41766 #ifdef __LITTLE_ENDIAN__
vaddlv_s8(int8x8_t __p0)41767 __ai int16_t vaddlv_s8(int8x8_t __p0) {
41768   int16_t __ret;
41769   __ret = (int16_t) __builtin_neon_vaddlv_s8(__p0);
41770   return __ret;
41771 }
41772 #else
vaddlv_s8(int8x8_t __p0)41773 __ai int16_t vaddlv_s8(int8x8_t __p0) {
41774   int16_t __ret;
41775   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41776   __ret = (int16_t) __builtin_neon_vaddlv_s8(__rev0);
41777   return __ret;
41778 }
41779 #endif
41780 
41781 #ifdef __LITTLE_ENDIAN__
vaddlv_s32(int32x2_t __p0)41782 __ai int64_t vaddlv_s32(int32x2_t __p0) {
41783   int64_t __ret;
41784   __ret = (int64_t) __builtin_neon_vaddlv_s32(__p0);
41785   return __ret;
41786 }
41787 #else
vaddlv_s32(int32x2_t __p0)41788 __ai int64_t vaddlv_s32(int32x2_t __p0) {
41789   int64_t __ret;
41790   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41791   __ret = (int64_t) __builtin_neon_vaddlv_s32(__rev0);
41792   return __ret;
41793 }
41794 #endif
41795 
41796 #ifdef __LITTLE_ENDIAN__
vaddlv_s16(int16x4_t __p0)41797 __ai int32_t vaddlv_s16(int16x4_t __p0) {
41798   int32_t __ret;
41799   __ret = (int32_t) __builtin_neon_vaddlv_s16(__p0);
41800   return __ret;
41801 }
41802 #else
vaddlv_s16(int16x4_t __p0)41803 __ai int32_t vaddlv_s16(int16x4_t __p0) {
41804   int32_t __ret;
41805   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41806   __ret = (int32_t) __builtin_neon_vaddlv_s16(__rev0);
41807   return __ret;
41808 }
41809 #endif
41810 
41811 #ifdef __LITTLE_ENDIAN__
vaddvq_u8(uint8x16_t __p0)41812 __ai uint8_t vaddvq_u8(uint8x16_t __p0) {
41813   uint8_t __ret;
41814   __ret = (uint8_t) __builtin_neon_vaddvq_u8(__p0);
41815   return __ret;
41816 }
41817 #else
vaddvq_u8(uint8x16_t __p0)41818 __ai uint8_t vaddvq_u8(uint8x16_t __p0) {
41819   uint8_t __ret;
41820   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41821   __ret = (uint8_t) __builtin_neon_vaddvq_u8(__rev0);
41822   return __ret;
41823 }
41824 #endif
41825 
41826 #ifdef __LITTLE_ENDIAN__
vaddvq_u32(uint32x4_t __p0)41827 __ai uint32_t vaddvq_u32(uint32x4_t __p0) {
41828   uint32_t __ret;
41829   __ret = (uint32_t) __builtin_neon_vaddvq_u32(__p0);
41830   return __ret;
41831 }
41832 #else
vaddvq_u32(uint32x4_t __p0)41833 __ai uint32_t vaddvq_u32(uint32x4_t __p0) {
41834   uint32_t __ret;
41835   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41836   __ret = (uint32_t) __builtin_neon_vaddvq_u32(__rev0);
41837   return __ret;
41838 }
41839 #endif
41840 
41841 #ifdef __LITTLE_ENDIAN__
vaddvq_u64(uint64x2_t __p0)41842 __ai uint64_t vaddvq_u64(uint64x2_t __p0) {
41843   uint64_t __ret;
41844   __ret = (uint64_t) __builtin_neon_vaddvq_u64(__p0);
41845   return __ret;
41846 }
41847 #else
vaddvq_u64(uint64x2_t __p0)41848 __ai uint64_t vaddvq_u64(uint64x2_t __p0) {
41849   uint64_t __ret;
41850   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41851   __ret = (uint64_t) __builtin_neon_vaddvq_u64(__rev0);
41852   return __ret;
41853 }
41854 #endif
41855 
41856 #ifdef __LITTLE_ENDIAN__
vaddvq_u16(uint16x8_t __p0)41857 __ai uint16_t vaddvq_u16(uint16x8_t __p0) {
41858   uint16_t __ret;
41859   __ret = (uint16_t) __builtin_neon_vaddvq_u16(__p0);
41860   return __ret;
41861 }
41862 #else
vaddvq_u16(uint16x8_t __p0)41863 __ai uint16_t vaddvq_u16(uint16x8_t __p0) {
41864   uint16_t __ret;
41865   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41866   __ret = (uint16_t) __builtin_neon_vaddvq_u16(__rev0);
41867   return __ret;
41868 }
41869 #endif
41870 
41871 #ifdef __LITTLE_ENDIAN__
vaddvq_s8(int8x16_t __p0)41872 __ai int8_t vaddvq_s8(int8x16_t __p0) {
41873   int8_t __ret;
41874   __ret = (int8_t) __builtin_neon_vaddvq_s8(__p0);
41875   return __ret;
41876 }
41877 #else
vaddvq_s8(int8x16_t __p0)41878 __ai int8_t vaddvq_s8(int8x16_t __p0) {
41879   int8_t __ret;
41880   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41881   __ret = (int8_t) __builtin_neon_vaddvq_s8(__rev0);
41882   return __ret;
41883 }
41884 #endif
41885 
41886 #ifdef __LITTLE_ENDIAN__
vaddvq_f64(float64x2_t __p0)41887 __ai float64_t vaddvq_f64(float64x2_t __p0) {
41888   float64_t __ret;
41889   __ret = (float64_t) __builtin_neon_vaddvq_f64(__p0);
41890   return __ret;
41891 }
41892 #else
vaddvq_f64(float64x2_t __p0)41893 __ai float64_t vaddvq_f64(float64x2_t __p0) {
41894   float64_t __ret;
41895   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41896   __ret = (float64_t) __builtin_neon_vaddvq_f64(__rev0);
41897   return __ret;
41898 }
41899 #endif
41900 
41901 #ifdef __LITTLE_ENDIAN__
vaddvq_f32(float32x4_t __p0)41902 __ai float32_t vaddvq_f32(float32x4_t __p0) {
41903   float32_t __ret;
41904   __ret = (float32_t) __builtin_neon_vaddvq_f32(__p0);
41905   return __ret;
41906 }
41907 #else
vaddvq_f32(float32x4_t __p0)41908 __ai float32_t vaddvq_f32(float32x4_t __p0) {
41909   float32_t __ret;
41910   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41911   __ret = (float32_t) __builtin_neon_vaddvq_f32(__rev0);
41912   return __ret;
41913 }
41914 #endif
41915 
41916 #ifdef __LITTLE_ENDIAN__
vaddvq_s32(int32x4_t __p0)41917 __ai int32_t vaddvq_s32(int32x4_t __p0) {
41918   int32_t __ret;
41919   __ret = (int32_t) __builtin_neon_vaddvq_s32(__p0);
41920   return __ret;
41921 }
41922 #else
vaddvq_s32(int32x4_t __p0)41923 __ai int32_t vaddvq_s32(int32x4_t __p0) {
41924   int32_t __ret;
41925   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41926   __ret = (int32_t) __builtin_neon_vaddvq_s32(__rev0);
41927   return __ret;
41928 }
41929 #endif
41930 
41931 #ifdef __LITTLE_ENDIAN__
vaddvq_s64(int64x2_t __p0)41932 __ai int64_t vaddvq_s64(int64x2_t __p0) {
41933   int64_t __ret;
41934   __ret = (int64_t) __builtin_neon_vaddvq_s64(__p0);
41935   return __ret;
41936 }
41937 #else
vaddvq_s64(int64x2_t __p0)41938 __ai int64_t vaddvq_s64(int64x2_t __p0) {
41939   int64_t __ret;
41940   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41941   __ret = (int64_t) __builtin_neon_vaddvq_s64(__rev0);
41942   return __ret;
41943 }
41944 #endif
41945 
41946 #ifdef __LITTLE_ENDIAN__
vaddvq_s16(int16x8_t __p0)41947 __ai int16_t vaddvq_s16(int16x8_t __p0) {
41948   int16_t __ret;
41949   __ret = (int16_t) __builtin_neon_vaddvq_s16(__p0);
41950   return __ret;
41951 }
41952 #else
vaddvq_s16(int16x8_t __p0)41953 __ai int16_t vaddvq_s16(int16x8_t __p0) {
41954   int16_t __ret;
41955   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41956   __ret = (int16_t) __builtin_neon_vaddvq_s16(__rev0);
41957   return __ret;
41958 }
41959 #endif
41960 
41961 #ifdef __LITTLE_ENDIAN__
vaddv_u8(uint8x8_t __p0)41962 __ai uint8_t vaddv_u8(uint8x8_t __p0) {
41963   uint8_t __ret;
41964   __ret = (uint8_t) __builtin_neon_vaddv_u8(__p0);
41965   return __ret;
41966 }
41967 #else
vaddv_u8(uint8x8_t __p0)41968 __ai uint8_t vaddv_u8(uint8x8_t __p0) {
41969   uint8_t __ret;
41970   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41971   __ret = (uint8_t) __builtin_neon_vaddv_u8(__rev0);
41972   return __ret;
41973 }
41974 #endif
41975 
41976 #ifdef __LITTLE_ENDIAN__
vaddv_u32(uint32x2_t __p0)41977 __ai uint32_t vaddv_u32(uint32x2_t __p0) {
41978   uint32_t __ret;
41979   __ret = (uint32_t) __builtin_neon_vaddv_u32(__p0);
41980   return __ret;
41981 }
41982 #else
vaddv_u32(uint32x2_t __p0)41983 __ai uint32_t vaddv_u32(uint32x2_t __p0) {
41984   uint32_t __ret;
41985   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41986   __ret = (uint32_t) __builtin_neon_vaddv_u32(__rev0);
41987   return __ret;
41988 }
41989 #endif
41990 
41991 #ifdef __LITTLE_ENDIAN__
vaddv_u16(uint16x4_t __p0)41992 __ai uint16_t vaddv_u16(uint16x4_t __p0) {
41993   uint16_t __ret;
41994   __ret = (uint16_t) __builtin_neon_vaddv_u16(__p0);
41995   return __ret;
41996 }
41997 #else
vaddv_u16(uint16x4_t __p0)41998 __ai uint16_t vaddv_u16(uint16x4_t __p0) {
41999   uint16_t __ret;
42000   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42001   __ret = (uint16_t) __builtin_neon_vaddv_u16(__rev0);
42002   return __ret;
42003 }
42004 #endif
42005 
42006 #ifdef __LITTLE_ENDIAN__
vaddv_s8(int8x8_t __p0)42007 __ai int8_t vaddv_s8(int8x8_t __p0) {
42008   int8_t __ret;
42009   __ret = (int8_t) __builtin_neon_vaddv_s8(__p0);
42010   return __ret;
42011 }
42012 #else
vaddv_s8(int8x8_t __p0)42013 __ai int8_t vaddv_s8(int8x8_t __p0) {
42014   int8_t __ret;
42015   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42016   __ret = (int8_t) __builtin_neon_vaddv_s8(__rev0);
42017   return __ret;
42018 }
42019 #endif
42020 
42021 #ifdef __LITTLE_ENDIAN__
vaddv_f32(float32x2_t __p0)42022 __ai float32_t vaddv_f32(float32x2_t __p0) {
42023   float32_t __ret;
42024   __ret = (float32_t) __builtin_neon_vaddv_f32(__p0);
42025   return __ret;
42026 }
42027 #else
vaddv_f32(float32x2_t __p0)42028 __ai float32_t vaddv_f32(float32x2_t __p0) {
42029   float32_t __ret;
42030   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42031   __ret = (float32_t) __builtin_neon_vaddv_f32(__rev0);
42032   return __ret;
42033 }
42034 #endif
42035 
42036 #ifdef __LITTLE_ENDIAN__
vaddv_s32(int32x2_t __p0)42037 __ai int32_t vaddv_s32(int32x2_t __p0) {
42038   int32_t __ret;
42039   __ret = (int32_t) __builtin_neon_vaddv_s32(__p0);
42040   return __ret;
42041 }
42042 #else
vaddv_s32(int32x2_t __p0)42043 __ai int32_t vaddv_s32(int32x2_t __p0) {
42044   int32_t __ret;
42045   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42046   __ret = (int32_t) __builtin_neon_vaddv_s32(__rev0);
42047   return __ret;
42048 }
42049 #endif
42050 
42051 #ifdef __LITTLE_ENDIAN__
vaddv_s16(int16x4_t __p0)42052 __ai int16_t vaddv_s16(int16x4_t __p0) {
42053   int16_t __ret;
42054   __ret = (int16_t) __builtin_neon_vaddv_s16(__p0);
42055   return __ret;
42056 }
42057 #else
vaddv_s16(int16x4_t __p0)42058 __ai int16_t vaddv_s16(int16x4_t __p0) {
42059   int16_t __ret;
42060   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42061   __ret = (int16_t) __builtin_neon_vaddv_s16(__rev0);
42062   return __ret;
42063 }
42064 #endif
42065 
vbsl_p64(uint64x1_t __p0,poly64x1_t __p1,poly64x1_t __p2)42066 __ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
42067   poly64x1_t __ret;
42068   __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
42069   return __ret;
42070 }
42071 #ifdef __LITTLE_ENDIAN__
vbslq_p64(uint64x2_t __p0,poly64x2_t __p1,poly64x2_t __p2)42072 __ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
42073   poly64x2_t __ret;
42074   __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 38);
42075   return __ret;
42076 }
42077 #else
vbslq_p64(uint64x2_t __p0,poly64x2_t __p1,poly64x2_t __p2)42078 __ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
42079   poly64x2_t __ret;
42080   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42081   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42082   poly64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
42083   __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 38);
42084   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42085   return __ret;
42086 }
42087 #endif
42088 
42089 #ifdef __LITTLE_ENDIAN__
vbslq_f64(uint64x2_t __p0,float64x2_t __p1,float64x2_t __p2)42090 __ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
42091   float64x2_t __ret;
42092   __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
42093   return __ret;
42094 }
42095 #else
vbslq_f64(uint64x2_t __p0,float64x2_t __p1,float64x2_t __p2)42096 __ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
42097   float64x2_t __ret;
42098   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42099   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42100   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
42101   __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
42102   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42103   return __ret;
42104 }
42105 #endif
42106 
vbsl_f64(uint64x1_t __p0,float64x1_t __p1,float64x1_t __p2)42107 __ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
42108   float64x1_t __ret;
42109   __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
42110   return __ret;
42111 }
42112 #ifdef __LITTLE_ENDIAN__
vcageq_f64(float64x2_t __p0,float64x2_t __p1)42113 __ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
42114   uint64x2_t __ret;
42115   __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
42116   return __ret;
42117 }
42118 #else
vcageq_f64(float64x2_t __p0,float64x2_t __p1)42119 __ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
42120   uint64x2_t __ret;
42121   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42122   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42123   __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
42124   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42125   return __ret;
42126 }
42127 #endif
42128 
vcage_f64(float64x1_t __p0,float64x1_t __p1)42129 __ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
42130   uint64x1_t __ret;
42131   __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
42132   return __ret;
42133 }
vcaged_f64(float64_t __p0,float64_t __p1)42134 __ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
42135   uint64_t __ret;
42136   __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
42137   return __ret;
42138 }
vcages_f32(float32_t __p0,float32_t __p1)42139 __ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
42140   uint32_t __ret;
42141   __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
42142   return __ret;
42143 }
42144 #ifdef __LITTLE_ENDIAN__
vcagtq_f64(float64x2_t __p0,float64x2_t __p1)42145 __ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
42146   uint64x2_t __ret;
42147   __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
42148   return __ret;
42149 }
42150 #else
vcagtq_f64(float64x2_t __p0,float64x2_t __p1)42151 __ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
42152   uint64x2_t __ret;
42153   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42154   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42155   __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
42156   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42157   return __ret;
42158 }
42159 #endif
42160 
vcagt_f64(float64x1_t __p0,float64x1_t __p1)42161 __ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
42162   uint64x1_t __ret;
42163   __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
42164   return __ret;
42165 }
vcagtd_f64(float64_t __p0,float64_t __p1)42166 __ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
42167   uint64_t __ret;
42168   __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
42169   return __ret;
42170 }
vcagts_f32(float32_t __p0,float32_t __p1)42171 __ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
42172   uint32_t __ret;
42173   __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
42174   return __ret;
42175 }
42176 #ifdef __LITTLE_ENDIAN__
vcaleq_f64(float64x2_t __p0,float64x2_t __p1)42177 __ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
42178   uint64x2_t __ret;
42179   __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
42180   return __ret;
42181 }
42182 #else
vcaleq_f64(float64x2_t __p0,float64x2_t __p1)42183 __ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
42184   uint64x2_t __ret;
42185   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42186   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42187   __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
42188   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42189   return __ret;
42190 }
42191 #endif
42192 
vcale_f64(float64x1_t __p0,float64x1_t __p1)42193 __ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
42194   uint64x1_t __ret;
42195   __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
42196   return __ret;
42197 }
vcaled_f64(float64_t __p0,float64_t __p1)42198 __ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
42199   uint64_t __ret;
42200   __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
42201   return __ret;
42202 }
vcales_f32(float32_t __p0,float32_t __p1)42203 __ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
42204   uint32_t __ret;
42205   __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
42206   return __ret;
42207 }
42208 #ifdef __LITTLE_ENDIAN__
vcaltq_f64(float64x2_t __p0,float64x2_t __p1)42209 __ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
42210   uint64x2_t __ret;
42211   __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
42212   return __ret;
42213 }
42214 #else
vcaltq_f64(float64x2_t __p0,float64x2_t __p1)42215 __ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
42216   uint64x2_t __ret;
42217   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42218   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42219   __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
42220   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42221   return __ret;
42222 }
42223 #endif
42224 
vcalt_f64(float64x1_t __p0,float64x1_t __p1)42225 __ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
42226   uint64x1_t __ret;
42227   __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
42228   return __ret;
42229 }
vcaltd_f64(float64_t __p0,float64_t __p1)42230 __ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
42231   uint64_t __ret;
42232   __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
42233   return __ret;
42234 }
vcalts_f32(float32_t __p0,float32_t __p1)42235 __ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
42236   uint32_t __ret;
42237   __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
42238   return __ret;
42239 }
vceq_p64(poly64x1_t __p0,poly64x1_t __p1)42240 __ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
42241   uint64x1_t __ret;
42242   __ret = (uint64x1_t)(__p0 == __p1);
42243   return __ret;
42244 }
42245 #ifdef __LITTLE_ENDIAN__
vceqq_p64(poly64x2_t __p0,poly64x2_t __p1)42246 __ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
42247   uint64x2_t __ret;
42248   __ret = (uint64x2_t)(__p0 == __p1);
42249   return __ret;
42250 }
42251 #else
vceqq_p64(poly64x2_t __p0,poly64x2_t __p1)42252 __ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
42253   uint64x2_t __ret;
42254   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42255   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42256   __ret = (uint64x2_t)(__rev0 == __rev1);
42257   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42258   return __ret;
42259 }
42260 #endif
42261 
42262 #ifdef __LITTLE_ENDIAN__
vceqq_u64(uint64x2_t __p0,uint64x2_t __p1)42263 __ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42264   uint64x2_t __ret;
42265   __ret = (uint64x2_t)(__p0 == __p1);
42266   return __ret;
42267 }
42268 #else
vceqq_u64(uint64x2_t __p0,uint64x2_t __p1)42269 __ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42270   uint64x2_t __ret;
42271   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42272   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42273   __ret = (uint64x2_t)(__rev0 == __rev1);
42274   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42275   return __ret;
42276 }
42277 #endif
42278 
42279 #ifdef __LITTLE_ENDIAN__
vceqq_f64(float64x2_t __p0,float64x2_t __p1)42280 __ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
42281   uint64x2_t __ret;
42282   __ret = (uint64x2_t)(__p0 == __p1);
42283   return __ret;
42284 }
42285 #else
vceqq_f64(float64x2_t __p0,float64x2_t __p1)42286 __ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
42287   uint64x2_t __ret;
42288   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42289   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42290   __ret = (uint64x2_t)(__rev0 == __rev1);
42291   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42292   return __ret;
42293 }
42294 #endif
42295 
42296 #ifdef __LITTLE_ENDIAN__
vceqq_s64(int64x2_t __p0,int64x2_t __p1)42297 __ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
42298   uint64x2_t __ret;
42299   __ret = (uint64x2_t)(__p0 == __p1);
42300   return __ret;
42301 }
42302 #else
vceqq_s64(int64x2_t __p0,int64x2_t __p1)42303 __ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
42304   uint64x2_t __ret;
42305   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42306   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42307   __ret = (uint64x2_t)(__rev0 == __rev1);
42308   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42309   return __ret;
42310 }
42311 #endif
42312 
vceq_u64(uint64x1_t __p0,uint64x1_t __p1)42313 __ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
42314   uint64x1_t __ret;
42315   __ret = (uint64x1_t)(__p0 == __p1);
42316   return __ret;
42317 }
vceq_f64(float64x1_t __p0,float64x1_t __p1)42318 __ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
42319   uint64x1_t __ret;
42320   __ret = (uint64x1_t)(__p0 == __p1);
42321   return __ret;
42322 }
vceq_s64(int64x1_t __p0,int64x1_t __p1)42323 __ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
42324   uint64x1_t __ret;
42325   __ret = (uint64x1_t)(__p0 == __p1);
42326   return __ret;
42327 }
vceqd_u64(uint64_t __p0,uint64_t __p1)42328 __ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
42329   uint64_t __ret;
42330   __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
42331   return __ret;
42332 }
vceqd_s64(int64_t __p0,int64_t __p1)42333 __ai uint64_t vceqd_s64(int64_t __p0, int64_t __p1) {
42334   uint64_t __ret;
42335   __ret = (uint64_t) __builtin_neon_vceqd_s64(__p0, __p1);
42336   return __ret;
42337 }
vceqd_f64(float64_t __p0,float64_t __p1)42338 __ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
42339   uint64_t __ret;
42340   __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
42341   return __ret;
42342 }
vceqs_f32(float32_t __p0,float32_t __p1)42343 __ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
42344   uint32_t __ret;
42345   __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
42346   return __ret;
42347 }
42348 #ifdef __LITTLE_ENDIAN__
vceqz_p8(poly8x8_t __p0)42349 __ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
42350   uint8x8_t __ret;
42351   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42352   return __ret;
42353 }
42354 #else
vceqz_p8(poly8x8_t __p0)42355 __ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
42356   uint8x8_t __ret;
42357   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42358   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42359   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42360   return __ret;
42361 }
42362 #endif
42363 
vceqz_p64(poly64x1_t __p0)42364 __ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
42365   uint64x1_t __ret;
42366   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42367   return __ret;
42368 }
42369 #ifdef __LITTLE_ENDIAN__
vceqzq_p8(poly8x16_t __p0)42370 __ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
42371   uint8x16_t __ret;
42372   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42373   return __ret;
42374 }
42375 #else
vceqzq_p8(poly8x16_t __p0)42376 __ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
42377   uint8x16_t __ret;
42378   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42379   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42380   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42381   return __ret;
42382 }
42383 #endif
42384 
42385 #ifdef __LITTLE_ENDIAN__
vceqzq_p64(poly64x2_t __p0)42386 __ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
42387   uint64x2_t __ret;
42388   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42389   return __ret;
42390 }
42391 #else
vceqzq_p64(poly64x2_t __p0)42392 __ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
42393   uint64x2_t __ret;
42394   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42395   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42396   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42397   return __ret;
42398 }
42399 #endif
42400 
42401 #ifdef __LITTLE_ENDIAN__
vceqzq_u8(uint8x16_t __p0)42402 __ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
42403   uint8x16_t __ret;
42404   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42405   return __ret;
42406 }
42407 #else
vceqzq_u8(uint8x16_t __p0)42408 __ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
42409   uint8x16_t __ret;
42410   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42411   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42412   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42413   return __ret;
42414 }
42415 #endif
42416 
42417 #ifdef __LITTLE_ENDIAN__
vceqzq_u32(uint32x4_t __p0)42418 __ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
42419   uint32x4_t __ret;
42420   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42421   return __ret;
42422 }
42423 #else
vceqzq_u32(uint32x4_t __p0)42424 __ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
42425   uint32x4_t __ret;
42426   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42427   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42428   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42429   return __ret;
42430 }
42431 #endif
42432 
42433 #ifdef __LITTLE_ENDIAN__
vceqzq_u64(uint64x2_t __p0)42434 __ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
42435   uint64x2_t __ret;
42436   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42437   return __ret;
42438 }
42439 #else
vceqzq_u64(uint64x2_t __p0)42440 __ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
42441   uint64x2_t __ret;
42442   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42443   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42444   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42445   return __ret;
42446 }
42447 #endif
42448 
42449 #ifdef __LITTLE_ENDIAN__
vceqzq_u16(uint16x8_t __p0)42450 __ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
42451   uint16x8_t __ret;
42452   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42453   return __ret;
42454 }
42455 #else
vceqzq_u16(uint16x8_t __p0)42456 __ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
42457   uint16x8_t __ret;
42458   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42459   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42460   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42461   return __ret;
42462 }
42463 #endif
42464 
42465 #ifdef __LITTLE_ENDIAN__
vceqzq_s8(int8x16_t __p0)42466 __ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
42467   uint8x16_t __ret;
42468   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42469   return __ret;
42470 }
42471 #else
vceqzq_s8(int8x16_t __p0)42472 __ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
42473   uint8x16_t __ret;
42474   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42475   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42476   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42477   return __ret;
42478 }
42479 #endif
42480 
42481 #ifdef __LITTLE_ENDIAN__
vceqzq_f64(float64x2_t __p0)42482 __ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
42483   uint64x2_t __ret;
42484   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42485   return __ret;
42486 }
42487 #else
vceqzq_f64(float64x2_t __p0)42488 __ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
42489   uint64x2_t __ret;
42490   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42491   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42492   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42493   return __ret;
42494 }
42495 #endif
42496 
42497 #ifdef __LITTLE_ENDIAN__
vceqzq_f32(float32x4_t __p0)42498 __ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
42499   uint32x4_t __ret;
42500   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42501   return __ret;
42502 }
42503 #else
vceqzq_f32(float32x4_t __p0)42504 __ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
42505   uint32x4_t __ret;
42506   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42507   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42508   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42509   return __ret;
42510 }
42511 #endif
42512 
42513 #ifdef __LITTLE_ENDIAN__
vceqzq_s32(int32x4_t __p0)42514 __ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
42515   uint32x4_t __ret;
42516   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42517   return __ret;
42518 }
42519 #else
vceqzq_s32(int32x4_t __p0)42520 __ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
42521   uint32x4_t __ret;
42522   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42523   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42524   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42525   return __ret;
42526 }
42527 #endif
42528 
42529 #ifdef __LITTLE_ENDIAN__
vceqzq_s64(int64x2_t __p0)42530 __ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
42531   uint64x2_t __ret;
42532   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42533   return __ret;
42534 }
42535 #else
vceqzq_s64(int64x2_t __p0)42536 __ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
42537   uint64x2_t __ret;
42538   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42539   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42540   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42541   return __ret;
42542 }
42543 #endif
42544 
42545 #ifdef __LITTLE_ENDIAN__
vceqzq_s16(int16x8_t __p0)42546 __ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
42547   uint16x8_t __ret;
42548   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42549   return __ret;
42550 }
42551 #else
vceqzq_s16(int16x8_t __p0)42552 __ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
42553   uint16x8_t __ret;
42554   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42555   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42556   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42557   return __ret;
42558 }
42559 #endif
42560 
42561 #ifdef __LITTLE_ENDIAN__
vceqz_u8(uint8x8_t __p0)42562 __ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
42563   uint8x8_t __ret;
42564   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42565   return __ret;
42566 }
42567 #else
vceqz_u8(uint8x8_t __p0)42568 __ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
42569   uint8x8_t __ret;
42570   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42571   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42572   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42573   return __ret;
42574 }
42575 #endif
42576 
42577 #ifdef __LITTLE_ENDIAN__
vceqz_u32(uint32x2_t __p0)42578 __ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
42579   uint32x2_t __ret;
42580   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42581   return __ret;
42582 }
42583 #else
vceqz_u32(uint32x2_t __p0)42584 __ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
42585   uint32x2_t __ret;
42586   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42587   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42588   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42589   return __ret;
42590 }
42591 #endif
42592 
vceqz_u64(uint64x1_t __p0)42593 __ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
42594   uint64x1_t __ret;
42595   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42596   return __ret;
42597 }
42598 #ifdef __LITTLE_ENDIAN__
vceqz_u16(uint16x4_t __p0)42599 __ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
42600   uint16x4_t __ret;
42601   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42602   return __ret;
42603 }
42604 #else
vceqz_u16(uint16x4_t __p0)42605 __ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
42606   uint16x4_t __ret;
42607   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42608   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42609   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42610   return __ret;
42611 }
42612 #endif
42613 
42614 #ifdef __LITTLE_ENDIAN__
vceqz_s8(int8x8_t __p0)42615 __ai uint8x8_t vceqz_s8(int8x8_t __p0) {
42616   uint8x8_t __ret;
42617   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42618   return __ret;
42619 }
42620 #else
vceqz_s8(int8x8_t __p0)42621 __ai uint8x8_t vceqz_s8(int8x8_t __p0) {
42622   uint8x8_t __ret;
42623   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42624   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42625   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42626   return __ret;
42627 }
42628 #endif
42629 
vceqz_f64(float64x1_t __p0)42630 __ai uint64x1_t vceqz_f64(float64x1_t __p0) {
42631   uint64x1_t __ret;
42632   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42633   return __ret;
42634 }
42635 #ifdef __LITTLE_ENDIAN__
vceqz_f32(float32x2_t __p0)42636 __ai uint32x2_t vceqz_f32(float32x2_t __p0) {
42637   uint32x2_t __ret;
42638   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42639   return __ret;
42640 }
42641 #else
vceqz_f32(float32x2_t __p0)42642 __ai uint32x2_t vceqz_f32(float32x2_t __p0) {
42643   uint32x2_t __ret;
42644   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42645   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42646   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42647   return __ret;
42648 }
42649 #endif
42650 
42651 #ifdef __LITTLE_ENDIAN__
vceqz_s32(int32x2_t __p0)42652 __ai uint32x2_t vceqz_s32(int32x2_t __p0) {
42653   uint32x2_t __ret;
42654   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42655   return __ret;
42656 }
42657 #else
vceqz_s32(int32x2_t __p0)42658 __ai uint32x2_t vceqz_s32(int32x2_t __p0) {
42659   uint32x2_t __ret;
42660   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42661   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42662   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42663   return __ret;
42664 }
42665 #endif
42666 
vceqz_s64(int64x1_t __p0)42667 __ai uint64x1_t vceqz_s64(int64x1_t __p0) {
42668   uint64x1_t __ret;
42669   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42670   return __ret;
42671 }
42672 #ifdef __LITTLE_ENDIAN__
vceqz_s16(int16x4_t __p0)42673 __ai uint16x4_t vceqz_s16(int16x4_t __p0) {
42674   uint16x4_t __ret;
42675   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42676   return __ret;
42677 }
42678 #else
vceqz_s16(int16x4_t __p0)42679 __ai uint16x4_t vceqz_s16(int16x4_t __p0) {
42680   uint16x4_t __ret;
42681   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42682   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42683   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42684   return __ret;
42685 }
42686 #endif
42687 
vceqzd_u64(uint64_t __p0)42688 __ai uint64_t vceqzd_u64(uint64_t __p0) {
42689   uint64_t __ret;
42690   __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
42691   return __ret;
42692 }
vceqzd_s64(int64_t __p0)42693 __ai uint64_t vceqzd_s64(int64_t __p0) {
42694   uint64_t __ret;
42695   __ret = (uint64_t) __builtin_neon_vceqzd_s64(__p0);
42696   return __ret;
42697 }
vceqzd_f64(float64_t __p0)42698 __ai uint64_t vceqzd_f64(float64_t __p0) {
42699   uint64_t __ret;
42700   __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
42701   return __ret;
42702 }
vceqzs_f32(float32_t __p0)42703 __ai uint32_t vceqzs_f32(float32_t __p0) {
42704   uint32_t __ret;
42705   __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
42706   return __ret;
42707 }
42708 #ifdef __LITTLE_ENDIAN__
vcgeq_u64(uint64x2_t __p0,uint64x2_t __p1)42709 __ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42710   uint64x2_t __ret;
42711   __ret = (uint64x2_t)(__p0 >= __p1);
42712   return __ret;
42713 }
42714 #else
vcgeq_u64(uint64x2_t __p0,uint64x2_t __p1)42715 __ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42716   uint64x2_t __ret;
42717   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42718   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42719   __ret = (uint64x2_t)(__rev0 >= __rev1);
42720   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42721   return __ret;
42722 }
42723 #endif
42724 
42725 #ifdef __LITTLE_ENDIAN__
vcgeq_f64(float64x2_t __p0,float64x2_t __p1)42726 __ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
42727   uint64x2_t __ret;
42728   __ret = (uint64x2_t)(__p0 >= __p1);
42729   return __ret;
42730 }
42731 #else
vcgeq_f64(float64x2_t __p0,float64x2_t __p1)42732 __ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
42733   uint64x2_t __ret;
42734   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42735   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42736   __ret = (uint64x2_t)(__rev0 >= __rev1);
42737   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42738   return __ret;
42739 }
42740 #endif
42741 
42742 #ifdef __LITTLE_ENDIAN__
vcgeq_s64(int64x2_t __p0,int64x2_t __p1)42743 __ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
42744   uint64x2_t __ret;
42745   __ret = (uint64x2_t)(__p0 >= __p1);
42746   return __ret;
42747 }
42748 #else
vcgeq_s64(int64x2_t __p0,int64x2_t __p1)42749 __ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
42750   uint64x2_t __ret;
42751   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42752   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42753   __ret = (uint64x2_t)(__rev0 >= __rev1);
42754   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42755   return __ret;
42756 }
42757 #endif
42758 
vcge_u64(uint64x1_t __p0,uint64x1_t __p1)42759 __ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
42760   uint64x1_t __ret;
42761   __ret = (uint64x1_t)(__p0 >= __p1);
42762   return __ret;
42763 }
vcge_f64(float64x1_t __p0,float64x1_t __p1)42764 __ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
42765   uint64x1_t __ret;
42766   __ret = (uint64x1_t)(__p0 >= __p1);
42767   return __ret;
42768 }
vcge_s64(int64x1_t __p0,int64x1_t __p1)42769 __ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
42770   uint64x1_t __ret;
42771   __ret = (uint64x1_t)(__p0 >= __p1);
42772   return __ret;
42773 }
vcged_s64(int64_t __p0,int64_t __p1)42774 __ai uint64_t vcged_s64(int64_t __p0, int64_t __p1) {
42775   uint64_t __ret;
42776   __ret = (uint64_t) __builtin_neon_vcged_s64(__p0, __p1);
42777   return __ret;
42778 }
vcged_u64(uint64_t __p0,uint64_t __p1)42779 __ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
42780   uint64_t __ret;
42781   __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
42782   return __ret;
42783 }
vcged_f64(float64_t __p0,float64_t __p1)42784 __ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
42785   uint64_t __ret;
42786   __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
42787   return __ret;
42788 }
vcges_f32(float32_t __p0,float32_t __p1)42789 __ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
42790   uint32_t __ret;
42791   __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
42792   return __ret;
42793 }
42794 #ifdef __LITTLE_ENDIAN__
vcgezq_s8(int8x16_t __p0)42795 __ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
42796   uint8x16_t __ret;
42797   __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 48);
42798   return __ret;
42799 }
42800 #else
vcgezq_s8(int8x16_t __p0)42801 __ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
42802   uint8x16_t __ret;
42803   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42804   __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 48);
42805   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42806   return __ret;
42807 }
42808 #endif
42809 
42810 #ifdef __LITTLE_ENDIAN__
vcgezq_f64(float64x2_t __p0)42811 __ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
42812   uint64x2_t __ret;
42813   __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
42814   return __ret;
42815 }
42816 #else
vcgezq_f64(float64x2_t __p0)42817 __ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
42818   uint64x2_t __ret;
42819   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42820   __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
42821   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42822   return __ret;
42823 }
42824 #endif
42825 
42826 #ifdef __LITTLE_ENDIAN__
vcgezq_f32(float32x4_t __p0)42827 __ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
42828   uint32x4_t __ret;
42829   __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
42830   return __ret;
42831 }
42832 #else
vcgezq_f32(float32x4_t __p0)42833 __ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
42834   uint32x4_t __ret;
42835   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42836   __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
42837   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42838   return __ret;
42839 }
42840 #endif
42841 
42842 #ifdef __LITTLE_ENDIAN__
vcgezq_s32(int32x4_t __p0)42843 __ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
42844   uint32x4_t __ret;
42845   __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
42846   return __ret;
42847 }
42848 #else
vcgezq_s32(int32x4_t __p0)42849 __ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
42850   uint32x4_t __ret;
42851   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42852   __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
42853   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42854   return __ret;
42855 }
42856 #endif
42857 
42858 #ifdef __LITTLE_ENDIAN__
vcgezq_s64(int64x2_t __p0)42859 __ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
42860   uint64x2_t __ret;
42861   __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
42862   return __ret;
42863 }
42864 #else
vcgezq_s64(int64x2_t __p0)42865 __ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
42866   uint64x2_t __ret;
42867   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42868   __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
42869   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42870   return __ret;
42871 }
42872 #endif
42873 
42874 #ifdef __LITTLE_ENDIAN__
vcgezq_s16(int16x8_t __p0)42875 __ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
42876   uint16x8_t __ret;
42877   __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 49);
42878   return __ret;
42879 }
42880 #else
vcgezq_s16(int16x8_t __p0)42881 __ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
42882   uint16x8_t __ret;
42883   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42884   __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 49);
42885   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42886   return __ret;
42887 }
42888 #endif
42889 
42890 #ifdef __LITTLE_ENDIAN__
vcgez_s8(int8x8_t __p0)42891 __ai uint8x8_t vcgez_s8(int8x8_t __p0) {
42892   uint8x8_t __ret;
42893   __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 16);
42894   return __ret;
42895 }
42896 #else
vcgez_s8(int8x8_t __p0)42897 __ai uint8x8_t vcgez_s8(int8x8_t __p0) {
42898   uint8x8_t __ret;
42899   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42900   __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 16);
42901   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42902   return __ret;
42903 }
42904 #endif
42905 
vcgez_f64(float64x1_t __p0)42906 __ai uint64x1_t vcgez_f64(float64x1_t __p0) {
42907   uint64x1_t __ret;
42908   __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42909   return __ret;
42910 }
42911 #ifdef __LITTLE_ENDIAN__
vcgez_f32(float32x2_t __p0)42912 __ai uint32x2_t vcgez_f32(float32x2_t __p0) {
42913   uint32x2_t __ret;
42914   __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
42915   return __ret;
42916 }
42917 #else
vcgez_f32(float32x2_t __p0)42918 __ai uint32x2_t vcgez_f32(float32x2_t __p0) {
42919   uint32x2_t __ret;
42920   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42921   __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
42922   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42923   return __ret;
42924 }
42925 #endif
42926 
42927 #ifdef __LITTLE_ENDIAN__
vcgez_s32(int32x2_t __p0)42928 __ai uint32x2_t vcgez_s32(int32x2_t __p0) {
42929   uint32x2_t __ret;
42930   __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
42931   return __ret;
42932 }
42933 #else
vcgez_s32(int32x2_t __p0)42934 __ai uint32x2_t vcgez_s32(int32x2_t __p0) {
42935   uint32x2_t __ret;
42936   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42937   __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
42938   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42939   return __ret;
42940 }
42941 #endif
42942 
vcgez_s64(int64x1_t __p0)42943 __ai uint64x1_t vcgez_s64(int64x1_t __p0) {
42944   uint64x1_t __ret;
42945   __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42946   return __ret;
42947 }
42948 #ifdef __LITTLE_ENDIAN__
vcgez_s16(int16x4_t __p0)42949 __ai uint16x4_t vcgez_s16(int16x4_t __p0) {
42950   uint16x4_t __ret;
42951   __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 17);
42952   return __ret;
42953 }
42954 #else
vcgez_s16(int16x4_t __p0)42955 __ai uint16x4_t vcgez_s16(int16x4_t __p0) {
42956   uint16x4_t __ret;
42957   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42958   __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 17);
42959   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42960   return __ret;
42961 }
42962 #endif
42963 
vcgezd_s64(int64_t __p0)42964 __ai uint64_t vcgezd_s64(int64_t __p0) {
42965   uint64_t __ret;
42966   __ret = (uint64_t) __builtin_neon_vcgezd_s64(__p0);
42967   return __ret;
42968 }
vcgezd_f64(float64_t __p0)42969 __ai uint64_t vcgezd_f64(float64_t __p0) {
42970   uint64_t __ret;
42971   __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
42972   return __ret;
42973 }
vcgezs_f32(float32_t __p0)42974 __ai uint32_t vcgezs_f32(float32_t __p0) {
42975   uint32_t __ret;
42976   __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
42977   return __ret;
42978 }
42979 #ifdef __LITTLE_ENDIAN__
vcgtq_u64(uint64x2_t __p0,uint64x2_t __p1)42980 __ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42981   uint64x2_t __ret;
42982   __ret = (uint64x2_t)(__p0 > __p1);
42983   return __ret;
42984 }
42985 #else
vcgtq_u64(uint64x2_t __p0,uint64x2_t __p1)42986 __ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42987   uint64x2_t __ret;
42988   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42989   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42990   __ret = (uint64x2_t)(__rev0 > __rev1);
42991   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42992   return __ret;
42993 }
42994 #endif
42995 
42996 #ifdef __LITTLE_ENDIAN__
vcgtq_f64(float64x2_t __p0,float64x2_t __p1)42997 __ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
42998   uint64x2_t __ret;
42999   __ret = (uint64x2_t)(__p0 > __p1);
43000   return __ret;
43001 }
43002 #else
vcgtq_f64(float64x2_t __p0,float64x2_t __p1)43003 __ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
43004   uint64x2_t __ret;
43005   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43006   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43007   __ret = (uint64x2_t)(__rev0 > __rev1);
43008   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43009   return __ret;
43010 }
43011 #endif
43012 
43013 #ifdef __LITTLE_ENDIAN__
vcgtq_s64(int64x2_t __p0,int64x2_t __p1)43014 __ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
43015   uint64x2_t __ret;
43016   __ret = (uint64x2_t)(__p0 > __p1);
43017   return __ret;
43018 }
43019 #else
vcgtq_s64(int64x2_t __p0,int64x2_t __p1)43020 __ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
43021   uint64x2_t __ret;
43022   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43023   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43024   __ret = (uint64x2_t)(__rev0 > __rev1);
43025   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43026   return __ret;
43027 }
43028 #endif
43029 
vcgt_u64(uint64x1_t __p0,uint64x1_t __p1)43030 __ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43031   uint64x1_t __ret;
43032   __ret = (uint64x1_t)(__p0 > __p1);
43033   return __ret;
43034 }
vcgt_f64(float64x1_t __p0,float64x1_t __p1)43035 __ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
43036   uint64x1_t __ret;
43037   __ret = (uint64x1_t)(__p0 > __p1);
43038   return __ret;
43039 }
vcgt_s64(int64x1_t __p0,int64x1_t __p1)43040 __ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
43041   uint64x1_t __ret;
43042   __ret = (uint64x1_t)(__p0 > __p1);
43043   return __ret;
43044 }
vcgtd_s64(int64_t __p0,int64_t __p1)43045 __ai uint64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
43046   uint64_t __ret;
43047   __ret = (uint64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
43048   return __ret;
43049 }
vcgtd_u64(uint64_t __p0,uint64_t __p1)43050 __ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
43051   uint64_t __ret;
43052   __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
43053   return __ret;
43054 }
vcgtd_f64(float64_t __p0,float64_t __p1)43055 __ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
43056   uint64_t __ret;
43057   __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
43058   return __ret;
43059 }
vcgts_f32(float32_t __p0,float32_t __p1)43060 __ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
43061   uint32_t __ret;
43062   __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
43063   return __ret;
43064 }
43065 #ifdef __LITTLE_ENDIAN__
vcgtzq_s8(int8x16_t __p0)43066 __ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
43067   uint8x16_t __ret;
43068   __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 48);
43069   return __ret;
43070 }
43071 #else
vcgtzq_s8(int8x16_t __p0)43072 __ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
43073   uint8x16_t __ret;
43074   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43075   __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 48);
43076   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43077   return __ret;
43078 }
43079 #endif
43080 
43081 #ifdef __LITTLE_ENDIAN__
vcgtzq_f64(float64x2_t __p0)43082 __ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
43083   uint64x2_t __ret;
43084   __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
43085   return __ret;
43086 }
43087 #else
vcgtzq_f64(float64x2_t __p0)43088 __ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
43089   uint64x2_t __ret;
43090   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43091   __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
43092   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43093   return __ret;
43094 }
43095 #endif
43096 
43097 #ifdef __LITTLE_ENDIAN__
vcgtzq_f32(float32x4_t __p0)43098 __ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
43099   uint32x4_t __ret;
43100   __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
43101   return __ret;
43102 }
43103 #else
vcgtzq_f32(float32x4_t __p0)43104 __ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
43105   uint32x4_t __ret;
43106   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43107   __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
43108   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43109   return __ret;
43110 }
43111 #endif
43112 
43113 #ifdef __LITTLE_ENDIAN__
vcgtzq_s32(int32x4_t __p0)43114 __ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
43115   uint32x4_t __ret;
43116   __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
43117   return __ret;
43118 }
43119 #else
vcgtzq_s32(int32x4_t __p0)43120 __ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
43121   uint32x4_t __ret;
43122   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43123   __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
43124   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43125   return __ret;
43126 }
43127 #endif
43128 
43129 #ifdef __LITTLE_ENDIAN__
vcgtzq_s64(int64x2_t __p0)43130 __ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
43131   uint64x2_t __ret;
43132   __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
43133   return __ret;
43134 }
43135 #else
vcgtzq_s64(int64x2_t __p0)43136 __ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
43137   uint64x2_t __ret;
43138   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43139   __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
43140   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43141   return __ret;
43142 }
43143 #endif
43144 
43145 #ifdef __LITTLE_ENDIAN__
vcgtzq_s16(int16x8_t __p0)43146 __ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
43147   uint16x8_t __ret;
43148   __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 49);
43149   return __ret;
43150 }
43151 #else
vcgtzq_s16(int16x8_t __p0)43152 __ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
43153   uint16x8_t __ret;
43154   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43155   __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 49);
43156   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43157   return __ret;
43158 }
43159 #endif
43160 
43161 #ifdef __LITTLE_ENDIAN__
vcgtz_s8(int8x8_t __p0)43162 __ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
43163   uint8x8_t __ret;
43164   __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 16);
43165   return __ret;
43166 }
43167 #else
vcgtz_s8(int8x8_t __p0)43168 __ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
43169   uint8x8_t __ret;
43170   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43171   __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 16);
43172   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43173   return __ret;
43174 }
43175 #endif
43176 
vcgtz_f64(float64x1_t __p0)43177 __ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
43178   uint64x1_t __ret;
43179   __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43180   return __ret;
43181 }
43182 #ifdef __LITTLE_ENDIAN__
vcgtz_f32(float32x2_t __p0)43183 __ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
43184   uint32x2_t __ret;
43185   __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
43186   return __ret;
43187 }
43188 #else
vcgtz_f32(float32x2_t __p0)43189 __ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
43190   uint32x2_t __ret;
43191   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43192   __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
43193   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43194   return __ret;
43195 }
43196 #endif
43197 
43198 #ifdef __LITTLE_ENDIAN__
vcgtz_s32(int32x2_t __p0)43199 __ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
43200   uint32x2_t __ret;
43201   __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
43202   return __ret;
43203 }
43204 #else
vcgtz_s32(int32x2_t __p0)43205 __ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
43206   uint32x2_t __ret;
43207   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43208   __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
43209   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43210   return __ret;
43211 }
43212 #endif
43213 
vcgtz_s64(int64x1_t __p0)43214 __ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
43215   uint64x1_t __ret;
43216   __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43217   return __ret;
43218 }
43219 #ifdef __LITTLE_ENDIAN__
vcgtz_s16(int16x4_t __p0)43220 __ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
43221   uint16x4_t __ret;
43222   __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 17);
43223   return __ret;
43224 }
43225 #else
vcgtz_s16(int16x4_t __p0)43226 __ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
43227   uint16x4_t __ret;
43228   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43229   __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 17);
43230   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43231   return __ret;
43232 }
43233 #endif
43234 
vcgtzd_s64(int64_t __p0)43235 __ai uint64_t vcgtzd_s64(int64_t __p0) {
43236   uint64_t __ret;
43237   __ret = (uint64_t) __builtin_neon_vcgtzd_s64(__p0);
43238   return __ret;
43239 }
vcgtzd_f64(float64_t __p0)43240 __ai uint64_t vcgtzd_f64(float64_t __p0) {
43241   uint64_t __ret;
43242   __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
43243   return __ret;
43244 }
vcgtzs_f32(float32_t __p0)43245 __ai uint32_t vcgtzs_f32(float32_t __p0) {
43246   uint32_t __ret;
43247   __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
43248   return __ret;
43249 }
43250 #ifdef __LITTLE_ENDIAN__
vcleq_u64(uint64x2_t __p0,uint64x2_t __p1)43251 __ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43252   uint64x2_t __ret;
43253   __ret = (uint64x2_t)(__p0 <= __p1);
43254   return __ret;
43255 }
43256 #else
vcleq_u64(uint64x2_t __p0,uint64x2_t __p1)43257 __ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43258   uint64x2_t __ret;
43259   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43260   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43261   __ret = (uint64x2_t)(__rev0 <= __rev1);
43262   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43263   return __ret;
43264 }
43265 #endif
43266 
43267 #ifdef __LITTLE_ENDIAN__
vcleq_f64(float64x2_t __p0,float64x2_t __p1)43268 __ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
43269   uint64x2_t __ret;
43270   __ret = (uint64x2_t)(__p0 <= __p1);
43271   return __ret;
43272 }
43273 #else
vcleq_f64(float64x2_t __p0,float64x2_t __p1)43274 __ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
43275   uint64x2_t __ret;
43276   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43277   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43278   __ret = (uint64x2_t)(__rev0 <= __rev1);
43279   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43280   return __ret;
43281 }
43282 #endif
43283 
43284 #ifdef __LITTLE_ENDIAN__
vcleq_s64(int64x2_t __p0,int64x2_t __p1)43285 __ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
43286   uint64x2_t __ret;
43287   __ret = (uint64x2_t)(__p0 <= __p1);
43288   return __ret;
43289 }
43290 #else
vcleq_s64(int64x2_t __p0,int64x2_t __p1)43291 __ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
43292   uint64x2_t __ret;
43293   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43294   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43295   __ret = (uint64x2_t)(__rev0 <= __rev1);
43296   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43297   return __ret;
43298 }
43299 #endif
43300 
vcle_u64(uint64x1_t __p0,uint64x1_t __p1)43301 __ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
43302   uint64x1_t __ret;
43303   __ret = (uint64x1_t)(__p0 <= __p1);
43304   return __ret;
43305 }
vcle_f64(float64x1_t __p0,float64x1_t __p1)43306 __ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
43307   uint64x1_t __ret;
43308   __ret = (uint64x1_t)(__p0 <= __p1);
43309   return __ret;
43310 }
vcle_s64(int64x1_t __p0,int64x1_t __p1)43311 __ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
43312   uint64x1_t __ret;
43313   __ret = (uint64x1_t)(__p0 <= __p1);
43314   return __ret;
43315 }
vcled_u64(uint64_t __p0,uint64_t __p1)43316 __ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
43317   uint64_t __ret;
43318   __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
43319   return __ret;
43320 }
vcled_s64(int64_t __p0,int64_t __p1)43321 __ai uint64_t vcled_s64(int64_t __p0, int64_t __p1) {
43322   uint64_t __ret;
43323   __ret = (uint64_t) __builtin_neon_vcled_s64(__p0, __p1);
43324   return __ret;
43325 }
vcled_f64(float64_t __p0,float64_t __p1)43326 __ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
43327   uint64_t __ret;
43328   __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
43329   return __ret;
43330 }
vcles_f32(float32_t __p0,float32_t __p1)43331 __ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
43332   uint32_t __ret;
43333   __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
43334   return __ret;
43335 }
43336 #ifdef __LITTLE_ENDIAN__
vclezq_s8(int8x16_t __p0)43337 __ai uint8x16_t vclezq_s8(int8x16_t __p0) {
43338   uint8x16_t __ret;
43339   __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 48);
43340   return __ret;
43341 }
43342 #else
vclezq_s8(int8x16_t __p0)43343 __ai uint8x16_t vclezq_s8(int8x16_t __p0) {
43344   uint8x16_t __ret;
43345   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43346   __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 48);
43347   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43348   return __ret;
43349 }
43350 #endif
43351 
43352 #ifdef __LITTLE_ENDIAN__
vclezq_f64(float64x2_t __p0)43353 __ai uint64x2_t vclezq_f64(float64x2_t __p0) {
43354   uint64x2_t __ret;
43355   __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
43356   return __ret;
43357 }
43358 #else
vclezq_f64(float64x2_t __p0)43359 __ai uint64x2_t vclezq_f64(float64x2_t __p0) {
43360   uint64x2_t __ret;
43361   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43362   __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
43363   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43364   return __ret;
43365 }
43366 #endif
43367 
43368 #ifdef __LITTLE_ENDIAN__
vclezq_f32(float32x4_t __p0)43369 __ai uint32x4_t vclezq_f32(float32x4_t __p0) {
43370   uint32x4_t __ret;
43371   __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
43372   return __ret;
43373 }
43374 #else
vclezq_f32(float32x4_t __p0)43375 __ai uint32x4_t vclezq_f32(float32x4_t __p0) {
43376   uint32x4_t __ret;
43377   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43378   __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
43379   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43380   return __ret;
43381 }
43382 #endif
43383 
43384 #ifdef __LITTLE_ENDIAN__
vclezq_s32(int32x4_t __p0)43385 __ai uint32x4_t vclezq_s32(int32x4_t __p0) {
43386   uint32x4_t __ret;
43387   __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
43388   return __ret;
43389 }
43390 #else
vclezq_s32(int32x4_t __p0)43391 __ai uint32x4_t vclezq_s32(int32x4_t __p0) {
43392   uint32x4_t __ret;
43393   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43394   __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
43395   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43396   return __ret;
43397 }
43398 #endif
43399 
43400 #ifdef __LITTLE_ENDIAN__
vclezq_s64(int64x2_t __p0)43401 __ai uint64x2_t vclezq_s64(int64x2_t __p0) {
43402   uint64x2_t __ret;
43403   __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
43404   return __ret;
43405 }
43406 #else
vclezq_s64(int64x2_t __p0)43407 __ai uint64x2_t vclezq_s64(int64x2_t __p0) {
43408   uint64x2_t __ret;
43409   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43410   __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
43411   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43412   return __ret;
43413 }
43414 #endif
43415 
43416 #ifdef __LITTLE_ENDIAN__
vclezq_s16(int16x8_t __p0)43417 __ai uint16x8_t vclezq_s16(int16x8_t __p0) {
43418   uint16x8_t __ret;
43419   __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 49);
43420   return __ret;
43421 }
43422 #else
vclezq_s16(int16x8_t __p0)43423 __ai uint16x8_t vclezq_s16(int16x8_t __p0) {
43424   uint16x8_t __ret;
43425   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43426   __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 49);
43427   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43428   return __ret;
43429 }
43430 #endif
43431 
43432 #ifdef __LITTLE_ENDIAN__
vclez_s8(int8x8_t __p0)43433 __ai uint8x8_t vclez_s8(int8x8_t __p0) {
43434   uint8x8_t __ret;
43435   __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__p0, 16);
43436   return __ret;
43437 }
43438 #else
vclez_s8(int8x8_t __p0)43439 __ai uint8x8_t vclez_s8(int8x8_t __p0) {
43440   uint8x8_t __ret;
43441   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43442   __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 16);
43443   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43444   return __ret;
43445 }
43446 #endif
43447 
vclez_f64(float64x1_t __p0)43448 __ai uint64x1_t vclez_f64(float64x1_t __p0) {
43449   uint64x1_t __ret;
43450   __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43451   return __ret;
43452 }
43453 #ifdef __LITTLE_ENDIAN__
vclez_f32(float32x2_t __p0)43454 __ai uint32x2_t vclez_f32(float32x2_t __p0) {
43455   uint32x2_t __ret;
43456   __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
43457   return __ret;
43458 }
43459 #else
vclez_f32(float32x2_t __p0)43460 __ai uint32x2_t vclez_f32(float32x2_t __p0) {
43461   uint32x2_t __ret;
43462   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43463   __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
43464   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43465   return __ret;
43466 }
43467 #endif
43468 
43469 #ifdef __LITTLE_ENDIAN__
vclez_s32(int32x2_t __p0)43470 __ai uint32x2_t vclez_s32(int32x2_t __p0) {
43471   uint32x2_t __ret;
43472   __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
43473   return __ret;
43474 }
43475 #else
vclez_s32(int32x2_t __p0)43476 __ai uint32x2_t vclez_s32(int32x2_t __p0) {
43477   uint32x2_t __ret;
43478   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43479   __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
43480   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43481   return __ret;
43482 }
43483 #endif
43484 
vclez_s64(int64x1_t __p0)43485 __ai uint64x1_t vclez_s64(int64x1_t __p0) {
43486   uint64x1_t __ret;
43487   __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43488   return __ret;
43489 }
43490 #ifdef __LITTLE_ENDIAN__
vclez_s16(int16x4_t __p0)43491 __ai uint16x4_t vclez_s16(int16x4_t __p0) {
43492   uint16x4_t __ret;
43493   __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__p0, 17);
43494   return __ret;
43495 }
43496 #else
vclez_s16(int16x4_t __p0)43497 __ai uint16x4_t vclez_s16(int16x4_t __p0) {
43498   uint16x4_t __ret;
43499   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43500   __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 17);
43501   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43502   return __ret;
43503 }
43504 #endif
43505 
vclezd_s64(int64_t __p0)43506 __ai uint64_t vclezd_s64(int64_t __p0) {
43507   uint64_t __ret;
43508   __ret = (uint64_t) __builtin_neon_vclezd_s64(__p0);
43509   return __ret;
43510 }
vclezd_f64(float64_t __p0)43511 __ai uint64_t vclezd_f64(float64_t __p0) {
43512   uint64_t __ret;
43513   __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
43514   return __ret;
43515 }
vclezs_f32(float32_t __p0)43516 __ai uint32_t vclezs_f32(float32_t __p0) {
43517   uint32_t __ret;
43518   __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
43519   return __ret;
43520 }
43521 #ifdef __LITTLE_ENDIAN__
vcltq_u64(uint64x2_t __p0,uint64x2_t __p1)43522 __ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43523   uint64x2_t __ret;
43524   __ret = (uint64x2_t)(__p0 < __p1);
43525   return __ret;
43526 }
43527 #else
vcltq_u64(uint64x2_t __p0,uint64x2_t __p1)43528 __ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43529   uint64x2_t __ret;
43530   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43531   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43532   __ret = (uint64x2_t)(__rev0 < __rev1);
43533   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43534   return __ret;
43535 }
43536 #endif
43537 
43538 #ifdef __LITTLE_ENDIAN__
vcltq_f64(float64x2_t __p0,float64x2_t __p1)43539 __ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
43540   uint64x2_t __ret;
43541   __ret = (uint64x2_t)(__p0 < __p1);
43542   return __ret;
43543 }
43544 #else
vcltq_f64(float64x2_t __p0,float64x2_t __p1)43545 __ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
43546   uint64x2_t __ret;
43547   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43548   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43549   __ret = (uint64x2_t)(__rev0 < __rev1);
43550   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43551   return __ret;
43552 }
43553 #endif
43554 
43555 #ifdef __LITTLE_ENDIAN__
vcltq_s64(int64x2_t __p0,int64x2_t __p1)43556 __ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
43557   uint64x2_t __ret;
43558   __ret = (uint64x2_t)(__p0 < __p1);
43559   return __ret;
43560 }
43561 #else
vcltq_s64(int64x2_t __p0,int64x2_t __p1)43562 __ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
43563   uint64x2_t __ret;
43564   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43565   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43566   __ret = (uint64x2_t)(__rev0 < __rev1);
43567   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43568   return __ret;
43569 }
43570 #endif
43571 
vclt_u64(uint64x1_t __p0,uint64x1_t __p1)43572 __ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43573   uint64x1_t __ret;
43574   __ret = (uint64x1_t)(__p0 < __p1);
43575   return __ret;
43576 }
vclt_f64(float64x1_t __p0,float64x1_t __p1)43577 __ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
43578   uint64x1_t __ret;
43579   __ret = (uint64x1_t)(__p0 < __p1);
43580   return __ret;
43581 }
vclt_s64(int64x1_t __p0,int64x1_t __p1)43582 __ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
43583   uint64x1_t __ret;
43584   __ret = (uint64x1_t)(__p0 < __p1);
43585   return __ret;
43586 }
vcltd_u64(uint64_t __p0,uint64_t __p1)43587 __ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
43588   uint64_t __ret;
43589   __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
43590   return __ret;
43591 }
vcltd_s64(int64_t __p0,int64_t __p1)43592 __ai uint64_t vcltd_s64(int64_t __p0, int64_t __p1) {
43593   uint64_t __ret;
43594   __ret = (uint64_t) __builtin_neon_vcltd_s64(__p0, __p1);
43595   return __ret;
43596 }
vcltd_f64(float64_t __p0,float64_t __p1)43597 __ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
43598   uint64_t __ret;
43599   __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
43600   return __ret;
43601 }
vclts_f32(float32_t __p0,float32_t __p1)43602 __ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
43603   uint32_t __ret;
43604   __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
43605   return __ret;
43606 }
43607 #ifdef __LITTLE_ENDIAN__
vcltzq_s8(int8x16_t __p0)43608 __ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
43609   uint8x16_t __ret;
43610   __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 48);
43611   return __ret;
43612 }
43613 #else
vcltzq_s8(int8x16_t __p0)43614 __ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
43615   uint8x16_t __ret;
43616   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43617   __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 48);
43618   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43619   return __ret;
43620 }
43621 #endif
43622 
43623 #ifdef __LITTLE_ENDIAN__
vcltzq_f64(float64x2_t __p0)43624 __ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
43625   uint64x2_t __ret;
43626   __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
43627   return __ret;
43628 }
43629 #else
vcltzq_f64(float64x2_t __p0)43630 __ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
43631   uint64x2_t __ret;
43632   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43633   __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
43634   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43635   return __ret;
43636 }
43637 #endif
43638 
43639 #ifdef __LITTLE_ENDIAN__
vcltzq_f32(float32x4_t __p0)43640 __ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
43641   uint32x4_t __ret;
43642   __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
43643   return __ret;
43644 }
43645 #else
vcltzq_f32(float32x4_t __p0)43646 __ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
43647   uint32x4_t __ret;
43648   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43649   __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
43650   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43651   return __ret;
43652 }
43653 #endif
43654 
43655 #ifdef __LITTLE_ENDIAN__
vcltzq_s32(int32x4_t __p0)43656 __ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
43657   uint32x4_t __ret;
43658   __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
43659   return __ret;
43660 }
43661 #else
vcltzq_s32(int32x4_t __p0)43662 __ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
43663   uint32x4_t __ret;
43664   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43665   __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
43666   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43667   return __ret;
43668 }
43669 #endif
43670 
43671 #ifdef __LITTLE_ENDIAN__
vcltzq_s64(int64x2_t __p0)43672 __ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
43673   uint64x2_t __ret;
43674   __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
43675   return __ret;
43676 }
43677 #else
vcltzq_s64(int64x2_t __p0)43678 __ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
43679   uint64x2_t __ret;
43680   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43681   __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
43682   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43683   return __ret;
43684 }
43685 #endif
43686 
43687 #ifdef __LITTLE_ENDIAN__
vcltzq_s16(int16x8_t __p0)43688 __ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
43689   uint16x8_t __ret;
43690   __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 49);
43691   return __ret;
43692 }
43693 #else
vcltzq_s16(int16x8_t __p0)43694 __ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
43695   uint16x8_t __ret;
43696   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43697   __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 49);
43698   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43699   return __ret;
43700 }
43701 #endif
43702 
43703 #ifdef __LITTLE_ENDIAN__
vcltz_s8(int8x8_t __p0)43704 __ai uint8x8_t vcltz_s8(int8x8_t __p0) {
43705   uint8x8_t __ret;
43706   __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 16);
43707   return __ret;
43708 }
43709 #else
vcltz_s8(int8x8_t __p0)43710 __ai uint8x8_t vcltz_s8(int8x8_t __p0) {
43711   uint8x8_t __ret;
43712   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43713   __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 16);
43714   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43715   return __ret;
43716 }
43717 #endif
43718 
vcltz_f64(float64x1_t __p0)43719 __ai uint64x1_t vcltz_f64(float64x1_t __p0) {
43720   uint64x1_t __ret;
43721   __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
43722   return __ret;
43723 }
43724 #ifdef __LITTLE_ENDIAN__
vcltz_f32(float32x2_t __p0)43725 __ai uint32x2_t vcltz_f32(float32x2_t __p0) {
43726   uint32x2_t __ret;
43727   __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
43728   return __ret;
43729 }
43730 #else
vcltz_f32(float32x2_t __p0)43731 __ai uint32x2_t vcltz_f32(float32x2_t __p0) {
43732   uint32x2_t __ret;
43733   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43734   __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
43735   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43736   return __ret;
43737 }
43738 #endif
43739 
43740 #ifdef __LITTLE_ENDIAN__
vcltz_s32(int32x2_t __p0)43741 __ai uint32x2_t vcltz_s32(int32x2_t __p0) {
43742   uint32x2_t __ret;
43743   __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
43744   return __ret;
43745 }
43746 #else
vcltz_s32(int32x2_t __p0)43747 __ai uint32x2_t vcltz_s32(int32x2_t __p0) {
43748   uint32x2_t __ret;
43749   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43750   __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
43751   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43752   return __ret;
43753 }
43754 #endif
43755 
vcltz_s64(int64x1_t __p0)43756 __ai uint64x1_t vcltz_s64(int64x1_t __p0) {
43757   uint64x1_t __ret;
43758   __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
43759   return __ret;
43760 }
43761 #ifdef __LITTLE_ENDIAN__
vcltz_s16(int16x4_t __p0)43762 __ai uint16x4_t vcltz_s16(int16x4_t __p0) {
43763   uint16x4_t __ret;
43764   __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 17);
43765   return __ret;
43766 }
43767 #else
vcltz_s16(int16x4_t __p0)43768 __ai uint16x4_t vcltz_s16(int16x4_t __p0) {
43769   uint16x4_t __ret;
43770   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43771   __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 17);
43772   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43773   return __ret;
43774 }
43775 #endif
43776 
vcltzd_s64(int64_t __p0)43777 __ai uint64_t vcltzd_s64(int64_t __p0) {
43778   uint64_t __ret;
43779   __ret = (uint64_t) __builtin_neon_vcltzd_s64(__p0);
43780   return __ret;
43781 }
vcltzd_f64(float64_t __p0)43782 __ai uint64_t vcltzd_f64(float64_t __p0) {
43783   uint64_t __ret;
43784   __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
43785   return __ret;
43786 }
vcltzs_f32(float32_t __p0)43787 __ai uint32_t vcltzs_f32(float32_t __p0) {
43788   uint32_t __ret;
43789   __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
43790   return __ret;
43791 }
43792 #ifdef __LITTLE_ENDIAN__
vcombine_p64(poly64x1_t __p0,poly64x1_t __p1)43793 __ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
43794   poly64x2_t __ret;
43795   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
43796   return __ret;
43797 }
43798 #else
vcombine_p64(poly64x1_t __p0,poly64x1_t __p1)43799 __ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
43800   poly64x2_t __ret;
43801   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
43802   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43803   return __ret;
43804 }
43805 #endif
43806 
43807 #ifdef __LITTLE_ENDIAN__
vcombine_f64(float64x1_t __p0,float64x1_t __p1)43808 __ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
43809   float64x2_t __ret;
43810   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
43811   return __ret;
43812 }
43813 #else
vcombine_f64(float64x1_t __p0,float64x1_t __p1)43814 __ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
43815   float64x2_t __ret;
43816   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
43817   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43818   return __ret;
43819 }
43820 #endif
43821 
43822 #ifdef __LITTLE_ENDIAN__
43823 #define vcopyq_lane_p8(__p0_257, __p1_257, __p2_257, __p3_257) __extension__ ({ \
43824   poly8x16_t __ret_257; \
43825   poly8x16_t __s0_257 = __p0_257; \
43826   poly8x8_t __s2_257 = __p2_257; \
43827   __ret_257 = vsetq_lane_p8(vget_lane_p8(__s2_257, __p3_257), __s0_257, __p1_257); \
43828   __ret_257; \
43829 })
43830 #else
43831 #define vcopyq_lane_p8(__p0_258, __p1_258, __p2_258, __p3_258) __extension__ ({ \
43832   poly8x16_t __ret_258; \
43833   poly8x16_t __s0_258 = __p0_258; \
43834   poly8x8_t __s2_258 = __p2_258; \
43835   poly8x16_t __rev0_258;  __rev0_258 = __builtin_shufflevector(__s0_258, __s0_258, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43836   poly8x8_t __rev2_258;  __rev2_258 = __builtin_shufflevector(__s2_258, __s2_258, 7, 6, 5, 4, 3, 2, 1, 0); \
43837   __ret_258 = __noswap_vsetq_lane_p8(__noswap_vget_lane_p8(__rev2_258, __p3_258), __rev0_258, __p1_258); \
43838   __ret_258 = __builtin_shufflevector(__ret_258, __ret_258, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43839   __ret_258; \
43840 })
43841 #endif
43842 
43843 #ifdef __LITTLE_ENDIAN__
43844 #define vcopyq_lane_p16(__p0_259, __p1_259, __p2_259, __p3_259) __extension__ ({ \
43845   poly16x8_t __ret_259; \
43846   poly16x8_t __s0_259 = __p0_259; \
43847   poly16x4_t __s2_259 = __p2_259; \
43848   __ret_259 = vsetq_lane_p16(vget_lane_p16(__s2_259, __p3_259), __s0_259, __p1_259); \
43849   __ret_259; \
43850 })
43851 #else
43852 #define vcopyq_lane_p16(__p0_260, __p1_260, __p2_260, __p3_260) __extension__ ({ \
43853   poly16x8_t __ret_260; \
43854   poly16x8_t __s0_260 = __p0_260; \
43855   poly16x4_t __s2_260 = __p2_260; \
43856   poly16x8_t __rev0_260;  __rev0_260 = __builtin_shufflevector(__s0_260, __s0_260, 7, 6, 5, 4, 3, 2, 1, 0); \
43857   poly16x4_t __rev2_260;  __rev2_260 = __builtin_shufflevector(__s2_260, __s2_260, 3, 2, 1, 0); \
43858   __ret_260 = __noswap_vsetq_lane_p16(__noswap_vget_lane_p16(__rev2_260, __p3_260), __rev0_260, __p1_260); \
43859   __ret_260 = __builtin_shufflevector(__ret_260, __ret_260, 7, 6, 5, 4, 3, 2, 1, 0); \
43860   __ret_260; \
43861 })
43862 #endif
43863 
43864 #ifdef __LITTLE_ENDIAN__
43865 #define vcopyq_lane_u8(__p0_261, __p1_261, __p2_261, __p3_261) __extension__ ({ \
43866   uint8x16_t __ret_261; \
43867   uint8x16_t __s0_261 = __p0_261; \
43868   uint8x8_t __s2_261 = __p2_261; \
43869   __ret_261 = vsetq_lane_u8(vget_lane_u8(__s2_261, __p3_261), __s0_261, __p1_261); \
43870   __ret_261; \
43871 })
43872 #else
43873 #define vcopyq_lane_u8(__p0_262, __p1_262, __p2_262, __p3_262) __extension__ ({ \
43874   uint8x16_t __ret_262; \
43875   uint8x16_t __s0_262 = __p0_262; \
43876   uint8x8_t __s2_262 = __p2_262; \
43877   uint8x16_t __rev0_262;  __rev0_262 = __builtin_shufflevector(__s0_262, __s0_262, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43878   uint8x8_t __rev2_262;  __rev2_262 = __builtin_shufflevector(__s2_262, __s2_262, 7, 6, 5, 4, 3, 2, 1, 0); \
43879   __ret_262 = __noswap_vsetq_lane_u8(__noswap_vget_lane_u8(__rev2_262, __p3_262), __rev0_262, __p1_262); \
43880   __ret_262 = __builtin_shufflevector(__ret_262, __ret_262, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43881   __ret_262; \
43882 })
43883 #endif
43884 
43885 #ifdef __LITTLE_ENDIAN__
43886 #define vcopyq_lane_u32(__p0_263, __p1_263, __p2_263, __p3_263) __extension__ ({ \
43887   uint32x4_t __ret_263; \
43888   uint32x4_t __s0_263 = __p0_263; \
43889   uint32x2_t __s2_263 = __p2_263; \
43890   __ret_263 = vsetq_lane_u32(vget_lane_u32(__s2_263, __p3_263), __s0_263, __p1_263); \
43891   __ret_263; \
43892 })
43893 #else
43894 #define vcopyq_lane_u32(__p0_264, __p1_264, __p2_264, __p3_264) __extension__ ({ \
43895   uint32x4_t __ret_264; \
43896   uint32x4_t __s0_264 = __p0_264; \
43897   uint32x2_t __s2_264 = __p2_264; \
43898   uint32x4_t __rev0_264;  __rev0_264 = __builtin_shufflevector(__s0_264, __s0_264, 3, 2, 1, 0); \
43899   uint32x2_t __rev2_264;  __rev2_264 = __builtin_shufflevector(__s2_264, __s2_264, 1, 0); \
43900   __ret_264 = __noswap_vsetq_lane_u32(__noswap_vget_lane_u32(__rev2_264, __p3_264), __rev0_264, __p1_264); \
43901   __ret_264 = __builtin_shufflevector(__ret_264, __ret_264, 3, 2, 1, 0); \
43902   __ret_264; \
43903 })
43904 #endif
43905 
43906 #ifdef __LITTLE_ENDIAN__
43907 #define vcopyq_lane_u64(__p0_265, __p1_265, __p2_265, __p3_265) __extension__ ({ \
43908   uint64x2_t __ret_265; \
43909   uint64x2_t __s0_265 = __p0_265; \
43910   uint64x1_t __s2_265 = __p2_265; \
43911   __ret_265 = vsetq_lane_u64(vget_lane_u64(__s2_265, __p3_265), __s0_265, __p1_265); \
43912   __ret_265; \
43913 })
43914 #else
43915 #define vcopyq_lane_u64(__p0_266, __p1_266, __p2_266, __p3_266) __extension__ ({ \
43916   uint64x2_t __ret_266; \
43917   uint64x2_t __s0_266 = __p0_266; \
43918   uint64x1_t __s2_266 = __p2_266; \
43919   uint64x2_t __rev0_266;  __rev0_266 = __builtin_shufflevector(__s0_266, __s0_266, 1, 0); \
43920   __ret_266 = __noswap_vsetq_lane_u64(vget_lane_u64(__s2_266, __p3_266), __rev0_266, __p1_266); \
43921   __ret_266 = __builtin_shufflevector(__ret_266, __ret_266, 1, 0); \
43922   __ret_266; \
43923 })
43924 #endif
43925 
43926 #ifdef __LITTLE_ENDIAN__
43927 #define vcopyq_lane_u16(__p0_267, __p1_267, __p2_267, __p3_267) __extension__ ({ \
43928   uint16x8_t __ret_267; \
43929   uint16x8_t __s0_267 = __p0_267; \
43930   uint16x4_t __s2_267 = __p2_267; \
43931   __ret_267 = vsetq_lane_u16(vget_lane_u16(__s2_267, __p3_267), __s0_267, __p1_267); \
43932   __ret_267; \
43933 })
43934 #else
43935 #define vcopyq_lane_u16(__p0_268, __p1_268, __p2_268, __p3_268) __extension__ ({ \
43936   uint16x8_t __ret_268; \
43937   uint16x8_t __s0_268 = __p0_268; \
43938   uint16x4_t __s2_268 = __p2_268; \
43939   uint16x8_t __rev0_268;  __rev0_268 = __builtin_shufflevector(__s0_268, __s0_268, 7, 6, 5, 4, 3, 2, 1, 0); \
43940   uint16x4_t __rev2_268;  __rev2_268 = __builtin_shufflevector(__s2_268, __s2_268, 3, 2, 1, 0); \
43941   __ret_268 = __noswap_vsetq_lane_u16(__noswap_vget_lane_u16(__rev2_268, __p3_268), __rev0_268, __p1_268); \
43942   __ret_268 = __builtin_shufflevector(__ret_268, __ret_268, 7, 6, 5, 4, 3, 2, 1, 0); \
43943   __ret_268; \
43944 })
43945 #endif
43946 
43947 #ifdef __LITTLE_ENDIAN__
43948 #define vcopyq_lane_s8(__p0_269, __p1_269, __p2_269, __p3_269) __extension__ ({ \
43949   int8x16_t __ret_269; \
43950   int8x16_t __s0_269 = __p0_269; \
43951   int8x8_t __s2_269 = __p2_269; \
43952   __ret_269 = vsetq_lane_s8(vget_lane_s8(__s2_269, __p3_269), __s0_269, __p1_269); \
43953   __ret_269; \
43954 })
43955 #else
43956 #define vcopyq_lane_s8(__p0_270, __p1_270, __p2_270, __p3_270) __extension__ ({ \
43957   int8x16_t __ret_270; \
43958   int8x16_t __s0_270 = __p0_270; \
43959   int8x8_t __s2_270 = __p2_270; \
43960   int8x16_t __rev0_270;  __rev0_270 = __builtin_shufflevector(__s0_270, __s0_270, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43961   int8x8_t __rev2_270;  __rev2_270 = __builtin_shufflevector(__s2_270, __s2_270, 7, 6, 5, 4, 3, 2, 1, 0); \
43962   __ret_270 = __noswap_vsetq_lane_s8(__noswap_vget_lane_s8(__rev2_270, __p3_270), __rev0_270, __p1_270); \
43963   __ret_270 = __builtin_shufflevector(__ret_270, __ret_270, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43964   __ret_270; \
43965 })
43966 #endif
43967 
43968 #ifdef __LITTLE_ENDIAN__
43969 #define vcopyq_lane_f32(__p0_271, __p1_271, __p2_271, __p3_271) __extension__ ({ \
43970   float32x4_t __ret_271; \
43971   float32x4_t __s0_271 = __p0_271; \
43972   float32x2_t __s2_271 = __p2_271; \
43973   __ret_271 = vsetq_lane_f32(vget_lane_f32(__s2_271, __p3_271), __s0_271, __p1_271); \
43974   __ret_271; \
43975 })
43976 #else
43977 #define vcopyq_lane_f32(__p0_272, __p1_272, __p2_272, __p3_272) __extension__ ({ \
43978   float32x4_t __ret_272; \
43979   float32x4_t __s0_272 = __p0_272; \
43980   float32x2_t __s2_272 = __p2_272; \
43981   float32x4_t __rev0_272;  __rev0_272 = __builtin_shufflevector(__s0_272, __s0_272, 3, 2, 1, 0); \
43982   float32x2_t __rev2_272;  __rev2_272 = __builtin_shufflevector(__s2_272, __s2_272, 1, 0); \
43983   __ret_272 = __noswap_vsetq_lane_f32(__noswap_vget_lane_f32(__rev2_272, __p3_272), __rev0_272, __p1_272); \
43984   __ret_272 = __builtin_shufflevector(__ret_272, __ret_272, 3, 2, 1, 0); \
43985   __ret_272; \
43986 })
43987 #endif
43988 
43989 #ifdef __LITTLE_ENDIAN__
43990 #define vcopyq_lane_s32(__p0_273, __p1_273, __p2_273, __p3_273) __extension__ ({ \
43991   int32x4_t __ret_273; \
43992   int32x4_t __s0_273 = __p0_273; \
43993   int32x2_t __s2_273 = __p2_273; \
43994   __ret_273 = vsetq_lane_s32(vget_lane_s32(__s2_273, __p3_273), __s0_273, __p1_273); \
43995   __ret_273; \
43996 })
43997 #else
43998 #define vcopyq_lane_s32(__p0_274, __p1_274, __p2_274, __p3_274) __extension__ ({ \
43999   int32x4_t __ret_274; \
44000   int32x4_t __s0_274 = __p0_274; \
44001   int32x2_t __s2_274 = __p2_274; \
44002   int32x4_t __rev0_274;  __rev0_274 = __builtin_shufflevector(__s0_274, __s0_274, 3, 2, 1, 0); \
44003   int32x2_t __rev2_274;  __rev2_274 = __builtin_shufflevector(__s2_274, __s2_274, 1, 0); \
44004   __ret_274 = __noswap_vsetq_lane_s32(__noswap_vget_lane_s32(__rev2_274, __p3_274), __rev0_274, __p1_274); \
44005   __ret_274 = __builtin_shufflevector(__ret_274, __ret_274, 3, 2, 1, 0); \
44006   __ret_274; \
44007 })
44008 #endif
44009 
44010 #ifdef __LITTLE_ENDIAN__
44011 #define vcopyq_lane_s64(__p0_275, __p1_275, __p2_275, __p3_275) __extension__ ({ \
44012   int64x2_t __ret_275; \
44013   int64x2_t __s0_275 = __p0_275; \
44014   int64x1_t __s2_275 = __p2_275; \
44015   __ret_275 = vsetq_lane_s64(vget_lane_s64(__s2_275, __p3_275), __s0_275, __p1_275); \
44016   __ret_275; \
44017 })
44018 #else
44019 #define vcopyq_lane_s64(__p0_276, __p1_276, __p2_276, __p3_276) __extension__ ({ \
44020   int64x2_t __ret_276; \
44021   int64x2_t __s0_276 = __p0_276; \
44022   int64x1_t __s2_276 = __p2_276; \
44023   int64x2_t __rev0_276;  __rev0_276 = __builtin_shufflevector(__s0_276, __s0_276, 1, 0); \
44024   __ret_276 = __noswap_vsetq_lane_s64(vget_lane_s64(__s2_276, __p3_276), __rev0_276, __p1_276); \
44025   __ret_276 = __builtin_shufflevector(__ret_276, __ret_276, 1, 0); \
44026   __ret_276; \
44027 })
44028 #endif
44029 
44030 #ifdef __LITTLE_ENDIAN__
44031 #define vcopyq_lane_s16(__p0_277, __p1_277, __p2_277, __p3_277) __extension__ ({ \
44032   int16x8_t __ret_277; \
44033   int16x8_t __s0_277 = __p0_277; \
44034   int16x4_t __s2_277 = __p2_277; \
44035   __ret_277 = vsetq_lane_s16(vget_lane_s16(__s2_277, __p3_277), __s0_277, __p1_277); \
44036   __ret_277; \
44037 })
44038 #else
44039 #define vcopyq_lane_s16(__p0_278, __p1_278, __p2_278, __p3_278) __extension__ ({ \
44040   int16x8_t __ret_278; \
44041   int16x8_t __s0_278 = __p0_278; \
44042   int16x4_t __s2_278 = __p2_278; \
44043   int16x8_t __rev0_278;  __rev0_278 = __builtin_shufflevector(__s0_278, __s0_278, 7, 6, 5, 4, 3, 2, 1, 0); \
44044   int16x4_t __rev2_278;  __rev2_278 = __builtin_shufflevector(__s2_278, __s2_278, 3, 2, 1, 0); \
44045   __ret_278 = __noswap_vsetq_lane_s16(__noswap_vget_lane_s16(__rev2_278, __p3_278), __rev0_278, __p1_278); \
44046   __ret_278 = __builtin_shufflevector(__ret_278, __ret_278, 7, 6, 5, 4, 3, 2, 1, 0); \
44047   __ret_278; \
44048 })
44049 #endif
44050 
44051 #ifdef __LITTLE_ENDIAN__
44052 #define vcopy_lane_p8(__p0_279, __p1_279, __p2_279, __p3_279) __extension__ ({ \
44053   poly8x8_t __ret_279; \
44054   poly8x8_t __s0_279 = __p0_279; \
44055   poly8x8_t __s2_279 = __p2_279; \
44056   __ret_279 = vset_lane_p8(vget_lane_p8(__s2_279, __p3_279), __s0_279, __p1_279); \
44057   __ret_279; \
44058 })
44059 #else
44060 #define vcopy_lane_p8(__p0_280, __p1_280, __p2_280, __p3_280) __extension__ ({ \
44061   poly8x8_t __ret_280; \
44062   poly8x8_t __s0_280 = __p0_280; \
44063   poly8x8_t __s2_280 = __p2_280; \
44064   poly8x8_t __rev0_280;  __rev0_280 = __builtin_shufflevector(__s0_280, __s0_280, 7, 6, 5, 4, 3, 2, 1, 0); \
44065   poly8x8_t __rev2_280;  __rev2_280 = __builtin_shufflevector(__s2_280, __s2_280, 7, 6, 5, 4, 3, 2, 1, 0); \
44066   __ret_280 = __noswap_vset_lane_p8(__noswap_vget_lane_p8(__rev2_280, __p3_280), __rev0_280, __p1_280); \
44067   __ret_280 = __builtin_shufflevector(__ret_280, __ret_280, 7, 6, 5, 4, 3, 2, 1, 0); \
44068   __ret_280; \
44069 })
44070 #endif
44071 
44072 #ifdef __LITTLE_ENDIAN__
44073 #define vcopy_lane_p16(__p0_281, __p1_281, __p2_281, __p3_281) __extension__ ({ \
44074   poly16x4_t __ret_281; \
44075   poly16x4_t __s0_281 = __p0_281; \
44076   poly16x4_t __s2_281 = __p2_281; \
44077   __ret_281 = vset_lane_p16(vget_lane_p16(__s2_281, __p3_281), __s0_281, __p1_281); \
44078   __ret_281; \
44079 })
44080 #else
44081 #define vcopy_lane_p16(__p0_282, __p1_282, __p2_282, __p3_282) __extension__ ({ \
44082   poly16x4_t __ret_282; \
44083   poly16x4_t __s0_282 = __p0_282; \
44084   poly16x4_t __s2_282 = __p2_282; \
44085   poly16x4_t __rev0_282;  __rev0_282 = __builtin_shufflevector(__s0_282, __s0_282, 3, 2, 1, 0); \
44086   poly16x4_t __rev2_282;  __rev2_282 = __builtin_shufflevector(__s2_282, __s2_282, 3, 2, 1, 0); \
44087   __ret_282 = __noswap_vset_lane_p16(__noswap_vget_lane_p16(__rev2_282, __p3_282), __rev0_282, __p1_282); \
44088   __ret_282 = __builtin_shufflevector(__ret_282, __ret_282, 3, 2, 1, 0); \
44089   __ret_282; \
44090 })
44091 #endif
44092 
44093 #ifdef __LITTLE_ENDIAN__
44094 #define vcopy_lane_u8(__p0_283, __p1_283, __p2_283, __p3_283) __extension__ ({ \
44095   uint8x8_t __ret_283; \
44096   uint8x8_t __s0_283 = __p0_283; \
44097   uint8x8_t __s2_283 = __p2_283; \
44098   __ret_283 = vset_lane_u8(vget_lane_u8(__s2_283, __p3_283), __s0_283, __p1_283); \
44099   __ret_283; \
44100 })
44101 #else
44102 #define vcopy_lane_u8(__p0_284, __p1_284, __p2_284, __p3_284) __extension__ ({ \
44103   uint8x8_t __ret_284; \
44104   uint8x8_t __s0_284 = __p0_284; \
44105   uint8x8_t __s2_284 = __p2_284; \
44106   uint8x8_t __rev0_284;  __rev0_284 = __builtin_shufflevector(__s0_284, __s0_284, 7, 6, 5, 4, 3, 2, 1, 0); \
44107   uint8x8_t __rev2_284;  __rev2_284 = __builtin_shufflevector(__s2_284, __s2_284, 7, 6, 5, 4, 3, 2, 1, 0); \
44108   __ret_284 = __noswap_vset_lane_u8(__noswap_vget_lane_u8(__rev2_284, __p3_284), __rev0_284, __p1_284); \
44109   __ret_284 = __builtin_shufflevector(__ret_284, __ret_284, 7, 6, 5, 4, 3, 2, 1, 0); \
44110   __ret_284; \
44111 })
44112 #endif
44113 
44114 #ifdef __LITTLE_ENDIAN__
44115 #define vcopy_lane_u32(__p0_285, __p1_285, __p2_285, __p3_285) __extension__ ({ \
44116   uint32x2_t __ret_285; \
44117   uint32x2_t __s0_285 = __p0_285; \
44118   uint32x2_t __s2_285 = __p2_285; \
44119   __ret_285 = vset_lane_u32(vget_lane_u32(__s2_285, __p3_285), __s0_285, __p1_285); \
44120   __ret_285; \
44121 })
44122 #else
44123 #define vcopy_lane_u32(__p0_286, __p1_286, __p2_286, __p3_286) __extension__ ({ \
44124   uint32x2_t __ret_286; \
44125   uint32x2_t __s0_286 = __p0_286; \
44126   uint32x2_t __s2_286 = __p2_286; \
44127   uint32x2_t __rev0_286;  __rev0_286 = __builtin_shufflevector(__s0_286, __s0_286, 1, 0); \
44128   uint32x2_t __rev2_286;  __rev2_286 = __builtin_shufflevector(__s2_286, __s2_286, 1, 0); \
44129   __ret_286 = __noswap_vset_lane_u32(__noswap_vget_lane_u32(__rev2_286, __p3_286), __rev0_286, __p1_286); \
44130   __ret_286 = __builtin_shufflevector(__ret_286, __ret_286, 1, 0); \
44131   __ret_286; \
44132 })
44133 #endif
44134 
44135 #define vcopy_lane_u64(__p0_287, __p1_287, __p2_287, __p3_287) __extension__ ({ \
44136   uint64x1_t __ret_287; \
44137   uint64x1_t __s0_287 = __p0_287; \
44138   uint64x1_t __s2_287 = __p2_287; \
44139   __ret_287 = vset_lane_u64(vget_lane_u64(__s2_287, __p3_287), __s0_287, __p1_287); \
44140   __ret_287; \
44141 })
44142 #ifdef __LITTLE_ENDIAN__
44143 #define vcopy_lane_u16(__p0_288, __p1_288, __p2_288, __p3_288) __extension__ ({ \
44144   uint16x4_t __ret_288; \
44145   uint16x4_t __s0_288 = __p0_288; \
44146   uint16x4_t __s2_288 = __p2_288; \
44147   __ret_288 = vset_lane_u16(vget_lane_u16(__s2_288, __p3_288), __s0_288, __p1_288); \
44148   __ret_288; \
44149 })
44150 #else
44151 #define vcopy_lane_u16(__p0_289, __p1_289, __p2_289, __p3_289) __extension__ ({ \
44152   uint16x4_t __ret_289; \
44153   uint16x4_t __s0_289 = __p0_289; \
44154   uint16x4_t __s2_289 = __p2_289; \
44155   uint16x4_t __rev0_289;  __rev0_289 = __builtin_shufflevector(__s0_289, __s0_289, 3, 2, 1, 0); \
44156   uint16x4_t __rev2_289;  __rev2_289 = __builtin_shufflevector(__s2_289, __s2_289, 3, 2, 1, 0); \
44157   __ret_289 = __noswap_vset_lane_u16(__noswap_vget_lane_u16(__rev2_289, __p3_289), __rev0_289, __p1_289); \
44158   __ret_289 = __builtin_shufflevector(__ret_289, __ret_289, 3, 2, 1, 0); \
44159   __ret_289; \
44160 })
44161 #endif
44162 
44163 #ifdef __LITTLE_ENDIAN__
44164 #define vcopy_lane_s8(__p0_290, __p1_290, __p2_290, __p3_290) __extension__ ({ \
44165   int8x8_t __ret_290; \
44166   int8x8_t __s0_290 = __p0_290; \
44167   int8x8_t __s2_290 = __p2_290; \
44168   __ret_290 = vset_lane_s8(vget_lane_s8(__s2_290, __p3_290), __s0_290, __p1_290); \
44169   __ret_290; \
44170 })
44171 #else
44172 #define vcopy_lane_s8(__p0_291, __p1_291, __p2_291, __p3_291) __extension__ ({ \
44173   int8x8_t __ret_291; \
44174   int8x8_t __s0_291 = __p0_291; \
44175   int8x8_t __s2_291 = __p2_291; \
44176   int8x8_t __rev0_291;  __rev0_291 = __builtin_shufflevector(__s0_291, __s0_291, 7, 6, 5, 4, 3, 2, 1, 0); \
44177   int8x8_t __rev2_291;  __rev2_291 = __builtin_shufflevector(__s2_291, __s2_291, 7, 6, 5, 4, 3, 2, 1, 0); \
44178   __ret_291 = __noswap_vset_lane_s8(__noswap_vget_lane_s8(__rev2_291, __p3_291), __rev0_291, __p1_291); \
44179   __ret_291 = __builtin_shufflevector(__ret_291, __ret_291, 7, 6, 5, 4, 3, 2, 1, 0); \
44180   __ret_291; \
44181 })
44182 #endif
44183 
44184 #ifdef __LITTLE_ENDIAN__
44185 #define vcopy_lane_f32(__p0_292, __p1_292, __p2_292, __p3_292) __extension__ ({ \
44186   float32x2_t __ret_292; \
44187   float32x2_t __s0_292 = __p0_292; \
44188   float32x2_t __s2_292 = __p2_292; \
44189   __ret_292 = vset_lane_f32(vget_lane_f32(__s2_292, __p3_292), __s0_292, __p1_292); \
44190   __ret_292; \
44191 })
44192 #else
44193 #define vcopy_lane_f32(__p0_293, __p1_293, __p2_293, __p3_293) __extension__ ({ \
44194   float32x2_t __ret_293; \
44195   float32x2_t __s0_293 = __p0_293; \
44196   float32x2_t __s2_293 = __p2_293; \
44197   float32x2_t __rev0_293;  __rev0_293 = __builtin_shufflevector(__s0_293, __s0_293, 1, 0); \
44198   float32x2_t __rev2_293;  __rev2_293 = __builtin_shufflevector(__s2_293, __s2_293, 1, 0); \
44199   __ret_293 = __noswap_vset_lane_f32(__noswap_vget_lane_f32(__rev2_293, __p3_293), __rev0_293, __p1_293); \
44200   __ret_293 = __builtin_shufflevector(__ret_293, __ret_293, 1, 0); \
44201   __ret_293; \
44202 })
44203 #endif
44204 
44205 #ifdef __LITTLE_ENDIAN__
44206 #define vcopy_lane_s32(__p0_294, __p1_294, __p2_294, __p3_294) __extension__ ({ \
44207   int32x2_t __ret_294; \
44208   int32x2_t __s0_294 = __p0_294; \
44209   int32x2_t __s2_294 = __p2_294; \
44210   __ret_294 = vset_lane_s32(vget_lane_s32(__s2_294, __p3_294), __s0_294, __p1_294); \
44211   __ret_294; \
44212 })
44213 #else
44214 #define vcopy_lane_s32(__p0_295, __p1_295, __p2_295, __p3_295) __extension__ ({ \
44215   int32x2_t __ret_295; \
44216   int32x2_t __s0_295 = __p0_295; \
44217   int32x2_t __s2_295 = __p2_295; \
44218   int32x2_t __rev0_295;  __rev0_295 = __builtin_shufflevector(__s0_295, __s0_295, 1, 0); \
44219   int32x2_t __rev2_295;  __rev2_295 = __builtin_shufflevector(__s2_295, __s2_295, 1, 0); \
44220   __ret_295 = __noswap_vset_lane_s32(__noswap_vget_lane_s32(__rev2_295, __p3_295), __rev0_295, __p1_295); \
44221   __ret_295 = __builtin_shufflevector(__ret_295, __ret_295, 1, 0); \
44222   __ret_295; \
44223 })
44224 #endif
44225 
44226 #define vcopy_lane_s64(__p0_296, __p1_296, __p2_296, __p3_296) __extension__ ({ \
44227   int64x1_t __ret_296; \
44228   int64x1_t __s0_296 = __p0_296; \
44229   int64x1_t __s2_296 = __p2_296; \
44230   __ret_296 = vset_lane_s64(vget_lane_s64(__s2_296, __p3_296), __s0_296, __p1_296); \
44231   __ret_296; \
44232 })
44233 #ifdef __LITTLE_ENDIAN__
44234 #define vcopy_lane_s16(__p0_297, __p1_297, __p2_297, __p3_297) __extension__ ({ \
44235   int16x4_t __ret_297; \
44236   int16x4_t __s0_297 = __p0_297; \
44237   int16x4_t __s2_297 = __p2_297; \
44238   __ret_297 = vset_lane_s16(vget_lane_s16(__s2_297, __p3_297), __s0_297, __p1_297); \
44239   __ret_297; \
44240 })
44241 #else
44242 #define vcopy_lane_s16(__p0_298, __p1_298, __p2_298, __p3_298) __extension__ ({ \
44243   int16x4_t __ret_298; \
44244   int16x4_t __s0_298 = __p0_298; \
44245   int16x4_t __s2_298 = __p2_298; \
44246   int16x4_t __rev0_298;  __rev0_298 = __builtin_shufflevector(__s0_298, __s0_298, 3, 2, 1, 0); \
44247   int16x4_t __rev2_298;  __rev2_298 = __builtin_shufflevector(__s2_298, __s2_298, 3, 2, 1, 0); \
44248   __ret_298 = __noswap_vset_lane_s16(__noswap_vget_lane_s16(__rev2_298, __p3_298), __rev0_298, __p1_298); \
44249   __ret_298 = __builtin_shufflevector(__ret_298, __ret_298, 3, 2, 1, 0); \
44250   __ret_298; \
44251 })
44252 #endif
44253 
44254 #ifdef __LITTLE_ENDIAN__
44255 #define vcopyq_laneq_p8(__p0_299, __p1_299, __p2_299, __p3_299) __extension__ ({ \
44256   poly8x16_t __ret_299; \
44257   poly8x16_t __s0_299 = __p0_299; \
44258   poly8x16_t __s2_299 = __p2_299; \
44259   __ret_299 = vsetq_lane_p8(vgetq_lane_p8(__s2_299, __p3_299), __s0_299, __p1_299); \
44260   __ret_299; \
44261 })
44262 #else
44263 #define vcopyq_laneq_p8(__p0_300, __p1_300, __p2_300, __p3_300) __extension__ ({ \
44264   poly8x16_t __ret_300; \
44265   poly8x16_t __s0_300 = __p0_300; \
44266   poly8x16_t __s2_300 = __p2_300; \
44267   poly8x16_t __rev0_300;  __rev0_300 = __builtin_shufflevector(__s0_300, __s0_300, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44268   poly8x16_t __rev2_300;  __rev2_300 = __builtin_shufflevector(__s2_300, __s2_300, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44269   __ret_300 = __noswap_vsetq_lane_p8(__noswap_vgetq_lane_p8(__rev2_300, __p3_300), __rev0_300, __p1_300); \
44270   __ret_300 = __builtin_shufflevector(__ret_300, __ret_300, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44271   __ret_300; \
44272 })
44273 #endif
44274 
44275 #ifdef __LITTLE_ENDIAN__
44276 #define vcopyq_laneq_p16(__p0_301, __p1_301, __p2_301, __p3_301) __extension__ ({ \
44277   poly16x8_t __ret_301; \
44278   poly16x8_t __s0_301 = __p0_301; \
44279   poly16x8_t __s2_301 = __p2_301; \
44280   __ret_301 = vsetq_lane_p16(vgetq_lane_p16(__s2_301, __p3_301), __s0_301, __p1_301); \
44281   __ret_301; \
44282 })
44283 #else
44284 #define vcopyq_laneq_p16(__p0_302, __p1_302, __p2_302, __p3_302) __extension__ ({ \
44285   poly16x8_t __ret_302; \
44286   poly16x8_t __s0_302 = __p0_302; \
44287   poly16x8_t __s2_302 = __p2_302; \
44288   poly16x8_t __rev0_302;  __rev0_302 = __builtin_shufflevector(__s0_302, __s0_302, 7, 6, 5, 4, 3, 2, 1, 0); \
44289   poly16x8_t __rev2_302;  __rev2_302 = __builtin_shufflevector(__s2_302, __s2_302, 7, 6, 5, 4, 3, 2, 1, 0); \
44290   __ret_302 = __noswap_vsetq_lane_p16(__noswap_vgetq_lane_p16(__rev2_302, __p3_302), __rev0_302, __p1_302); \
44291   __ret_302 = __builtin_shufflevector(__ret_302, __ret_302, 7, 6, 5, 4, 3, 2, 1, 0); \
44292   __ret_302; \
44293 })
44294 #endif
44295 
44296 #ifdef __LITTLE_ENDIAN__
44297 #define vcopyq_laneq_u8(__p0_303, __p1_303, __p2_303, __p3_303) __extension__ ({ \
44298   uint8x16_t __ret_303; \
44299   uint8x16_t __s0_303 = __p0_303; \
44300   uint8x16_t __s2_303 = __p2_303; \
44301   __ret_303 = vsetq_lane_u8(vgetq_lane_u8(__s2_303, __p3_303), __s0_303, __p1_303); \
44302   __ret_303; \
44303 })
44304 #else
44305 #define vcopyq_laneq_u8(__p0_304, __p1_304, __p2_304, __p3_304) __extension__ ({ \
44306   uint8x16_t __ret_304; \
44307   uint8x16_t __s0_304 = __p0_304; \
44308   uint8x16_t __s2_304 = __p2_304; \
44309   uint8x16_t __rev0_304;  __rev0_304 = __builtin_shufflevector(__s0_304, __s0_304, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44310   uint8x16_t __rev2_304;  __rev2_304 = __builtin_shufflevector(__s2_304, __s2_304, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44311   __ret_304 = __noswap_vsetq_lane_u8(__noswap_vgetq_lane_u8(__rev2_304, __p3_304), __rev0_304, __p1_304); \
44312   __ret_304 = __builtin_shufflevector(__ret_304, __ret_304, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44313   __ret_304; \
44314 })
44315 #endif
44316 
44317 #ifdef __LITTLE_ENDIAN__
44318 #define vcopyq_laneq_u32(__p0_305, __p1_305, __p2_305, __p3_305) __extension__ ({ \
44319   uint32x4_t __ret_305; \
44320   uint32x4_t __s0_305 = __p0_305; \
44321   uint32x4_t __s2_305 = __p2_305; \
44322   __ret_305 = vsetq_lane_u32(vgetq_lane_u32(__s2_305, __p3_305), __s0_305, __p1_305); \
44323   __ret_305; \
44324 })
44325 #else
44326 #define vcopyq_laneq_u32(__p0_306, __p1_306, __p2_306, __p3_306) __extension__ ({ \
44327   uint32x4_t __ret_306; \
44328   uint32x4_t __s0_306 = __p0_306; \
44329   uint32x4_t __s2_306 = __p2_306; \
44330   uint32x4_t __rev0_306;  __rev0_306 = __builtin_shufflevector(__s0_306, __s0_306, 3, 2, 1, 0); \
44331   uint32x4_t __rev2_306;  __rev2_306 = __builtin_shufflevector(__s2_306, __s2_306, 3, 2, 1, 0); \
44332   __ret_306 = __noswap_vsetq_lane_u32(__noswap_vgetq_lane_u32(__rev2_306, __p3_306), __rev0_306, __p1_306); \
44333   __ret_306 = __builtin_shufflevector(__ret_306, __ret_306, 3, 2, 1, 0); \
44334   __ret_306; \
44335 })
44336 #endif
44337 
44338 #ifdef __LITTLE_ENDIAN__
44339 #define vcopyq_laneq_u64(__p0_307, __p1_307, __p2_307, __p3_307) __extension__ ({ \
44340   uint64x2_t __ret_307; \
44341   uint64x2_t __s0_307 = __p0_307; \
44342   uint64x2_t __s2_307 = __p2_307; \
44343   __ret_307 = vsetq_lane_u64(vgetq_lane_u64(__s2_307, __p3_307), __s0_307, __p1_307); \
44344   __ret_307; \
44345 })
44346 #else
44347 #define vcopyq_laneq_u64(__p0_308, __p1_308, __p2_308, __p3_308) __extension__ ({ \
44348   uint64x2_t __ret_308; \
44349   uint64x2_t __s0_308 = __p0_308; \
44350   uint64x2_t __s2_308 = __p2_308; \
44351   uint64x2_t __rev0_308;  __rev0_308 = __builtin_shufflevector(__s0_308, __s0_308, 1, 0); \
44352   uint64x2_t __rev2_308;  __rev2_308 = __builtin_shufflevector(__s2_308, __s2_308, 1, 0); \
44353   __ret_308 = __noswap_vsetq_lane_u64(__noswap_vgetq_lane_u64(__rev2_308, __p3_308), __rev0_308, __p1_308); \
44354   __ret_308 = __builtin_shufflevector(__ret_308, __ret_308, 1, 0); \
44355   __ret_308; \
44356 })
44357 #endif
44358 
44359 #ifdef __LITTLE_ENDIAN__
44360 #define vcopyq_laneq_u16(__p0_309, __p1_309, __p2_309, __p3_309) __extension__ ({ \
44361   uint16x8_t __ret_309; \
44362   uint16x8_t __s0_309 = __p0_309; \
44363   uint16x8_t __s2_309 = __p2_309; \
44364   __ret_309 = vsetq_lane_u16(vgetq_lane_u16(__s2_309, __p3_309), __s0_309, __p1_309); \
44365   __ret_309; \
44366 })
44367 #else
44368 #define vcopyq_laneq_u16(__p0_310, __p1_310, __p2_310, __p3_310) __extension__ ({ \
44369   uint16x8_t __ret_310; \
44370   uint16x8_t __s0_310 = __p0_310; \
44371   uint16x8_t __s2_310 = __p2_310; \
44372   uint16x8_t __rev0_310;  __rev0_310 = __builtin_shufflevector(__s0_310, __s0_310, 7, 6, 5, 4, 3, 2, 1, 0); \
44373   uint16x8_t __rev2_310;  __rev2_310 = __builtin_shufflevector(__s2_310, __s2_310, 7, 6, 5, 4, 3, 2, 1, 0); \
44374   __ret_310 = __noswap_vsetq_lane_u16(__noswap_vgetq_lane_u16(__rev2_310, __p3_310), __rev0_310, __p1_310); \
44375   __ret_310 = __builtin_shufflevector(__ret_310, __ret_310, 7, 6, 5, 4, 3, 2, 1, 0); \
44376   __ret_310; \
44377 })
44378 #endif
44379 
44380 #ifdef __LITTLE_ENDIAN__
44381 #define vcopyq_laneq_s8(__p0_311, __p1_311, __p2_311, __p3_311) __extension__ ({ \
44382   int8x16_t __ret_311; \
44383   int8x16_t __s0_311 = __p0_311; \
44384   int8x16_t __s2_311 = __p2_311; \
44385   __ret_311 = vsetq_lane_s8(vgetq_lane_s8(__s2_311, __p3_311), __s0_311, __p1_311); \
44386   __ret_311; \
44387 })
44388 #else
44389 #define vcopyq_laneq_s8(__p0_312, __p1_312, __p2_312, __p3_312) __extension__ ({ \
44390   int8x16_t __ret_312; \
44391   int8x16_t __s0_312 = __p0_312; \
44392   int8x16_t __s2_312 = __p2_312; \
44393   int8x16_t __rev0_312;  __rev0_312 = __builtin_shufflevector(__s0_312, __s0_312, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44394   int8x16_t __rev2_312;  __rev2_312 = __builtin_shufflevector(__s2_312, __s2_312, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44395   __ret_312 = __noswap_vsetq_lane_s8(__noswap_vgetq_lane_s8(__rev2_312, __p3_312), __rev0_312, __p1_312); \
44396   __ret_312 = __builtin_shufflevector(__ret_312, __ret_312, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44397   __ret_312; \
44398 })
44399 #endif
44400 
44401 #ifdef __LITTLE_ENDIAN__
44402 #define vcopyq_laneq_f32(__p0_313, __p1_313, __p2_313, __p3_313) __extension__ ({ \
44403   float32x4_t __ret_313; \
44404   float32x4_t __s0_313 = __p0_313; \
44405   float32x4_t __s2_313 = __p2_313; \
44406   __ret_313 = vsetq_lane_f32(vgetq_lane_f32(__s2_313, __p3_313), __s0_313, __p1_313); \
44407   __ret_313; \
44408 })
44409 #else
44410 #define vcopyq_laneq_f32(__p0_314, __p1_314, __p2_314, __p3_314) __extension__ ({ \
44411   float32x4_t __ret_314; \
44412   float32x4_t __s0_314 = __p0_314; \
44413   float32x4_t __s2_314 = __p2_314; \
44414   float32x4_t __rev0_314;  __rev0_314 = __builtin_shufflevector(__s0_314, __s0_314, 3, 2, 1, 0); \
44415   float32x4_t __rev2_314;  __rev2_314 = __builtin_shufflevector(__s2_314, __s2_314, 3, 2, 1, 0); \
44416   __ret_314 = __noswap_vsetq_lane_f32(__noswap_vgetq_lane_f32(__rev2_314, __p3_314), __rev0_314, __p1_314); \
44417   __ret_314 = __builtin_shufflevector(__ret_314, __ret_314, 3, 2, 1, 0); \
44418   __ret_314; \
44419 })
44420 #endif
44421 
44422 #ifdef __LITTLE_ENDIAN__
44423 #define vcopyq_laneq_s32(__p0_315, __p1_315, __p2_315, __p3_315) __extension__ ({ \
44424   int32x4_t __ret_315; \
44425   int32x4_t __s0_315 = __p0_315; \
44426   int32x4_t __s2_315 = __p2_315; \
44427   __ret_315 = vsetq_lane_s32(vgetq_lane_s32(__s2_315, __p3_315), __s0_315, __p1_315); \
44428   __ret_315; \
44429 })
44430 #else
44431 #define vcopyq_laneq_s32(__p0_316, __p1_316, __p2_316, __p3_316) __extension__ ({ \
44432   int32x4_t __ret_316; \
44433   int32x4_t __s0_316 = __p0_316; \
44434   int32x4_t __s2_316 = __p2_316; \
44435   int32x4_t __rev0_316;  __rev0_316 = __builtin_shufflevector(__s0_316, __s0_316, 3, 2, 1, 0); \
44436   int32x4_t __rev2_316;  __rev2_316 = __builtin_shufflevector(__s2_316, __s2_316, 3, 2, 1, 0); \
44437   __ret_316 = __noswap_vsetq_lane_s32(__noswap_vgetq_lane_s32(__rev2_316, __p3_316), __rev0_316, __p1_316); \
44438   __ret_316 = __builtin_shufflevector(__ret_316, __ret_316, 3, 2, 1, 0); \
44439   __ret_316; \
44440 })
44441 #endif
44442 
44443 #ifdef __LITTLE_ENDIAN__
44444 #define vcopyq_laneq_s64(__p0_317, __p1_317, __p2_317, __p3_317) __extension__ ({ \
44445   int64x2_t __ret_317; \
44446   int64x2_t __s0_317 = __p0_317; \
44447   int64x2_t __s2_317 = __p2_317; \
44448   __ret_317 = vsetq_lane_s64(vgetq_lane_s64(__s2_317, __p3_317), __s0_317, __p1_317); \
44449   __ret_317; \
44450 })
44451 #else
44452 #define vcopyq_laneq_s64(__p0_318, __p1_318, __p2_318, __p3_318) __extension__ ({ \
44453   int64x2_t __ret_318; \
44454   int64x2_t __s0_318 = __p0_318; \
44455   int64x2_t __s2_318 = __p2_318; \
44456   int64x2_t __rev0_318;  __rev0_318 = __builtin_shufflevector(__s0_318, __s0_318, 1, 0); \
44457   int64x2_t __rev2_318;  __rev2_318 = __builtin_shufflevector(__s2_318, __s2_318, 1, 0); \
44458   __ret_318 = __noswap_vsetq_lane_s64(__noswap_vgetq_lane_s64(__rev2_318, __p3_318), __rev0_318, __p1_318); \
44459   __ret_318 = __builtin_shufflevector(__ret_318, __ret_318, 1, 0); \
44460   __ret_318; \
44461 })
44462 #endif
44463 
44464 #ifdef __LITTLE_ENDIAN__
44465 #define vcopyq_laneq_s16(__p0_319, __p1_319, __p2_319, __p3_319) __extension__ ({ \
44466   int16x8_t __ret_319; \
44467   int16x8_t __s0_319 = __p0_319; \
44468   int16x8_t __s2_319 = __p2_319; \
44469   __ret_319 = vsetq_lane_s16(vgetq_lane_s16(__s2_319, __p3_319), __s0_319, __p1_319); \
44470   __ret_319; \
44471 })
44472 #else
44473 #define vcopyq_laneq_s16(__p0_320, __p1_320, __p2_320, __p3_320) __extension__ ({ \
44474   int16x8_t __ret_320; \
44475   int16x8_t __s0_320 = __p0_320; \
44476   int16x8_t __s2_320 = __p2_320; \
44477   int16x8_t __rev0_320;  __rev0_320 = __builtin_shufflevector(__s0_320, __s0_320, 7, 6, 5, 4, 3, 2, 1, 0); \
44478   int16x8_t __rev2_320;  __rev2_320 = __builtin_shufflevector(__s2_320, __s2_320, 7, 6, 5, 4, 3, 2, 1, 0); \
44479   __ret_320 = __noswap_vsetq_lane_s16(__noswap_vgetq_lane_s16(__rev2_320, __p3_320), __rev0_320, __p1_320); \
44480   __ret_320 = __builtin_shufflevector(__ret_320, __ret_320, 7, 6, 5, 4, 3, 2, 1, 0); \
44481   __ret_320; \
44482 })
44483 #endif
44484 
44485 #ifdef __LITTLE_ENDIAN__
44486 #define vcopy_laneq_p8(__p0_321, __p1_321, __p2_321, __p3_321) __extension__ ({ \
44487   poly8x8_t __ret_321; \
44488   poly8x8_t __s0_321 = __p0_321; \
44489   poly8x16_t __s2_321 = __p2_321; \
44490   __ret_321 = vset_lane_p8(vgetq_lane_p8(__s2_321, __p3_321), __s0_321, __p1_321); \
44491   __ret_321; \
44492 })
44493 #else
44494 #define vcopy_laneq_p8(__p0_322, __p1_322, __p2_322, __p3_322) __extension__ ({ \
44495   poly8x8_t __ret_322; \
44496   poly8x8_t __s0_322 = __p0_322; \
44497   poly8x16_t __s2_322 = __p2_322; \
44498   poly8x8_t __rev0_322;  __rev0_322 = __builtin_shufflevector(__s0_322, __s0_322, 7, 6, 5, 4, 3, 2, 1, 0); \
44499   poly8x16_t __rev2_322;  __rev2_322 = __builtin_shufflevector(__s2_322, __s2_322, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44500   __ret_322 = __noswap_vset_lane_p8(__noswap_vgetq_lane_p8(__rev2_322, __p3_322), __rev0_322, __p1_322); \
44501   __ret_322 = __builtin_shufflevector(__ret_322, __ret_322, 7, 6, 5, 4, 3, 2, 1, 0); \
44502   __ret_322; \
44503 })
44504 #endif
44505 
44506 #ifdef __LITTLE_ENDIAN__
44507 #define vcopy_laneq_p16(__p0_323, __p1_323, __p2_323, __p3_323) __extension__ ({ \
44508   poly16x4_t __ret_323; \
44509   poly16x4_t __s0_323 = __p0_323; \
44510   poly16x8_t __s2_323 = __p2_323; \
44511   __ret_323 = vset_lane_p16(vgetq_lane_p16(__s2_323, __p3_323), __s0_323, __p1_323); \
44512   __ret_323; \
44513 })
44514 #else
44515 #define vcopy_laneq_p16(__p0_324, __p1_324, __p2_324, __p3_324) __extension__ ({ \
44516   poly16x4_t __ret_324; \
44517   poly16x4_t __s0_324 = __p0_324; \
44518   poly16x8_t __s2_324 = __p2_324; \
44519   poly16x4_t __rev0_324;  __rev0_324 = __builtin_shufflevector(__s0_324, __s0_324, 3, 2, 1, 0); \
44520   poly16x8_t __rev2_324;  __rev2_324 = __builtin_shufflevector(__s2_324, __s2_324, 7, 6, 5, 4, 3, 2, 1, 0); \
44521   __ret_324 = __noswap_vset_lane_p16(__noswap_vgetq_lane_p16(__rev2_324, __p3_324), __rev0_324, __p1_324); \
44522   __ret_324 = __builtin_shufflevector(__ret_324, __ret_324, 3, 2, 1, 0); \
44523   __ret_324; \
44524 })
44525 #endif
44526 
44527 #ifdef __LITTLE_ENDIAN__
44528 #define vcopy_laneq_u8(__p0_325, __p1_325, __p2_325, __p3_325) __extension__ ({ \
44529   uint8x8_t __ret_325; \
44530   uint8x8_t __s0_325 = __p0_325; \
44531   uint8x16_t __s2_325 = __p2_325; \
44532   __ret_325 = vset_lane_u8(vgetq_lane_u8(__s2_325, __p3_325), __s0_325, __p1_325); \
44533   __ret_325; \
44534 })
44535 #else
44536 #define vcopy_laneq_u8(__p0_326, __p1_326, __p2_326, __p3_326) __extension__ ({ \
44537   uint8x8_t __ret_326; \
44538   uint8x8_t __s0_326 = __p0_326; \
44539   uint8x16_t __s2_326 = __p2_326; \
44540   uint8x8_t __rev0_326;  __rev0_326 = __builtin_shufflevector(__s0_326, __s0_326, 7, 6, 5, 4, 3, 2, 1, 0); \
44541   uint8x16_t __rev2_326;  __rev2_326 = __builtin_shufflevector(__s2_326, __s2_326, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44542   __ret_326 = __noswap_vset_lane_u8(__noswap_vgetq_lane_u8(__rev2_326, __p3_326), __rev0_326, __p1_326); \
44543   __ret_326 = __builtin_shufflevector(__ret_326, __ret_326, 7, 6, 5, 4, 3, 2, 1, 0); \
44544   __ret_326; \
44545 })
44546 #endif
44547 
44548 #ifdef __LITTLE_ENDIAN__
44549 #define vcopy_laneq_u32(__p0_327, __p1_327, __p2_327, __p3_327) __extension__ ({ \
44550   uint32x2_t __ret_327; \
44551   uint32x2_t __s0_327 = __p0_327; \
44552   uint32x4_t __s2_327 = __p2_327; \
44553   __ret_327 = vset_lane_u32(vgetq_lane_u32(__s2_327, __p3_327), __s0_327, __p1_327); \
44554   __ret_327; \
44555 })
44556 #else
44557 #define vcopy_laneq_u32(__p0_328, __p1_328, __p2_328, __p3_328) __extension__ ({ \
44558   uint32x2_t __ret_328; \
44559   uint32x2_t __s0_328 = __p0_328; \
44560   uint32x4_t __s2_328 = __p2_328; \
44561   uint32x2_t __rev0_328;  __rev0_328 = __builtin_shufflevector(__s0_328, __s0_328, 1, 0); \
44562   uint32x4_t __rev2_328;  __rev2_328 = __builtin_shufflevector(__s2_328, __s2_328, 3, 2, 1, 0); \
44563   __ret_328 = __noswap_vset_lane_u32(__noswap_vgetq_lane_u32(__rev2_328, __p3_328), __rev0_328, __p1_328); \
44564   __ret_328 = __builtin_shufflevector(__ret_328, __ret_328, 1, 0); \
44565   __ret_328; \
44566 })
44567 #endif
44568 
44569 #ifdef __LITTLE_ENDIAN__
44570 #define vcopy_laneq_u64(__p0_329, __p1_329, __p2_329, __p3_329) __extension__ ({ \
44571   uint64x1_t __ret_329; \
44572   uint64x1_t __s0_329 = __p0_329; \
44573   uint64x2_t __s2_329 = __p2_329; \
44574   __ret_329 = vset_lane_u64(vgetq_lane_u64(__s2_329, __p3_329), __s0_329, __p1_329); \
44575   __ret_329; \
44576 })
44577 #else
44578 #define vcopy_laneq_u64(__p0_330, __p1_330, __p2_330, __p3_330) __extension__ ({ \
44579   uint64x1_t __ret_330; \
44580   uint64x1_t __s0_330 = __p0_330; \
44581   uint64x2_t __s2_330 = __p2_330; \
44582   uint64x2_t __rev2_330;  __rev2_330 = __builtin_shufflevector(__s2_330, __s2_330, 1, 0); \
44583   __ret_330 = vset_lane_u64(__noswap_vgetq_lane_u64(__rev2_330, __p3_330), __s0_330, __p1_330); \
44584   __ret_330; \
44585 })
44586 #endif
44587 
44588 #ifdef __LITTLE_ENDIAN__
44589 #define vcopy_laneq_u16(__p0_331, __p1_331, __p2_331, __p3_331) __extension__ ({ \
44590   uint16x4_t __ret_331; \
44591   uint16x4_t __s0_331 = __p0_331; \
44592   uint16x8_t __s2_331 = __p2_331; \
44593   __ret_331 = vset_lane_u16(vgetq_lane_u16(__s2_331, __p3_331), __s0_331, __p1_331); \
44594   __ret_331; \
44595 })
44596 #else
44597 #define vcopy_laneq_u16(__p0_332, __p1_332, __p2_332, __p3_332) __extension__ ({ \
44598   uint16x4_t __ret_332; \
44599   uint16x4_t __s0_332 = __p0_332; \
44600   uint16x8_t __s2_332 = __p2_332; \
44601   uint16x4_t __rev0_332;  __rev0_332 = __builtin_shufflevector(__s0_332, __s0_332, 3, 2, 1, 0); \
44602   uint16x8_t __rev2_332;  __rev2_332 = __builtin_shufflevector(__s2_332, __s2_332, 7, 6, 5, 4, 3, 2, 1, 0); \
44603   __ret_332 = __noswap_vset_lane_u16(__noswap_vgetq_lane_u16(__rev2_332, __p3_332), __rev0_332, __p1_332); \
44604   __ret_332 = __builtin_shufflevector(__ret_332, __ret_332, 3, 2, 1, 0); \
44605   __ret_332; \
44606 })
44607 #endif
44608 
44609 #ifdef __LITTLE_ENDIAN__
44610 #define vcopy_laneq_s8(__p0_333, __p1_333, __p2_333, __p3_333) __extension__ ({ \
44611   int8x8_t __ret_333; \
44612   int8x8_t __s0_333 = __p0_333; \
44613   int8x16_t __s2_333 = __p2_333; \
44614   __ret_333 = vset_lane_s8(vgetq_lane_s8(__s2_333, __p3_333), __s0_333, __p1_333); \
44615   __ret_333; \
44616 })
44617 #else
44618 #define vcopy_laneq_s8(__p0_334, __p1_334, __p2_334, __p3_334) __extension__ ({ \
44619   int8x8_t __ret_334; \
44620   int8x8_t __s0_334 = __p0_334; \
44621   int8x16_t __s2_334 = __p2_334; \
44622   int8x8_t __rev0_334;  __rev0_334 = __builtin_shufflevector(__s0_334, __s0_334, 7, 6, 5, 4, 3, 2, 1, 0); \
44623   int8x16_t __rev2_334;  __rev2_334 = __builtin_shufflevector(__s2_334, __s2_334, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44624   __ret_334 = __noswap_vset_lane_s8(__noswap_vgetq_lane_s8(__rev2_334, __p3_334), __rev0_334, __p1_334); \
44625   __ret_334 = __builtin_shufflevector(__ret_334, __ret_334, 7, 6, 5, 4, 3, 2, 1, 0); \
44626   __ret_334; \
44627 })
44628 #endif
44629 
44630 #ifdef __LITTLE_ENDIAN__
44631 #define vcopy_laneq_f32(__p0_335, __p1_335, __p2_335, __p3_335) __extension__ ({ \
44632   float32x2_t __ret_335; \
44633   float32x2_t __s0_335 = __p0_335; \
44634   float32x4_t __s2_335 = __p2_335; \
44635   __ret_335 = vset_lane_f32(vgetq_lane_f32(__s2_335, __p3_335), __s0_335, __p1_335); \
44636   __ret_335; \
44637 })
44638 #else
44639 #define vcopy_laneq_f32(__p0_336, __p1_336, __p2_336, __p3_336) __extension__ ({ \
44640   float32x2_t __ret_336; \
44641   float32x2_t __s0_336 = __p0_336; \
44642   float32x4_t __s2_336 = __p2_336; \
44643   float32x2_t __rev0_336;  __rev0_336 = __builtin_shufflevector(__s0_336, __s0_336, 1, 0); \
44644   float32x4_t __rev2_336;  __rev2_336 = __builtin_shufflevector(__s2_336, __s2_336, 3, 2, 1, 0); \
44645   __ret_336 = __noswap_vset_lane_f32(__noswap_vgetq_lane_f32(__rev2_336, __p3_336), __rev0_336, __p1_336); \
44646   __ret_336 = __builtin_shufflevector(__ret_336, __ret_336, 1, 0); \
44647   __ret_336; \
44648 })
44649 #endif
44650 
44651 #ifdef __LITTLE_ENDIAN__
44652 #define vcopy_laneq_s32(__p0_337, __p1_337, __p2_337, __p3_337) __extension__ ({ \
44653   int32x2_t __ret_337; \
44654   int32x2_t __s0_337 = __p0_337; \
44655   int32x4_t __s2_337 = __p2_337; \
44656   __ret_337 = vset_lane_s32(vgetq_lane_s32(__s2_337, __p3_337), __s0_337, __p1_337); \
44657   __ret_337; \
44658 })
44659 #else
44660 #define vcopy_laneq_s32(__p0_338, __p1_338, __p2_338, __p3_338) __extension__ ({ \
44661   int32x2_t __ret_338; \
44662   int32x2_t __s0_338 = __p0_338; \
44663   int32x4_t __s2_338 = __p2_338; \
44664   int32x2_t __rev0_338;  __rev0_338 = __builtin_shufflevector(__s0_338, __s0_338, 1, 0); \
44665   int32x4_t __rev2_338;  __rev2_338 = __builtin_shufflevector(__s2_338, __s2_338, 3, 2, 1, 0); \
44666   __ret_338 = __noswap_vset_lane_s32(__noswap_vgetq_lane_s32(__rev2_338, __p3_338), __rev0_338, __p1_338); \
44667   __ret_338 = __builtin_shufflevector(__ret_338, __ret_338, 1, 0); \
44668   __ret_338; \
44669 })
44670 #endif
44671 
44672 #ifdef __LITTLE_ENDIAN__
44673 #define vcopy_laneq_s64(__p0_339, __p1_339, __p2_339, __p3_339) __extension__ ({ \
44674   int64x1_t __ret_339; \
44675   int64x1_t __s0_339 = __p0_339; \
44676   int64x2_t __s2_339 = __p2_339; \
44677   __ret_339 = vset_lane_s64(vgetq_lane_s64(__s2_339, __p3_339), __s0_339, __p1_339); \
44678   __ret_339; \
44679 })
44680 #else
44681 #define vcopy_laneq_s64(__p0_340, __p1_340, __p2_340, __p3_340) __extension__ ({ \
44682   int64x1_t __ret_340; \
44683   int64x1_t __s0_340 = __p0_340; \
44684   int64x2_t __s2_340 = __p2_340; \
44685   int64x2_t __rev2_340;  __rev2_340 = __builtin_shufflevector(__s2_340, __s2_340, 1, 0); \
44686   __ret_340 = vset_lane_s64(__noswap_vgetq_lane_s64(__rev2_340, __p3_340), __s0_340, __p1_340); \
44687   __ret_340; \
44688 })
44689 #endif
44690 
44691 #ifdef __LITTLE_ENDIAN__
44692 #define vcopy_laneq_s16(__p0_341, __p1_341, __p2_341, __p3_341) __extension__ ({ \
44693   int16x4_t __ret_341; \
44694   int16x4_t __s0_341 = __p0_341; \
44695   int16x8_t __s2_341 = __p2_341; \
44696   __ret_341 = vset_lane_s16(vgetq_lane_s16(__s2_341, __p3_341), __s0_341, __p1_341); \
44697   __ret_341; \
44698 })
44699 #else
44700 #define vcopy_laneq_s16(__p0_342, __p1_342, __p2_342, __p3_342) __extension__ ({ \
44701   int16x4_t __ret_342; \
44702   int16x4_t __s0_342 = __p0_342; \
44703   int16x8_t __s2_342 = __p2_342; \
44704   int16x4_t __rev0_342;  __rev0_342 = __builtin_shufflevector(__s0_342, __s0_342, 3, 2, 1, 0); \
44705   int16x8_t __rev2_342;  __rev2_342 = __builtin_shufflevector(__s2_342, __s2_342, 7, 6, 5, 4, 3, 2, 1, 0); \
44706   __ret_342 = __noswap_vset_lane_s16(__noswap_vgetq_lane_s16(__rev2_342, __p3_342), __rev0_342, __p1_342); \
44707   __ret_342 = __builtin_shufflevector(__ret_342, __ret_342, 3, 2, 1, 0); \
44708   __ret_342; \
44709 })
44710 #endif
44711 
44712 #define vcreate_p64(__p0) __extension__ ({ \
44713   poly64x1_t __ret; \
44714   uint64_t __promote = __p0; \
44715   __ret = (poly64x1_t)(__promote); \
44716   __ret; \
44717 })
44718 #define vcreate_f64(__p0) __extension__ ({ \
44719   float64x1_t __ret; \
44720   uint64_t __promote = __p0; \
44721   __ret = (float64x1_t)(__promote); \
44722   __ret; \
44723 })
vcvts_f32_s32(int32_t __p0)44724 __ai float32_t vcvts_f32_s32(int32_t __p0) {
44725   float32_t __ret;
44726   __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
44727   return __ret;
44728 }
vcvts_f32_u32(uint32_t __p0)44729 __ai float32_t vcvts_f32_u32(uint32_t __p0) {
44730   float32_t __ret;
44731   __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
44732   return __ret;
44733 }
44734 #ifdef __LITTLE_ENDIAN__
vcvt_f32_f64(float64x2_t __p0)44735 __ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
44736   float32x2_t __ret;
44737   __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
44738   return __ret;
44739 }
44740 #else
vcvt_f32_f64(float64x2_t __p0)44741 __ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
44742   float32x2_t __ret;
44743   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44744   __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__rev0, 9);
44745   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44746   return __ret;
44747 }
__noswap_vcvt_f32_f64(float64x2_t __p0)44748 __ai float32x2_t __noswap_vcvt_f32_f64(float64x2_t __p0) {
44749   float32x2_t __ret;
44750   __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
44751   return __ret;
44752 }
44753 #endif
44754 
vcvtd_f64_s64(int64_t __p0)44755 __ai float64_t vcvtd_f64_s64(int64_t __p0) {
44756   float64_t __ret;
44757   __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
44758   return __ret;
44759 }
vcvtd_f64_u64(uint64_t __p0)44760 __ai float64_t vcvtd_f64_u64(uint64_t __p0) {
44761   float64_t __ret;
44762   __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
44763   return __ret;
44764 }
44765 #ifdef __LITTLE_ENDIAN__
vcvtq_f64_u64(uint64x2_t __p0)44766 __ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
44767   float64x2_t __ret;
44768   __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 51);
44769   return __ret;
44770 }
44771 #else
vcvtq_f64_u64(uint64x2_t __p0)44772 __ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
44773   float64x2_t __ret;
44774   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44775   __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 51);
44776   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44777   return __ret;
44778 }
44779 #endif
44780 
44781 #ifdef __LITTLE_ENDIAN__
vcvtq_f64_s64(int64x2_t __p0)44782 __ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
44783   float64x2_t __ret;
44784   __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 35);
44785   return __ret;
44786 }
44787 #else
vcvtq_f64_s64(int64x2_t __p0)44788 __ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
44789   float64x2_t __ret;
44790   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44791   __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 35);
44792   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44793   return __ret;
44794 }
44795 #endif
44796 
vcvt_f64_u64(uint64x1_t __p0)44797 __ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
44798   float64x1_t __ret;
44799   __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
44800   return __ret;
44801 }
vcvt_f64_s64(int64x1_t __p0)44802 __ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
44803   float64x1_t __ret;
44804   __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
44805   return __ret;
44806 }
44807 #ifdef __LITTLE_ENDIAN__
vcvt_f64_f32(float32x2_t __p0)44808 __ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
44809   float64x2_t __ret;
44810   __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
44811   return __ret;
44812 }
44813 #else
vcvt_f64_f32(float32x2_t __p0)44814 __ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
44815   float64x2_t __ret;
44816   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44817   __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__rev0, 42);
44818   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44819   return __ret;
44820 }
__noswap_vcvt_f64_f32(float32x2_t __p0)44821 __ai float64x2_t __noswap_vcvt_f64_f32(float32x2_t __p0) {
44822   float64x2_t __ret;
44823   __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
44824   return __ret;
44825 }
44826 #endif
44827 
44828 #ifdef __LITTLE_ENDIAN__
vcvt_high_f16_f32(float16x4_t __p0,float32x4_t __p1)44829 __ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
44830   float16x8_t __ret;
44831   __ret = vcombine_f16(__p0, vcvt_f16_f32(__p1));
44832   return __ret;
44833 }
44834 #else
vcvt_high_f16_f32(float16x4_t __p0,float32x4_t __p1)44835 __ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
44836   float16x8_t __ret;
44837   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44838   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44839   __ret = __noswap_vcombine_f16(__rev0, __noswap_vcvt_f16_f32(__rev1));
44840   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44841   return __ret;
44842 }
44843 #endif
44844 
44845 #ifdef __LITTLE_ENDIAN__
vcvt_high_f32_f16(float16x8_t __p0)44846 __ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
44847   float32x4_t __ret;
44848   __ret = vcvt_f32_f16(vget_high_f16(__p0));
44849   return __ret;
44850 }
44851 #else
vcvt_high_f32_f16(float16x8_t __p0)44852 __ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
44853   float32x4_t __ret;
44854   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44855   __ret = __noswap_vcvt_f32_f16(__noswap_vget_high_f16(__rev0));
44856   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44857   return __ret;
44858 }
44859 #endif
44860 
44861 #ifdef __LITTLE_ENDIAN__
vcvt_high_f32_f64(float32x2_t __p0,float64x2_t __p1)44862 __ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
44863   float32x4_t __ret;
44864   __ret = vcombine_f32(__p0, vcvt_f32_f64(__p1));
44865   return __ret;
44866 }
44867 #else
vcvt_high_f32_f64(float32x2_t __p0,float64x2_t __p1)44868 __ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
44869   float32x4_t __ret;
44870   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44871   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
44872   __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvt_f32_f64(__rev1));
44873   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44874   return __ret;
44875 }
44876 #endif
44877 
44878 #ifdef __LITTLE_ENDIAN__
vcvt_high_f64_f32(float32x4_t __p0)44879 __ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
44880   float64x2_t __ret;
44881   __ret = vcvt_f64_f32(vget_high_f32(__p0));
44882   return __ret;
44883 }
44884 #else
vcvt_high_f64_f32(float32x4_t __p0)44885 __ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
44886   float64x2_t __ret;
44887   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44888   __ret = __noswap_vcvt_f64_f32(__noswap_vget_high_f32(__rev0));
44889   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44890   return __ret;
44891 }
44892 #endif
44893 
44894 #define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
44895   float32_t __ret; \
44896   uint32_t __s0 = __p0; \
44897   __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
44898   __ret; \
44899 })
44900 #define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
44901   float32_t __ret; \
44902   int32_t __s0 = __p0; \
44903   __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
44904   __ret; \
44905 })
44906 #ifdef __LITTLE_ENDIAN__
44907 #define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
44908   float64x2_t __ret; \
44909   uint64x2_t __s0 = __p0; \
44910   __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 51); \
44911   __ret; \
44912 })
44913 #else
44914 #define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
44915   float64x2_t __ret; \
44916   uint64x2_t __s0 = __p0; \
44917   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
44918   __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 51); \
44919   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
44920   __ret; \
44921 })
44922 #endif
44923 
44924 #ifdef __LITTLE_ENDIAN__
44925 #define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
44926   float64x2_t __ret; \
44927   int64x2_t __s0 = __p0; \
44928   __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 35); \
44929   __ret; \
44930 })
44931 #else
44932 #define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
44933   float64x2_t __ret; \
44934   int64x2_t __s0 = __p0; \
44935   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
44936   __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 35); \
44937   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
44938   __ret; \
44939 })
44940 #endif
44941 
44942 #define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
44943   float64x1_t __ret; \
44944   uint64x1_t __s0 = __p0; \
44945   __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
44946   __ret; \
44947 })
44948 #define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
44949   float64x1_t __ret; \
44950   int64x1_t __s0 = __p0; \
44951   __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
44952   __ret; \
44953 })
44954 #define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
44955   float64_t __ret; \
44956   uint64_t __s0 = __p0; \
44957   __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
44958   __ret; \
44959 })
44960 #define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
44961   float64_t __ret; \
44962   int64_t __s0 = __p0; \
44963   __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
44964   __ret; \
44965 })
44966 #define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
44967   int32_t __ret; \
44968   float32_t __s0 = __p0; \
44969   __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
44970   __ret; \
44971 })
44972 #ifdef __LITTLE_ENDIAN__
44973 #define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
44974   int64x2_t __ret; \
44975   float64x2_t __s0 = __p0; \
44976   __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__s0, __p1, 35); \
44977   __ret; \
44978 })
44979 #else
44980 #define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
44981   int64x2_t __ret; \
44982   float64x2_t __s0 = __p0; \
44983   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
44984   __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__rev0, __p1, 35); \
44985   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
44986   __ret; \
44987 })
44988 #endif
44989 
44990 #define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
44991   int64x1_t __ret; \
44992   float64x1_t __s0 = __p0; \
44993   __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
44994   __ret; \
44995 })
44996 #define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
44997   int64_t __ret; \
44998   float64_t __s0 = __p0; \
44999   __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
45000   __ret; \
45001 })
45002 #define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
45003   uint32_t __ret; \
45004   float32_t __s0 = __p0; \
45005   __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
45006   __ret; \
45007 })
45008 #ifdef __LITTLE_ENDIAN__
45009 #define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
45010   uint64x2_t __ret; \
45011   float64x2_t __s0 = __p0; \
45012   __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__s0, __p1, 51); \
45013   __ret; \
45014 })
45015 #else
45016 #define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
45017   uint64x2_t __ret; \
45018   float64x2_t __s0 = __p0; \
45019   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45020   __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__rev0, __p1, 51); \
45021   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45022   __ret; \
45023 })
45024 #endif
45025 
45026 #define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
45027   uint64x1_t __ret; \
45028   float64x1_t __s0 = __p0; \
45029   __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
45030   __ret; \
45031 })
45032 #define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
45033   uint64_t __ret; \
45034   float64_t __s0 = __p0; \
45035   __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
45036   __ret; \
45037 })
vcvts_s32_f32(float32_t __p0)45038 __ai int32_t vcvts_s32_f32(float32_t __p0) {
45039   int32_t __ret;
45040   __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
45041   return __ret;
45042 }
vcvtd_s64_f64(float64_t __p0)45043 __ai int64_t vcvtd_s64_f64(float64_t __p0) {
45044   int64_t __ret;
45045   __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
45046   return __ret;
45047 }
45048 #ifdef __LITTLE_ENDIAN__
vcvtq_s64_f64(float64x2_t __p0)45049 __ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
45050   int64x2_t __ret;
45051   __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__p0, 35);
45052   return __ret;
45053 }
45054 #else
vcvtq_s64_f64(float64x2_t __p0)45055 __ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
45056   int64x2_t __ret;
45057   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45058   __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__rev0, 35);
45059   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45060   return __ret;
45061 }
45062 #endif
45063 
vcvt_s64_f64(float64x1_t __p0)45064 __ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
45065   int64x1_t __ret;
45066   __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
45067   return __ret;
45068 }
vcvts_u32_f32(float32_t __p0)45069 __ai uint32_t vcvts_u32_f32(float32_t __p0) {
45070   uint32_t __ret;
45071   __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
45072   return __ret;
45073 }
vcvtd_u64_f64(float64_t __p0)45074 __ai uint64_t vcvtd_u64_f64(float64_t __p0) {
45075   uint64_t __ret;
45076   __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
45077   return __ret;
45078 }
45079 #ifdef __LITTLE_ENDIAN__
vcvtq_u64_f64(float64x2_t __p0)45080 __ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
45081   uint64x2_t __ret;
45082   __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__p0, 51);
45083   return __ret;
45084 }
45085 #else
vcvtq_u64_f64(float64x2_t __p0)45086 __ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
45087   uint64x2_t __ret;
45088   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45089   __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__rev0, 51);
45090   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45091   return __ret;
45092 }
45093 #endif
45094 
vcvt_u64_f64(float64x1_t __p0)45095 __ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
45096   uint64x1_t __ret;
45097   __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
45098   return __ret;
45099 }
vcvtas_s32_f32(float32_t __p0)45100 __ai int32_t vcvtas_s32_f32(float32_t __p0) {
45101   int32_t __ret;
45102   __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
45103   return __ret;
45104 }
45105 #ifdef __LITTLE_ENDIAN__
vcvtaq_s64_f64(float64x2_t __p0)45106 __ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
45107   int64x2_t __ret;
45108   __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__p0, 35);
45109   return __ret;
45110 }
45111 #else
vcvtaq_s64_f64(float64x2_t __p0)45112 __ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
45113   int64x2_t __ret;
45114   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45115   __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__rev0, 35);
45116   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45117   return __ret;
45118 }
45119 #endif
45120 
vcvta_s64_f64(float64x1_t __p0)45121 __ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
45122   int64x1_t __ret;
45123   __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
45124   return __ret;
45125 }
vcvtad_s64_f64(float64_t __p0)45126 __ai int64_t vcvtad_s64_f64(float64_t __p0) {
45127   int64_t __ret;
45128   __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
45129   return __ret;
45130 }
vcvtas_u32_f32(float32_t __p0)45131 __ai uint32_t vcvtas_u32_f32(float32_t __p0) {
45132   uint32_t __ret;
45133   __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
45134   return __ret;
45135 }
45136 #ifdef __LITTLE_ENDIAN__
vcvtaq_u64_f64(float64x2_t __p0)45137 __ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
45138   uint64x2_t __ret;
45139   __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__p0, 51);
45140   return __ret;
45141 }
45142 #else
vcvtaq_u64_f64(float64x2_t __p0)45143 __ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
45144   uint64x2_t __ret;
45145   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45146   __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__rev0, 51);
45147   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45148   return __ret;
45149 }
45150 #endif
45151 
vcvta_u64_f64(float64x1_t __p0)45152 __ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
45153   uint64x1_t __ret;
45154   __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
45155   return __ret;
45156 }
vcvtad_u64_f64(float64_t __p0)45157 __ai uint64_t vcvtad_u64_f64(float64_t __p0) {
45158   uint64_t __ret;
45159   __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
45160   return __ret;
45161 }
vcvtms_s32_f32(float32_t __p0)45162 __ai int32_t vcvtms_s32_f32(float32_t __p0) {
45163   int32_t __ret;
45164   __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
45165   return __ret;
45166 }
45167 #ifdef __LITTLE_ENDIAN__
vcvtmq_s64_f64(float64x2_t __p0)45168 __ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
45169   int64x2_t __ret;
45170   __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__p0, 35);
45171   return __ret;
45172 }
45173 #else
vcvtmq_s64_f64(float64x2_t __p0)45174 __ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
45175   int64x2_t __ret;
45176   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45177   __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__rev0, 35);
45178   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45179   return __ret;
45180 }
45181 #endif
45182 
vcvtm_s64_f64(float64x1_t __p0)45183 __ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
45184   int64x1_t __ret;
45185   __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
45186   return __ret;
45187 }
vcvtmd_s64_f64(float64_t __p0)45188 __ai int64_t vcvtmd_s64_f64(float64_t __p0) {
45189   int64_t __ret;
45190   __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
45191   return __ret;
45192 }
vcvtms_u32_f32(float32_t __p0)45193 __ai uint32_t vcvtms_u32_f32(float32_t __p0) {
45194   uint32_t __ret;
45195   __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
45196   return __ret;
45197 }
45198 #ifdef __LITTLE_ENDIAN__
vcvtmq_u64_f64(float64x2_t __p0)45199 __ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
45200   uint64x2_t __ret;
45201   __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__p0, 51);
45202   return __ret;
45203 }
45204 #else
vcvtmq_u64_f64(float64x2_t __p0)45205 __ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
45206   uint64x2_t __ret;
45207   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45208   __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__rev0, 51);
45209   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45210   return __ret;
45211 }
45212 #endif
45213 
vcvtm_u64_f64(float64x1_t __p0)45214 __ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
45215   uint64x1_t __ret;
45216   __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
45217   return __ret;
45218 }
vcvtmd_u64_f64(float64_t __p0)45219 __ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
45220   uint64_t __ret;
45221   __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
45222   return __ret;
45223 }
vcvtns_s32_f32(float32_t __p0)45224 __ai int32_t vcvtns_s32_f32(float32_t __p0) {
45225   int32_t __ret;
45226   __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
45227   return __ret;
45228 }
45229 #ifdef __LITTLE_ENDIAN__
vcvtnq_s64_f64(float64x2_t __p0)45230 __ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
45231   int64x2_t __ret;
45232   __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__p0, 35);
45233   return __ret;
45234 }
45235 #else
vcvtnq_s64_f64(float64x2_t __p0)45236 __ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
45237   int64x2_t __ret;
45238   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45239   __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__rev0, 35);
45240   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45241   return __ret;
45242 }
45243 #endif
45244 
vcvtn_s64_f64(float64x1_t __p0)45245 __ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
45246   int64x1_t __ret;
45247   __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
45248   return __ret;
45249 }
vcvtnd_s64_f64(float64_t __p0)45250 __ai int64_t vcvtnd_s64_f64(float64_t __p0) {
45251   int64_t __ret;
45252   __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
45253   return __ret;
45254 }
vcvtns_u32_f32(float32_t __p0)45255 __ai uint32_t vcvtns_u32_f32(float32_t __p0) {
45256   uint32_t __ret;
45257   __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
45258   return __ret;
45259 }
45260 #ifdef __LITTLE_ENDIAN__
vcvtnq_u64_f64(float64x2_t __p0)45261 __ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
45262   uint64x2_t __ret;
45263   __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__p0, 51);
45264   return __ret;
45265 }
45266 #else
vcvtnq_u64_f64(float64x2_t __p0)45267 __ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
45268   uint64x2_t __ret;
45269   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45270   __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__rev0, 51);
45271   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45272   return __ret;
45273 }
45274 #endif
45275 
vcvtn_u64_f64(float64x1_t __p0)45276 __ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
45277   uint64x1_t __ret;
45278   __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
45279   return __ret;
45280 }
vcvtnd_u64_f64(float64_t __p0)45281 __ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
45282   uint64_t __ret;
45283   __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
45284   return __ret;
45285 }
vcvtps_s32_f32(float32_t __p0)45286 __ai int32_t vcvtps_s32_f32(float32_t __p0) {
45287   int32_t __ret;
45288   __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
45289   return __ret;
45290 }
45291 #ifdef __LITTLE_ENDIAN__
vcvtpq_s64_f64(float64x2_t __p0)45292 __ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
45293   int64x2_t __ret;
45294   __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__p0, 35);
45295   return __ret;
45296 }
45297 #else
vcvtpq_s64_f64(float64x2_t __p0)45298 __ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
45299   int64x2_t __ret;
45300   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45301   __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__rev0, 35);
45302   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45303   return __ret;
45304 }
45305 #endif
45306 
vcvtp_s64_f64(float64x1_t __p0)45307 __ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
45308   int64x1_t __ret;
45309   __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
45310   return __ret;
45311 }
vcvtpd_s64_f64(float64_t __p0)45312 __ai int64_t vcvtpd_s64_f64(float64_t __p0) {
45313   int64_t __ret;
45314   __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
45315   return __ret;
45316 }
vcvtps_u32_f32(float32_t __p0)45317 __ai uint32_t vcvtps_u32_f32(float32_t __p0) {
45318   uint32_t __ret;
45319   __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
45320   return __ret;
45321 }
45322 #ifdef __LITTLE_ENDIAN__
vcvtpq_u64_f64(float64x2_t __p0)45323 __ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
45324   uint64x2_t __ret;
45325   __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__p0, 51);
45326   return __ret;
45327 }
45328 #else
vcvtpq_u64_f64(float64x2_t __p0)45329 __ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
45330   uint64x2_t __ret;
45331   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45332   __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__rev0, 51);
45333   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45334   return __ret;
45335 }
45336 #endif
45337 
vcvtp_u64_f64(float64x1_t __p0)45338 __ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
45339   uint64x1_t __ret;
45340   __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
45341   return __ret;
45342 }
vcvtpd_u64_f64(float64_t __p0)45343 __ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
45344   uint64_t __ret;
45345   __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
45346   return __ret;
45347 }
vcvtxd_f32_f64(float64_t __p0)45348 __ai float32_t vcvtxd_f32_f64(float64_t __p0) {
45349   float32_t __ret;
45350   __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
45351   return __ret;
45352 }
45353 #ifdef __LITTLE_ENDIAN__
vcvtx_f32_f64(float64x2_t __p0)45354 __ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
45355   float32x2_t __ret;
45356   __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
45357   return __ret;
45358 }
45359 #else
vcvtx_f32_f64(float64x2_t __p0)45360 __ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
45361   float32x2_t __ret;
45362   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45363   __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__rev0, 42);
45364   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45365   return __ret;
45366 }
__noswap_vcvtx_f32_f64(float64x2_t __p0)45367 __ai float32x2_t __noswap_vcvtx_f32_f64(float64x2_t __p0) {
45368   float32x2_t __ret;
45369   __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
45370   return __ret;
45371 }
45372 #endif
45373 
45374 #ifdef __LITTLE_ENDIAN__
vcvtx_high_f32_f64(float32x2_t __p0,float64x2_t __p1)45375 __ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45376   float32x4_t __ret;
45377   __ret = vcombine_f32(__p0, vcvtx_f32_f64(__p1));
45378   return __ret;
45379 }
45380 #else
vcvtx_high_f32_f64(float32x2_t __p0,float64x2_t __p1)45381 __ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45382   float32x4_t __ret;
45383   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45384   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45385   __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvtx_f32_f64(__rev1));
45386   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45387   return __ret;
45388 }
45389 #endif
45390 
45391 #ifdef __LITTLE_ENDIAN__
vdivq_f64(float64x2_t __p0,float64x2_t __p1)45392 __ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
45393   float64x2_t __ret;
45394   __ret = __p0 / __p1;
45395   return __ret;
45396 }
45397 #else
vdivq_f64(float64x2_t __p0,float64x2_t __p1)45398 __ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
45399   float64x2_t __ret;
45400   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45401   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45402   __ret = __rev0 / __rev1;
45403   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45404   return __ret;
45405 }
45406 #endif
45407 
45408 #ifdef __LITTLE_ENDIAN__
vdivq_f32(float32x4_t __p0,float32x4_t __p1)45409 __ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
45410   float32x4_t __ret;
45411   __ret = __p0 / __p1;
45412   return __ret;
45413 }
45414 #else
vdivq_f32(float32x4_t __p0,float32x4_t __p1)45415 __ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
45416   float32x4_t __ret;
45417   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45418   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
45419   __ret = __rev0 / __rev1;
45420   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45421   return __ret;
45422 }
45423 #endif
45424 
vdiv_f64(float64x1_t __p0,float64x1_t __p1)45425 __ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
45426   float64x1_t __ret;
45427   __ret = __p0 / __p1;
45428   return __ret;
45429 }
45430 #ifdef __LITTLE_ENDIAN__
vdiv_f32(float32x2_t __p0,float32x2_t __p1)45431 __ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
45432   float32x2_t __ret;
45433   __ret = __p0 / __p1;
45434   return __ret;
45435 }
45436 #else
vdiv_f32(float32x2_t __p0,float32x2_t __p1)45437 __ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
45438   float32x2_t __ret;
45439   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45440   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45441   __ret = __rev0 / __rev1;
45442   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45443   return __ret;
45444 }
45445 #endif
45446 
45447 #ifdef __LITTLE_ENDIAN__
45448 #define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
45449   poly8_t __ret; \
45450   poly8x8_t __s0 = __p0; \
45451   __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((poly8x8_t)__s0, __p1); \
45452   __ret; \
45453 })
45454 #else
45455 #define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
45456   poly8_t __ret; \
45457   poly8x8_t __s0 = __p0; \
45458   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45459   __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((poly8x8_t)__rev0, __p1); \
45460   __ret; \
45461 })
45462 #endif
45463 
45464 #ifdef __LITTLE_ENDIAN__
45465 #define vduph_lane_p16(__p0, __p1) __extension__ ({ \
45466   poly16_t __ret; \
45467   poly16x4_t __s0 = __p0; \
45468   __ret = (poly16_t) __builtin_neon_vduph_lane_i16((poly16x4_t)__s0, __p1); \
45469   __ret; \
45470 })
45471 #else
45472 #define vduph_lane_p16(__p0, __p1) __extension__ ({ \
45473   poly16_t __ret; \
45474   poly16x4_t __s0 = __p0; \
45475   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45476   __ret = (poly16_t) __builtin_neon_vduph_lane_i16((poly16x4_t)__rev0, __p1); \
45477   __ret; \
45478 })
45479 #endif
45480 
45481 #ifdef __LITTLE_ENDIAN__
45482 #define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
45483   uint8_t __ret; \
45484   uint8x8_t __s0 = __p0; \
45485   __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
45486   __ret; \
45487 })
45488 #else
45489 #define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
45490   uint8_t __ret; \
45491   uint8x8_t __s0 = __p0; \
45492   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45493   __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
45494   __ret; \
45495 })
45496 #endif
45497 
45498 #ifdef __LITTLE_ENDIAN__
45499 #define vdups_lane_u32(__p0, __p1) __extension__ ({ \
45500   uint32_t __ret; \
45501   uint32x2_t __s0 = __p0; \
45502   __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int32x2_t)__s0, __p1); \
45503   __ret; \
45504 })
45505 #else
45506 #define vdups_lane_u32(__p0, __p1) __extension__ ({ \
45507   uint32_t __ret; \
45508   uint32x2_t __s0 = __p0; \
45509   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45510   __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int32x2_t)__rev0, __p1); \
45511   __ret; \
45512 })
45513 #endif
45514 
45515 #define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
45516   uint64_t __ret; \
45517   uint64x1_t __s0 = __p0; \
45518   __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int64x1_t)__s0, __p1); \
45519   __ret; \
45520 })
45521 #ifdef __LITTLE_ENDIAN__
45522 #define vduph_lane_u16(__p0, __p1) __extension__ ({ \
45523   uint16_t __ret; \
45524   uint16x4_t __s0 = __p0; \
45525   __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int16x4_t)__s0, __p1); \
45526   __ret; \
45527 })
45528 #else
45529 #define vduph_lane_u16(__p0, __p1) __extension__ ({ \
45530   uint16_t __ret; \
45531   uint16x4_t __s0 = __p0; \
45532   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45533   __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int16x4_t)__rev0, __p1); \
45534   __ret; \
45535 })
45536 #endif
45537 
45538 #ifdef __LITTLE_ENDIAN__
45539 #define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
45540   int8_t __ret; \
45541   int8x8_t __s0 = __p0; \
45542   __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
45543   __ret; \
45544 })
45545 #else
45546 #define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
45547   int8_t __ret; \
45548   int8x8_t __s0 = __p0; \
45549   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45550   __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
45551   __ret; \
45552 })
45553 #endif
45554 
45555 #define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
45556   float64_t __ret; \
45557   float64x1_t __s0 = __p0; \
45558   __ret = (float64_t) __builtin_neon_vdupd_lane_f64((float64x1_t)__s0, __p1); \
45559   __ret; \
45560 })
45561 #ifdef __LITTLE_ENDIAN__
45562 #define vdups_lane_f32(__p0, __p1) __extension__ ({ \
45563   float32_t __ret; \
45564   float32x2_t __s0 = __p0; \
45565   __ret = (float32_t) __builtin_neon_vdups_lane_f32((float32x2_t)__s0, __p1); \
45566   __ret; \
45567 })
45568 #else
45569 #define vdups_lane_f32(__p0, __p1) __extension__ ({ \
45570   float32_t __ret; \
45571   float32x2_t __s0 = __p0; \
45572   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45573   __ret = (float32_t) __builtin_neon_vdups_lane_f32((float32x2_t)__rev0, __p1); \
45574   __ret; \
45575 })
45576 #endif
45577 
45578 #ifdef __LITTLE_ENDIAN__
45579 #define vdups_lane_s32(__p0, __p1) __extension__ ({ \
45580   int32_t __ret; \
45581   int32x2_t __s0 = __p0; \
45582   __ret = (int32_t) __builtin_neon_vdups_lane_i32((int32x2_t)__s0, __p1); \
45583   __ret; \
45584 })
45585 #else
45586 #define vdups_lane_s32(__p0, __p1) __extension__ ({ \
45587   int32_t __ret; \
45588   int32x2_t __s0 = __p0; \
45589   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45590   __ret = (int32_t) __builtin_neon_vdups_lane_i32((int32x2_t)__rev0, __p1); \
45591   __ret; \
45592 })
45593 #endif
45594 
45595 #define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
45596   int64_t __ret; \
45597   int64x1_t __s0 = __p0; \
45598   __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int64x1_t)__s0, __p1); \
45599   __ret; \
45600 })
45601 #ifdef __LITTLE_ENDIAN__
45602 #define vduph_lane_s16(__p0, __p1) __extension__ ({ \
45603   int16_t __ret; \
45604   int16x4_t __s0 = __p0; \
45605   __ret = (int16_t) __builtin_neon_vduph_lane_i16((int16x4_t)__s0, __p1); \
45606   __ret; \
45607 })
45608 #else
45609 #define vduph_lane_s16(__p0, __p1) __extension__ ({ \
45610   int16_t __ret; \
45611   int16x4_t __s0 = __p0; \
45612   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45613   __ret = (int16_t) __builtin_neon_vduph_lane_i16((int16x4_t)__rev0, __p1); \
45614   __ret; \
45615 })
45616 #endif
45617 
45618 #define vdup_lane_p64(__p0_343, __p1_343) __extension__ ({ \
45619   poly64x1_t __ret_343; \
45620   poly64x1_t __s0_343 = __p0_343; \
45621   __ret_343 = splat_lane_p64(__s0_343, __p1_343); \
45622   __ret_343; \
45623 })
45624 #ifdef __LITTLE_ENDIAN__
45625 #define vdupq_lane_p64(__p0_344, __p1_344) __extension__ ({ \
45626   poly64x2_t __ret_344; \
45627   poly64x1_t __s0_344 = __p0_344; \
45628   __ret_344 = splatq_lane_p64(__s0_344, __p1_344); \
45629   __ret_344; \
45630 })
45631 #else
45632 #define vdupq_lane_p64(__p0_345, __p1_345) __extension__ ({ \
45633   poly64x2_t __ret_345; \
45634   poly64x1_t __s0_345 = __p0_345; \
45635   __ret_345 = __noswap_splatq_lane_p64(__s0_345, __p1_345); \
45636   __ret_345 = __builtin_shufflevector(__ret_345, __ret_345, 1, 0); \
45637   __ret_345; \
45638 })
45639 #endif
45640 
45641 #ifdef __LITTLE_ENDIAN__
45642 #define vdupq_lane_f64(__p0_346, __p1_346) __extension__ ({ \
45643   float64x2_t __ret_346; \
45644   float64x1_t __s0_346 = __p0_346; \
45645   __ret_346 = splatq_lane_f64(__s0_346, __p1_346); \
45646   __ret_346; \
45647 })
45648 #else
45649 #define vdupq_lane_f64(__p0_347, __p1_347) __extension__ ({ \
45650   float64x2_t __ret_347; \
45651   float64x1_t __s0_347 = __p0_347; \
45652   __ret_347 = __noswap_splatq_lane_f64(__s0_347, __p1_347); \
45653   __ret_347 = __builtin_shufflevector(__ret_347, __ret_347, 1, 0); \
45654   __ret_347; \
45655 })
45656 #endif
45657 
45658 #define vdup_lane_f64(__p0_348, __p1_348) __extension__ ({ \
45659   float64x1_t __ret_348; \
45660   float64x1_t __s0_348 = __p0_348; \
45661   __ret_348 = splat_lane_f64(__s0_348, __p1_348); \
45662   __ret_348; \
45663 })
45664 #ifdef __LITTLE_ENDIAN__
45665 #define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
45666   poly8_t __ret; \
45667   poly8x16_t __s0 = __p0; \
45668   __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((poly8x16_t)__s0, __p1); \
45669   __ret; \
45670 })
45671 #else
45672 #define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
45673   poly8_t __ret; \
45674   poly8x16_t __s0 = __p0; \
45675   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45676   __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((poly8x16_t)__rev0, __p1); \
45677   __ret; \
45678 })
45679 #endif
45680 
45681 #ifdef __LITTLE_ENDIAN__
45682 #define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
45683   poly16_t __ret; \
45684   poly16x8_t __s0 = __p0; \
45685   __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((poly16x8_t)__s0, __p1); \
45686   __ret; \
45687 })
45688 #else
45689 #define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
45690   poly16_t __ret; \
45691   poly16x8_t __s0 = __p0; \
45692   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45693   __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((poly16x8_t)__rev0, __p1); \
45694   __ret; \
45695 })
45696 #endif
45697 
45698 #ifdef __LITTLE_ENDIAN__
45699 #define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
45700   uint8_t __ret; \
45701   uint8x16_t __s0 = __p0; \
45702   __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
45703   __ret; \
45704 })
45705 #else
45706 #define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
45707   uint8_t __ret; \
45708   uint8x16_t __s0 = __p0; \
45709   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45710   __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
45711   __ret; \
45712 })
45713 #endif
45714 
45715 #ifdef __LITTLE_ENDIAN__
45716 #define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
45717   uint32_t __ret; \
45718   uint32x4_t __s0 = __p0; \
45719   __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int32x4_t)__s0, __p1); \
45720   __ret; \
45721 })
45722 #else
45723 #define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
45724   uint32_t __ret; \
45725   uint32x4_t __s0 = __p0; \
45726   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45727   __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int32x4_t)__rev0, __p1); \
45728   __ret; \
45729 })
45730 #endif
45731 
45732 #ifdef __LITTLE_ENDIAN__
45733 #define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
45734   uint64_t __ret; \
45735   uint64x2_t __s0 = __p0; \
45736   __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int64x2_t)__s0, __p1); \
45737   __ret; \
45738 })
45739 #else
45740 #define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
45741   uint64_t __ret; \
45742   uint64x2_t __s0 = __p0; \
45743   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45744   __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int64x2_t)__rev0, __p1); \
45745   __ret; \
45746 })
45747 #endif
45748 
45749 #ifdef __LITTLE_ENDIAN__
45750 #define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
45751   uint16_t __ret; \
45752   uint16x8_t __s0 = __p0; \
45753   __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int16x8_t)__s0, __p1); \
45754   __ret; \
45755 })
45756 #else
45757 #define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
45758   uint16_t __ret; \
45759   uint16x8_t __s0 = __p0; \
45760   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45761   __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int16x8_t)__rev0, __p1); \
45762   __ret; \
45763 })
45764 #endif
45765 
45766 #ifdef __LITTLE_ENDIAN__
45767 #define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
45768   int8_t __ret; \
45769   int8x16_t __s0 = __p0; \
45770   __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
45771   __ret; \
45772 })
45773 #else
45774 #define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
45775   int8_t __ret; \
45776   int8x16_t __s0 = __p0; \
45777   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45778   __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
45779   __ret; \
45780 })
45781 #endif
45782 
45783 #ifdef __LITTLE_ENDIAN__
45784 #define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
45785   float64_t __ret; \
45786   float64x2_t __s0 = __p0; \
45787   __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((float64x2_t)__s0, __p1); \
45788   __ret; \
45789 })
45790 #else
45791 #define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
45792   float64_t __ret; \
45793   float64x2_t __s0 = __p0; \
45794   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45795   __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((float64x2_t)__rev0, __p1); \
45796   __ret; \
45797 })
45798 #endif
45799 
45800 #ifdef __LITTLE_ENDIAN__
45801 #define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
45802   float32_t __ret; \
45803   float32x4_t __s0 = __p0; \
45804   __ret = (float32_t) __builtin_neon_vdups_laneq_f32((float32x4_t)__s0, __p1); \
45805   __ret; \
45806 })
45807 #else
45808 #define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
45809   float32_t __ret; \
45810   float32x4_t __s0 = __p0; \
45811   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45812   __ret = (float32_t) __builtin_neon_vdups_laneq_f32((float32x4_t)__rev0, __p1); \
45813   __ret; \
45814 })
45815 #endif
45816 
45817 #ifdef __LITTLE_ENDIAN__
45818 #define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
45819   int32_t __ret; \
45820   int32x4_t __s0 = __p0; \
45821   __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int32x4_t)__s0, __p1); \
45822   __ret; \
45823 })
45824 #else
45825 #define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
45826   int32_t __ret; \
45827   int32x4_t __s0 = __p0; \
45828   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45829   __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int32x4_t)__rev0, __p1); \
45830   __ret; \
45831 })
45832 #endif
45833 
45834 #ifdef __LITTLE_ENDIAN__
45835 #define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
45836   int64_t __ret; \
45837   int64x2_t __s0 = __p0; \
45838   __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int64x2_t)__s0, __p1); \
45839   __ret; \
45840 })
45841 #else
45842 #define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
45843   int64_t __ret; \
45844   int64x2_t __s0 = __p0; \
45845   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45846   __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int64x2_t)__rev0, __p1); \
45847   __ret; \
45848 })
45849 #endif
45850 
45851 #ifdef __LITTLE_ENDIAN__
45852 #define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
45853   int16_t __ret; \
45854   int16x8_t __s0 = __p0; \
45855   __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int16x8_t)__s0, __p1); \
45856   __ret; \
45857 })
45858 #else
45859 #define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
45860   int16_t __ret; \
45861   int16x8_t __s0 = __p0; \
45862   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45863   __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int16x8_t)__rev0, __p1); \
45864   __ret; \
45865 })
45866 #endif
45867 
45868 #ifdef __LITTLE_ENDIAN__
45869 #define vdup_laneq_p8(__p0_349, __p1_349) __extension__ ({ \
45870   poly8x8_t __ret_349; \
45871   poly8x16_t __s0_349 = __p0_349; \
45872   __ret_349 = splat_laneq_p8(__s0_349, __p1_349); \
45873   __ret_349; \
45874 })
45875 #else
45876 #define vdup_laneq_p8(__p0_350, __p1_350) __extension__ ({ \
45877   poly8x8_t __ret_350; \
45878   poly8x16_t __s0_350 = __p0_350; \
45879   poly8x16_t __rev0_350;  __rev0_350 = __builtin_shufflevector(__s0_350, __s0_350, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45880   __ret_350 = __noswap_splat_laneq_p8(__rev0_350, __p1_350); \
45881   __ret_350 = __builtin_shufflevector(__ret_350, __ret_350, 7, 6, 5, 4, 3, 2, 1, 0); \
45882   __ret_350; \
45883 })
45884 #endif
45885 
45886 #ifdef __LITTLE_ENDIAN__
45887 #define vdup_laneq_p64(__p0_351, __p1_351) __extension__ ({ \
45888   poly64x1_t __ret_351; \
45889   poly64x2_t __s0_351 = __p0_351; \
45890   __ret_351 = splat_laneq_p64(__s0_351, __p1_351); \
45891   __ret_351; \
45892 })
45893 #else
45894 #define vdup_laneq_p64(__p0_352, __p1_352) __extension__ ({ \
45895   poly64x1_t __ret_352; \
45896   poly64x2_t __s0_352 = __p0_352; \
45897   poly64x2_t __rev0_352;  __rev0_352 = __builtin_shufflevector(__s0_352, __s0_352, 1, 0); \
45898   __ret_352 = __noswap_splat_laneq_p64(__rev0_352, __p1_352); \
45899   __ret_352; \
45900 })
45901 #endif
45902 
45903 #ifdef __LITTLE_ENDIAN__
45904 #define vdup_laneq_p16(__p0_353, __p1_353) __extension__ ({ \
45905   poly16x4_t __ret_353; \
45906   poly16x8_t __s0_353 = __p0_353; \
45907   __ret_353 = splat_laneq_p16(__s0_353, __p1_353); \
45908   __ret_353; \
45909 })
45910 #else
45911 #define vdup_laneq_p16(__p0_354, __p1_354) __extension__ ({ \
45912   poly16x4_t __ret_354; \
45913   poly16x8_t __s0_354 = __p0_354; \
45914   poly16x8_t __rev0_354;  __rev0_354 = __builtin_shufflevector(__s0_354, __s0_354, 7, 6, 5, 4, 3, 2, 1, 0); \
45915   __ret_354 = __noswap_splat_laneq_p16(__rev0_354, __p1_354); \
45916   __ret_354 = __builtin_shufflevector(__ret_354, __ret_354, 3, 2, 1, 0); \
45917   __ret_354; \
45918 })
45919 #endif
45920 
45921 #ifdef __LITTLE_ENDIAN__
45922 #define vdupq_laneq_p8(__p0_355, __p1_355) __extension__ ({ \
45923   poly8x16_t __ret_355; \
45924   poly8x16_t __s0_355 = __p0_355; \
45925   __ret_355 = splatq_laneq_p8(__s0_355, __p1_355); \
45926   __ret_355; \
45927 })
45928 #else
45929 #define vdupq_laneq_p8(__p0_356, __p1_356) __extension__ ({ \
45930   poly8x16_t __ret_356; \
45931   poly8x16_t __s0_356 = __p0_356; \
45932   poly8x16_t __rev0_356;  __rev0_356 = __builtin_shufflevector(__s0_356, __s0_356, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45933   __ret_356 = __noswap_splatq_laneq_p8(__rev0_356, __p1_356); \
45934   __ret_356 = __builtin_shufflevector(__ret_356, __ret_356, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45935   __ret_356; \
45936 })
45937 #endif
45938 
45939 #ifdef __LITTLE_ENDIAN__
45940 #define vdupq_laneq_p64(__p0_357, __p1_357) __extension__ ({ \
45941   poly64x2_t __ret_357; \
45942   poly64x2_t __s0_357 = __p0_357; \
45943   __ret_357 = splatq_laneq_p64(__s0_357, __p1_357); \
45944   __ret_357; \
45945 })
45946 #else
45947 #define vdupq_laneq_p64(__p0_358, __p1_358) __extension__ ({ \
45948   poly64x2_t __ret_358; \
45949   poly64x2_t __s0_358 = __p0_358; \
45950   poly64x2_t __rev0_358;  __rev0_358 = __builtin_shufflevector(__s0_358, __s0_358, 1, 0); \
45951   __ret_358 = __noswap_splatq_laneq_p64(__rev0_358, __p1_358); \
45952   __ret_358 = __builtin_shufflevector(__ret_358, __ret_358, 1, 0); \
45953   __ret_358; \
45954 })
45955 #endif
45956 
45957 #ifdef __LITTLE_ENDIAN__
45958 #define vdupq_laneq_p16(__p0_359, __p1_359) __extension__ ({ \
45959   poly16x8_t __ret_359; \
45960   poly16x8_t __s0_359 = __p0_359; \
45961   __ret_359 = splatq_laneq_p16(__s0_359, __p1_359); \
45962   __ret_359; \
45963 })
45964 #else
45965 #define vdupq_laneq_p16(__p0_360, __p1_360) __extension__ ({ \
45966   poly16x8_t __ret_360; \
45967   poly16x8_t __s0_360 = __p0_360; \
45968   poly16x8_t __rev0_360;  __rev0_360 = __builtin_shufflevector(__s0_360, __s0_360, 7, 6, 5, 4, 3, 2, 1, 0); \
45969   __ret_360 = __noswap_splatq_laneq_p16(__rev0_360, __p1_360); \
45970   __ret_360 = __builtin_shufflevector(__ret_360, __ret_360, 7, 6, 5, 4, 3, 2, 1, 0); \
45971   __ret_360; \
45972 })
45973 #endif
45974 
45975 #ifdef __LITTLE_ENDIAN__
45976 #define vdupq_laneq_u8(__p0_361, __p1_361) __extension__ ({ \
45977   uint8x16_t __ret_361; \
45978   uint8x16_t __s0_361 = __p0_361; \
45979   __ret_361 = splatq_laneq_u8(__s0_361, __p1_361); \
45980   __ret_361; \
45981 })
45982 #else
45983 #define vdupq_laneq_u8(__p0_362, __p1_362) __extension__ ({ \
45984   uint8x16_t __ret_362; \
45985   uint8x16_t __s0_362 = __p0_362; \
45986   uint8x16_t __rev0_362;  __rev0_362 = __builtin_shufflevector(__s0_362, __s0_362, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45987   __ret_362 = __noswap_splatq_laneq_u8(__rev0_362, __p1_362); \
45988   __ret_362 = __builtin_shufflevector(__ret_362, __ret_362, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45989   __ret_362; \
45990 })
45991 #endif
45992 
45993 #ifdef __LITTLE_ENDIAN__
45994 #define vdupq_laneq_u32(__p0_363, __p1_363) __extension__ ({ \
45995   uint32x4_t __ret_363; \
45996   uint32x4_t __s0_363 = __p0_363; \
45997   __ret_363 = splatq_laneq_u32(__s0_363, __p1_363); \
45998   __ret_363; \
45999 })
46000 #else
46001 #define vdupq_laneq_u32(__p0_364, __p1_364) __extension__ ({ \
46002   uint32x4_t __ret_364; \
46003   uint32x4_t __s0_364 = __p0_364; \
46004   uint32x4_t __rev0_364;  __rev0_364 = __builtin_shufflevector(__s0_364, __s0_364, 3, 2, 1, 0); \
46005   __ret_364 = __noswap_splatq_laneq_u32(__rev0_364, __p1_364); \
46006   __ret_364 = __builtin_shufflevector(__ret_364, __ret_364, 3, 2, 1, 0); \
46007   __ret_364; \
46008 })
46009 #endif
46010 
46011 #ifdef __LITTLE_ENDIAN__
46012 #define vdupq_laneq_u64(__p0_365, __p1_365) __extension__ ({ \
46013   uint64x2_t __ret_365; \
46014   uint64x2_t __s0_365 = __p0_365; \
46015   __ret_365 = splatq_laneq_u64(__s0_365, __p1_365); \
46016   __ret_365; \
46017 })
46018 #else
46019 #define vdupq_laneq_u64(__p0_366, __p1_366) __extension__ ({ \
46020   uint64x2_t __ret_366; \
46021   uint64x2_t __s0_366 = __p0_366; \
46022   uint64x2_t __rev0_366;  __rev0_366 = __builtin_shufflevector(__s0_366, __s0_366, 1, 0); \
46023   __ret_366 = __noswap_splatq_laneq_u64(__rev0_366, __p1_366); \
46024   __ret_366 = __builtin_shufflevector(__ret_366, __ret_366, 1, 0); \
46025   __ret_366; \
46026 })
46027 #endif
46028 
46029 #ifdef __LITTLE_ENDIAN__
46030 #define vdupq_laneq_u16(__p0_367, __p1_367) __extension__ ({ \
46031   uint16x8_t __ret_367; \
46032   uint16x8_t __s0_367 = __p0_367; \
46033   __ret_367 = splatq_laneq_u16(__s0_367, __p1_367); \
46034   __ret_367; \
46035 })
46036 #else
46037 #define vdupq_laneq_u16(__p0_368, __p1_368) __extension__ ({ \
46038   uint16x8_t __ret_368; \
46039   uint16x8_t __s0_368 = __p0_368; \
46040   uint16x8_t __rev0_368;  __rev0_368 = __builtin_shufflevector(__s0_368, __s0_368, 7, 6, 5, 4, 3, 2, 1, 0); \
46041   __ret_368 = __noswap_splatq_laneq_u16(__rev0_368, __p1_368); \
46042   __ret_368 = __builtin_shufflevector(__ret_368, __ret_368, 7, 6, 5, 4, 3, 2, 1, 0); \
46043   __ret_368; \
46044 })
46045 #endif
46046 
46047 #ifdef __LITTLE_ENDIAN__
46048 #define vdupq_laneq_s8(__p0_369, __p1_369) __extension__ ({ \
46049   int8x16_t __ret_369; \
46050   int8x16_t __s0_369 = __p0_369; \
46051   __ret_369 = splatq_laneq_s8(__s0_369, __p1_369); \
46052   __ret_369; \
46053 })
46054 #else
46055 #define vdupq_laneq_s8(__p0_370, __p1_370) __extension__ ({ \
46056   int8x16_t __ret_370; \
46057   int8x16_t __s0_370 = __p0_370; \
46058   int8x16_t __rev0_370;  __rev0_370 = __builtin_shufflevector(__s0_370, __s0_370, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46059   __ret_370 = __noswap_splatq_laneq_s8(__rev0_370, __p1_370); \
46060   __ret_370 = __builtin_shufflevector(__ret_370, __ret_370, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46061   __ret_370; \
46062 })
46063 #endif
46064 
46065 #ifdef __LITTLE_ENDIAN__
46066 #define vdupq_laneq_f64(__p0_371, __p1_371) __extension__ ({ \
46067   float64x2_t __ret_371; \
46068   float64x2_t __s0_371 = __p0_371; \
46069   __ret_371 = splatq_laneq_f64(__s0_371, __p1_371); \
46070   __ret_371; \
46071 })
46072 #else
46073 #define vdupq_laneq_f64(__p0_372, __p1_372) __extension__ ({ \
46074   float64x2_t __ret_372; \
46075   float64x2_t __s0_372 = __p0_372; \
46076   float64x2_t __rev0_372;  __rev0_372 = __builtin_shufflevector(__s0_372, __s0_372, 1, 0); \
46077   __ret_372 = __noswap_splatq_laneq_f64(__rev0_372, __p1_372); \
46078   __ret_372 = __builtin_shufflevector(__ret_372, __ret_372, 1, 0); \
46079   __ret_372; \
46080 })
46081 #endif
46082 
46083 #ifdef __LITTLE_ENDIAN__
46084 #define vdupq_laneq_f32(__p0_373, __p1_373) __extension__ ({ \
46085   float32x4_t __ret_373; \
46086   float32x4_t __s0_373 = __p0_373; \
46087   __ret_373 = splatq_laneq_f32(__s0_373, __p1_373); \
46088   __ret_373; \
46089 })
46090 #else
46091 #define vdupq_laneq_f32(__p0_374, __p1_374) __extension__ ({ \
46092   float32x4_t __ret_374; \
46093   float32x4_t __s0_374 = __p0_374; \
46094   float32x4_t __rev0_374;  __rev0_374 = __builtin_shufflevector(__s0_374, __s0_374, 3, 2, 1, 0); \
46095   __ret_374 = __noswap_splatq_laneq_f32(__rev0_374, __p1_374); \
46096   __ret_374 = __builtin_shufflevector(__ret_374, __ret_374, 3, 2, 1, 0); \
46097   __ret_374; \
46098 })
46099 #endif
46100 
46101 #ifdef __LITTLE_ENDIAN__
46102 #define vdupq_laneq_f16(__p0_375, __p1_375) __extension__ ({ \
46103   float16x8_t __ret_375; \
46104   float16x8_t __s0_375 = __p0_375; \
46105   __ret_375 = splatq_laneq_f16(__s0_375, __p1_375); \
46106   __ret_375; \
46107 })
46108 #else
46109 #define vdupq_laneq_f16(__p0_376, __p1_376) __extension__ ({ \
46110   float16x8_t __ret_376; \
46111   float16x8_t __s0_376 = __p0_376; \
46112   float16x8_t __rev0_376;  __rev0_376 = __builtin_shufflevector(__s0_376, __s0_376, 7, 6, 5, 4, 3, 2, 1, 0); \
46113   __ret_376 = __noswap_splatq_laneq_f16(__rev0_376, __p1_376); \
46114   __ret_376 = __builtin_shufflevector(__ret_376, __ret_376, 7, 6, 5, 4, 3, 2, 1, 0); \
46115   __ret_376; \
46116 })
46117 #endif
46118 
46119 #ifdef __LITTLE_ENDIAN__
46120 #define vdupq_laneq_s32(__p0_377, __p1_377) __extension__ ({ \
46121   int32x4_t __ret_377; \
46122   int32x4_t __s0_377 = __p0_377; \
46123   __ret_377 = splatq_laneq_s32(__s0_377, __p1_377); \
46124   __ret_377; \
46125 })
46126 #else
46127 #define vdupq_laneq_s32(__p0_378, __p1_378) __extension__ ({ \
46128   int32x4_t __ret_378; \
46129   int32x4_t __s0_378 = __p0_378; \
46130   int32x4_t __rev0_378;  __rev0_378 = __builtin_shufflevector(__s0_378, __s0_378, 3, 2, 1, 0); \
46131   __ret_378 = __noswap_splatq_laneq_s32(__rev0_378, __p1_378); \
46132   __ret_378 = __builtin_shufflevector(__ret_378, __ret_378, 3, 2, 1, 0); \
46133   __ret_378; \
46134 })
46135 #endif
46136 
46137 #ifdef __LITTLE_ENDIAN__
46138 #define vdupq_laneq_s64(__p0_379, __p1_379) __extension__ ({ \
46139   int64x2_t __ret_379; \
46140   int64x2_t __s0_379 = __p0_379; \
46141   __ret_379 = splatq_laneq_s64(__s0_379, __p1_379); \
46142   __ret_379; \
46143 })
46144 #else
46145 #define vdupq_laneq_s64(__p0_380, __p1_380) __extension__ ({ \
46146   int64x2_t __ret_380; \
46147   int64x2_t __s0_380 = __p0_380; \
46148   int64x2_t __rev0_380;  __rev0_380 = __builtin_shufflevector(__s0_380, __s0_380, 1, 0); \
46149   __ret_380 = __noswap_splatq_laneq_s64(__rev0_380, __p1_380); \
46150   __ret_380 = __builtin_shufflevector(__ret_380, __ret_380, 1, 0); \
46151   __ret_380; \
46152 })
46153 #endif
46154 
46155 #ifdef __LITTLE_ENDIAN__
46156 #define vdupq_laneq_s16(__p0_381, __p1_381) __extension__ ({ \
46157   int16x8_t __ret_381; \
46158   int16x8_t __s0_381 = __p0_381; \
46159   __ret_381 = splatq_laneq_s16(__s0_381, __p1_381); \
46160   __ret_381; \
46161 })
46162 #else
46163 #define vdupq_laneq_s16(__p0_382, __p1_382) __extension__ ({ \
46164   int16x8_t __ret_382; \
46165   int16x8_t __s0_382 = __p0_382; \
46166   int16x8_t __rev0_382;  __rev0_382 = __builtin_shufflevector(__s0_382, __s0_382, 7, 6, 5, 4, 3, 2, 1, 0); \
46167   __ret_382 = __noswap_splatq_laneq_s16(__rev0_382, __p1_382); \
46168   __ret_382 = __builtin_shufflevector(__ret_382, __ret_382, 7, 6, 5, 4, 3, 2, 1, 0); \
46169   __ret_382; \
46170 })
46171 #endif
46172 
46173 #ifdef __LITTLE_ENDIAN__
46174 #define vdup_laneq_u8(__p0_383, __p1_383) __extension__ ({ \
46175   uint8x8_t __ret_383; \
46176   uint8x16_t __s0_383 = __p0_383; \
46177   __ret_383 = splat_laneq_u8(__s0_383, __p1_383); \
46178   __ret_383; \
46179 })
46180 #else
46181 #define vdup_laneq_u8(__p0_384, __p1_384) __extension__ ({ \
46182   uint8x8_t __ret_384; \
46183   uint8x16_t __s0_384 = __p0_384; \
46184   uint8x16_t __rev0_384;  __rev0_384 = __builtin_shufflevector(__s0_384, __s0_384, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46185   __ret_384 = __noswap_splat_laneq_u8(__rev0_384, __p1_384); \
46186   __ret_384 = __builtin_shufflevector(__ret_384, __ret_384, 7, 6, 5, 4, 3, 2, 1, 0); \
46187   __ret_384; \
46188 })
46189 #endif
46190 
46191 #ifdef __LITTLE_ENDIAN__
46192 #define vdup_laneq_u32(__p0_385, __p1_385) __extension__ ({ \
46193   uint32x2_t __ret_385; \
46194   uint32x4_t __s0_385 = __p0_385; \
46195   __ret_385 = splat_laneq_u32(__s0_385, __p1_385); \
46196   __ret_385; \
46197 })
46198 #else
46199 #define vdup_laneq_u32(__p0_386, __p1_386) __extension__ ({ \
46200   uint32x2_t __ret_386; \
46201   uint32x4_t __s0_386 = __p0_386; \
46202   uint32x4_t __rev0_386;  __rev0_386 = __builtin_shufflevector(__s0_386, __s0_386, 3, 2, 1, 0); \
46203   __ret_386 = __noswap_splat_laneq_u32(__rev0_386, __p1_386); \
46204   __ret_386 = __builtin_shufflevector(__ret_386, __ret_386, 1, 0); \
46205   __ret_386; \
46206 })
46207 #endif
46208 
46209 #ifdef __LITTLE_ENDIAN__
46210 #define vdup_laneq_u64(__p0_387, __p1_387) __extension__ ({ \
46211   uint64x1_t __ret_387; \
46212   uint64x2_t __s0_387 = __p0_387; \
46213   __ret_387 = splat_laneq_u64(__s0_387, __p1_387); \
46214   __ret_387; \
46215 })
46216 #else
46217 #define vdup_laneq_u64(__p0_388, __p1_388) __extension__ ({ \
46218   uint64x1_t __ret_388; \
46219   uint64x2_t __s0_388 = __p0_388; \
46220   uint64x2_t __rev0_388;  __rev0_388 = __builtin_shufflevector(__s0_388, __s0_388, 1, 0); \
46221   __ret_388 = __noswap_splat_laneq_u64(__rev0_388, __p1_388); \
46222   __ret_388; \
46223 })
46224 #endif
46225 
46226 #ifdef __LITTLE_ENDIAN__
46227 #define vdup_laneq_u16(__p0_389, __p1_389) __extension__ ({ \
46228   uint16x4_t __ret_389; \
46229   uint16x8_t __s0_389 = __p0_389; \
46230   __ret_389 = splat_laneq_u16(__s0_389, __p1_389); \
46231   __ret_389; \
46232 })
46233 #else
46234 #define vdup_laneq_u16(__p0_390, __p1_390) __extension__ ({ \
46235   uint16x4_t __ret_390; \
46236   uint16x8_t __s0_390 = __p0_390; \
46237   uint16x8_t __rev0_390;  __rev0_390 = __builtin_shufflevector(__s0_390, __s0_390, 7, 6, 5, 4, 3, 2, 1, 0); \
46238   __ret_390 = __noswap_splat_laneq_u16(__rev0_390, __p1_390); \
46239   __ret_390 = __builtin_shufflevector(__ret_390, __ret_390, 3, 2, 1, 0); \
46240   __ret_390; \
46241 })
46242 #endif
46243 
46244 #ifdef __LITTLE_ENDIAN__
46245 #define vdup_laneq_s8(__p0_391, __p1_391) __extension__ ({ \
46246   int8x8_t __ret_391; \
46247   int8x16_t __s0_391 = __p0_391; \
46248   __ret_391 = splat_laneq_s8(__s0_391, __p1_391); \
46249   __ret_391; \
46250 })
46251 #else
46252 #define vdup_laneq_s8(__p0_392, __p1_392) __extension__ ({ \
46253   int8x8_t __ret_392; \
46254   int8x16_t __s0_392 = __p0_392; \
46255   int8x16_t __rev0_392;  __rev0_392 = __builtin_shufflevector(__s0_392, __s0_392, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46256   __ret_392 = __noswap_splat_laneq_s8(__rev0_392, __p1_392); \
46257   __ret_392 = __builtin_shufflevector(__ret_392, __ret_392, 7, 6, 5, 4, 3, 2, 1, 0); \
46258   __ret_392; \
46259 })
46260 #endif
46261 
46262 #ifdef __LITTLE_ENDIAN__
46263 #define vdup_laneq_f64(__p0_393, __p1_393) __extension__ ({ \
46264   float64x1_t __ret_393; \
46265   float64x2_t __s0_393 = __p0_393; \
46266   __ret_393 = splat_laneq_f64(__s0_393, __p1_393); \
46267   __ret_393; \
46268 })
46269 #else
46270 #define vdup_laneq_f64(__p0_394, __p1_394) __extension__ ({ \
46271   float64x1_t __ret_394; \
46272   float64x2_t __s0_394 = __p0_394; \
46273   float64x2_t __rev0_394;  __rev0_394 = __builtin_shufflevector(__s0_394, __s0_394, 1, 0); \
46274   __ret_394 = __noswap_splat_laneq_f64(__rev0_394, __p1_394); \
46275   __ret_394; \
46276 })
46277 #endif
46278 
46279 #ifdef __LITTLE_ENDIAN__
46280 #define vdup_laneq_f32(__p0_395, __p1_395) __extension__ ({ \
46281   float32x2_t __ret_395; \
46282   float32x4_t __s0_395 = __p0_395; \
46283   __ret_395 = splat_laneq_f32(__s0_395, __p1_395); \
46284   __ret_395; \
46285 })
46286 #else
46287 #define vdup_laneq_f32(__p0_396, __p1_396) __extension__ ({ \
46288   float32x2_t __ret_396; \
46289   float32x4_t __s0_396 = __p0_396; \
46290   float32x4_t __rev0_396;  __rev0_396 = __builtin_shufflevector(__s0_396, __s0_396, 3, 2, 1, 0); \
46291   __ret_396 = __noswap_splat_laneq_f32(__rev0_396, __p1_396); \
46292   __ret_396 = __builtin_shufflevector(__ret_396, __ret_396, 1, 0); \
46293   __ret_396; \
46294 })
46295 #endif
46296 
46297 #ifdef __LITTLE_ENDIAN__
46298 #define vdup_laneq_f16(__p0_397, __p1_397) __extension__ ({ \
46299   float16x4_t __ret_397; \
46300   float16x8_t __s0_397 = __p0_397; \
46301   __ret_397 = splat_laneq_f16(__s0_397, __p1_397); \
46302   __ret_397; \
46303 })
46304 #else
46305 #define vdup_laneq_f16(__p0_398, __p1_398) __extension__ ({ \
46306   float16x4_t __ret_398; \
46307   float16x8_t __s0_398 = __p0_398; \
46308   float16x8_t __rev0_398;  __rev0_398 = __builtin_shufflevector(__s0_398, __s0_398, 7, 6, 5, 4, 3, 2, 1, 0); \
46309   __ret_398 = __noswap_splat_laneq_f16(__rev0_398, __p1_398); \
46310   __ret_398 = __builtin_shufflevector(__ret_398, __ret_398, 3, 2, 1, 0); \
46311   __ret_398; \
46312 })
46313 #endif
46314 
46315 #ifdef __LITTLE_ENDIAN__
46316 #define vdup_laneq_s32(__p0_399, __p1_399) __extension__ ({ \
46317   int32x2_t __ret_399; \
46318   int32x4_t __s0_399 = __p0_399; \
46319   __ret_399 = splat_laneq_s32(__s0_399, __p1_399); \
46320   __ret_399; \
46321 })
46322 #else
46323 #define vdup_laneq_s32(__p0_400, __p1_400) __extension__ ({ \
46324   int32x2_t __ret_400; \
46325   int32x4_t __s0_400 = __p0_400; \
46326   int32x4_t __rev0_400;  __rev0_400 = __builtin_shufflevector(__s0_400, __s0_400, 3, 2, 1, 0); \
46327   __ret_400 = __noswap_splat_laneq_s32(__rev0_400, __p1_400); \
46328   __ret_400 = __builtin_shufflevector(__ret_400, __ret_400, 1, 0); \
46329   __ret_400; \
46330 })
46331 #endif
46332 
46333 #ifdef __LITTLE_ENDIAN__
46334 #define vdup_laneq_s64(__p0_401, __p1_401) __extension__ ({ \
46335   int64x1_t __ret_401; \
46336   int64x2_t __s0_401 = __p0_401; \
46337   __ret_401 = splat_laneq_s64(__s0_401, __p1_401); \
46338   __ret_401; \
46339 })
46340 #else
46341 #define vdup_laneq_s64(__p0_402, __p1_402) __extension__ ({ \
46342   int64x1_t __ret_402; \
46343   int64x2_t __s0_402 = __p0_402; \
46344   int64x2_t __rev0_402;  __rev0_402 = __builtin_shufflevector(__s0_402, __s0_402, 1, 0); \
46345   __ret_402 = __noswap_splat_laneq_s64(__rev0_402, __p1_402); \
46346   __ret_402; \
46347 })
46348 #endif
46349 
46350 #ifdef __LITTLE_ENDIAN__
46351 #define vdup_laneq_s16(__p0_403, __p1_403) __extension__ ({ \
46352   int16x4_t __ret_403; \
46353   int16x8_t __s0_403 = __p0_403; \
46354   __ret_403 = splat_laneq_s16(__s0_403, __p1_403); \
46355   __ret_403; \
46356 })
46357 #else
46358 #define vdup_laneq_s16(__p0_404, __p1_404) __extension__ ({ \
46359   int16x4_t __ret_404; \
46360   int16x8_t __s0_404 = __p0_404; \
46361   int16x8_t __rev0_404;  __rev0_404 = __builtin_shufflevector(__s0_404, __s0_404, 7, 6, 5, 4, 3, 2, 1, 0); \
46362   __ret_404 = __noswap_splat_laneq_s16(__rev0_404, __p1_404); \
46363   __ret_404 = __builtin_shufflevector(__ret_404, __ret_404, 3, 2, 1, 0); \
46364   __ret_404; \
46365 })
46366 #endif
46367 
vdup_n_p64(poly64_t __p0)46368 __ai poly64x1_t vdup_n_p64(poly64_t __p0) {
46369   poly64x1_t __ret;
46370   __ret = (poly64x1_t) {__p0};
46371   return __ret;
46372 }
46373 #ifdef __LITTLE_ENDIAN__
vdupq_n_p64(poly64_t __p0)46374 __ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
46375   poly64x2_t __ret;
46376   __ret = (poly64x2_t) {__p0, __p0};
46377   return __ret;
46378 }
46379 #else
vdupq_n_p64(poly64_t __p0)46380 __ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
46381   poly64x2_t __ret;
46382   __ret = (poly64x2_t) {__p0, __p0};
46383   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46384   return __ret;
46385 }
46386 #endif
46387 
46388 #ifdef __LITTLE_ENDIAN__
vdupq_n_f64(float64_t __p0)46389 __ai float64x2_t vdupq_n_f64(float64_t __p0) {
46390   float64x2_t __ret;
46391   __ret = (float64x2_t) {__p0, __p0};
46392   return __ret;
46393 }
46394 #else
vdupq_n_f64(float64_t __p0)46395 __ai float64x2_t vdupq_n_f64(float64_t __p0) {
46396   float64x2_t __ret;
46397   __ret = (float64x2_t) {__p0, __p0};
46398   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46399   return __ret;
46400 }
46401 #endif
46402 
vdup_n_f64(float64_t __p0)46403 __ai float64x1_t vdup_n_f64(float64_t __p0) {
46404   float64x1_t __ret;
46405   __ret = (float64x1_t) {__p0};
46406   return __ret;
46407 }
46408 #define vext_p64(__p0, __p1, __p2) __extension__ ({ \
46409   poly64x1_t __ret; \
46410   poly64x1_t __s0 = __p0; \
46411   poly64x1_t __s1 = __p1; \
46412   __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
46413   __ret; \
46414 })
46415 #ifdef __LITTLE_ENDIAN__
46416 #define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
46417   poly64x2_t __ret; \
46418   poly64x2_t __s0 = __p0; \
46419   poly64x2_t __s1 = __p1; \
46420   __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
46421   __ret; \
46422 })
46423 #else
46424 #define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
46425   poly64x2_t __ret; \
46426   poly64x2_t __s0 = __p0; \
46427   poly64x2_t __s1 = __p1; \
46428   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46429   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46430   __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
46431   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46432   __ret; \
46433 })
46434 #endif
46435 
46436 #ifdef __LITTLE_ENDIAN__
46437 #define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
46438   float64x2_t __ret; \
46439   float64x2_t __s0 = __p0; \
46440   float64x2_t __s1 = __p1; \
46441   __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 42); \
46442   __ret; \
46443 })
46444 #else
46445 #define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
46446   float64x2_t __ret; \
46447   float64x2_t __s0 = __p0; \
46448   float64x2_t __s1 = __p1; \
46449   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46450   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46451   __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 42); \
46452   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46453   __ret; \
46454 })
46455 #endif
46456 
46457 #define vext_f64(__p0, __p1, __p2) __extension__ ({ \
46458   float64x1_t __ret; \
46459   float64x1_t __s0 = __p0; \
46460   float64x1_t __s1 = __p1; \
46461   __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
46462   __ret; \
46463 })
46464 #ifdef __LITTLE_ENDIAN__
vfmaq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)46465 __ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46466   float64x2_t __ret;
46467   __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
46468   return __ret;
46469 }
46470 #else
vfmaq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)46471 __ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46472   float64x2_t __ret;
46473   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46474   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46475   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
46476   __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
46477   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46478   return __ret;
46479 }
__noswap_vfmaq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)46480 __ai float64x2_t __noswap_vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46481   float64x2_t __ret;
46482   __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
46483   return __ret;
46484 }
46485 #endif
46486 
vfma_f64(float64x1_t __p0,float64x1_t __p1,float64x1_t __p2)46487 __ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
46488   float64x1_t __ret;
46489   __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
46490   return __ret;
46491 }
46492 #define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46493   float64_t __ret; \
46494   float64_t __s0 = __p0; \
46495   float64_t __s1 = __p1; \
46496   float64x1_t __s2 = __p2; \
46497   __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (float64x1_t)__s2, __p3); \
46498   __ret; \
46499 })
46500 #ifdef __LITTLE_ENDIAN__
46501 #define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46502   float32_t __ret; \
46503   float32_t __s0 = __p0; \
46504   float32_t __s1 = __p1; \
46505   float32x2_t __s2 = __p2; \
46506   __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (float32x2_t)__s2, __p3); \
46507   __ret; \
46508 })
46509 #else
46510 #define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46511   float32_t __ret; \
46512   float32_t __s0 = __p0; \
46513   float32_t __s1 = __p1; \
46514   float32x2_t __s2 = __p2; \
46515   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46516   __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (float32x2_t)__rev2, __p3); \
46517   __ret; \
46518 })
46519 #define __noswap_vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46520   float32_t __ret; \
46521   float32_t __s0 = __p0; \
46522   float32_t __s1 = __p1; \
46523   float32x2_t __s2 = __p2; \
46524   __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (float32x2_t)__s2, __p3); \
46525   __ret; \
46526 })
46527 #endif
46528 
46529 #ifdef __LITTLE_ENDIAN__
46530 #define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46531   float64x2_t __ret; \
46532   float64x2_t __s0 = __p0; \
46533   float64x2_t __s1 = __p1; \
46534   float64x1_t __s2 = __p2; \
46535   __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
46536   __ret; \
46537 })
46538 #else
46539 #define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46540   float64x2_t __ret; \
46541   float64x2_t __s0 = __p0; \
46542   float64x2_t __s1 = __p1; \
46543   float64x1_t __s2 = __p2; \
46544   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46545   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46546   __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__s2, __p3, 42); \
46547   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46548   __ret; \
46549 })
46550 #define __noswap_vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46551   float64x2_t __ret; \
46552   float64x2_t __s0 = __p0; \
46553   float64x2_t __s1 = __p1; \
46554   float64x1_t __s2 = __p2; \
46555   __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
46556   __ret; \
46557 })
46558 #endif
46559 
46560 #ifdef __LITTLE_ENDIAN__
46561 #define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46562   float32x4_t __ret; \
46563   float32x4_t __s0 = __p0; \
46564   float32x4_t __s1 = __p1; \
46565   float32x2_t __s2 = __p2; \
46566   __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
46567   __ret; \
46568 })
46569 #else
46570 #define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46571   float32x4_t __ret; \
46572   float32x4_t __s0 = __p0; \
46573   float32x4_t __s1 = __p1; \
46574   float32x2_t __s2 = __p2; \
46575   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46576   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
46577   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46578   __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, __p3, 41); \
46579   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46580   __ret; \
46581 })
46582 #define __noswap_vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46583   float32x4_t __ret; \
46584   float32x4_t __s0 = __p0; \
46585   float32x4_t __s1 = __p1; \
46586   float32x2_t __s2 = __p2; \
46587   __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
46588   __ret; \
46589 })
46590 #endif
46591 
46592 #define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46593   float64x1_t __ret; \
46594   float64x1_t __s0 = __p0; \
46595   float64x1_t __s1 = __p1; \
46596   float64x1_t __s2 = __p2; \
46597   __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
46598   __ret; \
46599 })
46600 #ifdef __LITTLE_ENDIAN__
46601 #define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46602   float32x2_t __ret; \
46603   float32x2_t __s0 = __p0; \
46604   float32x2_t __s1 = __p1; \
46605   float32x2_t __s2 = __p2; \
46606   __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
46607   __ret; \
46608 })
46609 #else
46610 #define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46611   float32x2_t __ret; \
46612   float32x2_t __s0 = __p0; \
46613   float32x2_t __s1 = __p1; \
46614   float32x2_t __s2 = __p2; \
46615   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46616   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46617   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46618   __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, __p3, 9); \
46619   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46620   __ret; \
46621 })
46622 #define __noswap_vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46623   float32x2_t __ret; \
46624   float32x2_t __s0 = __p0; \
46625   float32x2_t __s1 = __p1; \
46626   float32x2_t __s2 = __p2; \
46627   __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
46628   __ret; \
46629 })
46630 #endif
46631 
46632 #ifdef __LITTLE_ENDIAN__
46633 #define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46634   float64_t __ret; \
46635   float64_t __s0 = __p0; \
46636   float64_t __s1 = __p1; \
46637   float64x2_t __s2 = __p2; \
46638   __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (float64x2_t)__s2, __p3); \
46639   __ret; \
46640 })
46641 #else
46642 #define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46643   float64_t __ret; \
46644   float64_t __s0 = __p0; \
46645   float64_t __s1 = __p1; \
46646   float64x2_t __s2 = __p2; \
46647   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46648   __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (float64x2_t)__rev2, __p3); \
46649   __ret; \
46650 })
46651 #define __noswap_vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46652   float64_t __ret; \
46653   float64_t __s0 = __p0; \
46654   float64_t __s1 = __p1; \
46655   float64x2_t __s2 = __p2; \
46656   __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (float64x2_t)__s2, __p3); \
46657   __ret; \
46658 })
46659 #endif
46660 
46661 #ifdef __LITTLE_ENDIAN__
46662 #define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46663   float32_t __ret; \
46664   float32_t __s0 = __p0; \
46665   float32_t __s1 = __p1; \
46666   float32x4_t __s2 = __p2; \
46667   __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (float32x4_t)__s2, __p3); \
46668   __ret; \
46669 })
46670 #else
46671 #define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46672   float32_t __ret; \
46673   float32_t __s0 = __p0; \
46674   float32_t __s1 = __p1; \
46675   float32x4_t __s2 = __p2; \
46676   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
46677   __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (float32x4_t)__rev2, __p3); \
46678   __ret; \
46679 })
46680 #define __noswap_vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46681   float32_t __ret; \
46682   float32_t __s0 = __p0; \
46683   float32_t __s1 = __p1; \
46684   float32x4_t __s2 = __p2; \
46685   __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (float32x4_t)__s2, __p3); \
46686   __ret; \
46687 })
46688 #endif
46689 
46690 #ifdef __LITTLE_ENDIAN__
46691 #define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46692   float64x2_t __ret; \
46693   float64x2_t __s0 = __p0; \
46694   float64x2_t __s1 = __p1; \
46695   float64x2_t __s2 = __p2; \
46696   __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
46697   __ret; \
46698 })
46699 #else
46700 #define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46701   float64x2_t __ret; \
46702   float64x2_t __s0 = __p0; \
46703   float64x2_t __s1 = __p1; \
46704   float64x2_t __s2 = __p2; \
46705   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46706   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46707   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46708   __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 42); \
46709   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46710   __ret; \
46711 })
46712 #define __noswap_vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46713   float64x2_t __ret; \
46714   float64x2_t __s0 = __p0; \
46715   float64x2_t __s1 = __p1; \
46716   float64x2_t __s2 = __p2; \
46717   __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
46718   __ret; \
46719 })
46720 #endif
46721 
46722 #ifdef __LITTLE_ENDIAN__
46723 #define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46724   float32x4_t __ret; \
46725   float32x4_t __s0 = __p0; \
46726   float32x4_t __s1 = __p1; \
46727   float32x4_t __s2 = __p2; \
46728   __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
46729   __ret; \
46730 })
46731 #else
46732 #define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46733   float32x4_t __ret; \
46734   float32x4_t __s0 = __p0; \
46735   float32x4_t __s1 = __p1; \
46736   float32x4_t __s2 = __p2; \
46737   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46738   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
46739   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
46740   __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 41); \
46741   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46742   __ret; \
46743 })
46744 #define __noswap_vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46745   float32x4_t __ret; \
46746   float32x4_t __s0 = __p0; \
46747   float32x4_t __s1 = __p1; \
46748   float32x4_t __s2 = __p2; \
46749   __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
46750   __ret; \
46751 })
46752 #endif
46753 
46754 #ifdef __LITTLE_ENDIAN__
46755 #define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46756   float64x1_t __ret; \
46757   float64x1_t __s0 = __p0; \
46758   float64x1_t __s1 = __p1; \
46759   float64x2_t __s2 = __p2; \
46760   __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
46761   __ret; \
46762 })
46763 #else
46764 #define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46765   float64x1_t __ret; \
46766   float64x1_t __s0 = __p0; \
46767   float64x1_t __s1 = __p1; \
46768   float64x2_t __s2 = __p2; \
46769   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46770   __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__rev2, __p3, 10); \
46771   __ret; \
46772 })
46773 #define __noswap_vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46774   float64x1_t __ret; \
46775   float64x1_t __s0 = __p0; \
46776   float64x1_t __s1 = __p1; \
46777   float64x2_t __s2 = __p2; \
46778   __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
46779   __ret; \
46780 })
46781 #endif
46782 
46783 #ifdef __LITTLE_ENDIAN__
46784 #define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46785   float32x2_t __ret; \
46786   float32x2_t __s0 = __p0; \
46787   float32x2_t __s1 = __p1; \
46788   float32x4_t __s2 = __p2; \
46789   __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
46790   __ret; \
46791 })
46792 #else
46793 #define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46794   float32x2_t __ret; \
46795   float32x2_t __s0 = __p0; \
46796   float32x2_t __s1 = __p1; \
46797   float32x4_t __s2 = __p2; \
46798   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46799   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46800   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
46801   __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x16_t)__rev2, __p3, 9); \
46802   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46803   __ret; \
46804 })
46805 #define __noswap_vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46806   float32x2_t __ret; \
46807   float32x2_t __s0 = __p0; \
46808   float32x2_t __s1 = __p1; \
46809   float32x4_t __s2 = __p2; \
46810   __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
46811   __ret; \
46812 })
46813 #endif
46814 
46815 #ifdef __LITTLE_ENDIAN__
vfmaq_n_f64(float64x2_t __p0,float64x2_t __p1,float64_t __p2)46816 __ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
46817   float64x2_t __ret;
46818   __ret = vfmaq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
46819   return __ret;
46820 }
46821 #else
vfmaq_n_f64(float64x2_t __p0,float64x2_t __p1,float64_t __p2)46822 __ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
46823   float64x2_t __ret;
46824   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46825   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46826   __ret = __noswap_vfmaq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
46827   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46828   return __ret;
46829 }
46830 #endif
46831 
vfma_n_f64(float64x1_t __p0,float64x1_t __p1,float64_t __p2)46832 __ai float64x1_t vfma_n_f64(float64x1_t __p0, float64x1_t __p1, float64_t __p2) {
46833   float64x1_t __ret;
46834   __ret = vfma_f64(__p0, __p1, (float64x1_t) {__p2});
46835   return __ret;
46836 }
46837 #ifdef __LITTLE_ENDIAN__
vfmsq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)46838 __ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46839   float64x2_t __ret;
46840   __ret = vfmaq_f64(__p0, -__p1, __p2);
46841   return __ret;
46842 }
46843 #else
vfmsq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)46844 __ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46845   float64x2_t __ret;
46846   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46847   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46848   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
46849   __ret = __noswap_vfmaq_f64(__rev0, -__rev1, __rev2);
46850   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46851   return __ret;
46852 }
46853 #endif
46854 
vfms_f64(float64x1_t __p0,float64x1_t __p1,float64x1_t __p2)46855 __ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
46856   float64x1_t __ret;
46857   __ret = vfma_f64(__p0, -__p1, __p2);
46858   return __ret;
46859 }
46860 #define vfmsd_lane_f64(__p0_405, __p1_405, __p2_405, __p3_405) __extension__ ({ \
46861   float64_t __ret_405; \
46862   float64_t __s0_405 = __p0_405; \
46863   float64_t __s1_405 = __p1_405; \
46864   float64x1_t __s2_405 = __p2_405; \
46865   __ret_405 = vfmad_lane_f64(__s0_405, -__s1_405, __s2_405, __p3_405); \
46866   __ret_405; \
46867 })
46868 #ifdef __LITTLE_ENDIAN__
46869 #define vfmss_lane_f32(__p0_406, __p1_406, __p2_406, __p3_406) __extension__ ({ \
46870   float32_t __ret_406; \
46871   float32_t __s0_406 = __p0_406; \
46872   float32_t __s1_406 = __p1_406; \
46873   float32x2_t __s2_406 = __p2_406; \
46874   __ret_406 = vfmas_lane_f32(__s0_406, -__s1_406, __s2_406, __p3_406); \
46875   __ret_406; \
46876 })
46877 #else
46878 #define vfmss_lane_f32(__p0_407, __p1_407, __p2_407, __p3_407) __extension__ ({ \
46879   float32_t __ret_407; \
46880   float32_t __s0_407 = __p0_407; \
46881   float32_t __s1_407 = __p1_407; \
46882   float32x2_t __s2_407 = __p2_407; \
46883   float32x2_t __rev2_407;  __rev2_407 = __builtin_shufflevector(__s2_407, __s2_407, 1, 0); \
46884   __ret_407 = __noswap_vfmas_lane_f32(__s0_407, -__s1_407, __rev2_407, __p3_407); \
46885   __ret_407; \
46886 })
46887 #endif
46888 
46889 #ifdef __LITTLE_ENDIAN__
46890 #define vfmsq_lane_f64(__p0_408, __p1_408, __p2_408, __p3_408) __extension__ ({ \
46891   float64x2_t __ret_408; \
46892   float64x2_t __s0_408 = __p0_408; \
46893   float64x2_t __s1_408 = __p1_408; \
46894   float64x1_t __s2_408 = __p2_408; \
46895   __ret_408 = vfmaq_lane_f64(__s0_408, -__s1_408, __s2_408, __p3_408); \
46896   __ret_408; \
46897 })
46898 #else
46899 #define vfmsq_lane_f64(__p0_409, __p1_409, __p2_409, __p3_409) __extension__ ({ \
46900   float64x2_t __ret_409; \
46901   float64x2_t __s0_409 = __p0_409; \
46902   float64x2_t __s1_409 = __p1_409; \
46903   float64x1_t __s2_409 = __p2_409; \
46904   float64x2_t __rev0_409;  __rev0_409 = __builtin_shufflevector(__s0_409, __s0_409, 1, 0); \
46905   float64x2_t __rev1_409;  __rev1_409 = __builtin_shufflevector(__s1_409, __s1_409, 1, 0); \
46906   __ret_409 = __noswap_vfmaq_lane_f64(__rev0_409, -__rev1_409, __s2_409, __p3_409); \
46907   __ret_409 = __builtin_shufflevector(__ret_409, __ret_409, 1, 0); \
46908   __ret_409; \
46909 })
46910 #endif
46911 
46912 #ifdef __LITTLE_ENDIAN__
46913 #define vfmsq_lane_f32(__p0_410, __p1_410, __p2_410, __p3_410) __extension__ ({ \
46914   float32x4_t __ret_410; \
46915   float32x4_t __s0_410 = __p0_410; \
46916   float32x4_t __s1_410 = __p1_410; \
46917   float32x2_t __s2_410 = __p2_410; \
46918   __ret_410 = vfmaq_lane_f32(__s0_410, -__s1_410, __s2_410, __p3_410); \
46919   __ret_410; \
46920 })
46921 #else
46922 #define vfmsq_lane_f32(__p0_411, __p1_411, __p2_411, __p3_411) __extension__ ({ \
46923   float32x4_t __ret_411; \
46924   float32x4_t __s0_411 = __p0_411; \
46925   float32x4_t __s1_411 = __p1_411; \
46926   float32x2_t __s2_411 = __p2_411; \
46927   float32x4_t __rev0_411;  __rev0_411 = __builtin_shufflevector(__s0_411, __s0_411, 3, 2, 1, 0); \
46928   float32x4_t __rev1_411;  __rev1_411 = __builtin_shufflevector(__s1_411, __s1_411, 3, 2, 1, 0); \
46929   float32x2_t __rev2_411;  __rev2_411 = __builtin_shufflevector(__s2_411, __s2_411, 1, 0); \
46930   __ret_411 = __noswap_vfmaq_lane_f32(__rev0_411, -__rev1_411, __rev2_411, __p3_411); \
46931   __ret_411 = __builtin_shufflevector(__ret_411, __ret_411, 3, 2, 1, 0); \
46932   __ret_411; \
46933 })
46934 #endif
46935 
46936 #define vfms_lane_f64(__p0_412, __p1_412, __p2_412, __p3_412) __extension__ ({ \
46937   float64x1_t __ret_412; \
46938   float64x1_t __s0_412 = __p0_412; \
46939   float64x1_t __s1_412 = __p1_412; \
46940   float64x1_t __s2_412 = __p2_412; \
46941   __ret_412 = vfma_lane_f64(__s0_412, -__s1_412, __s2_412, __p3_412); \
46942   __ret_412; \
46943 })
46944 #ifdef __LITTLE_ENDIAN__
46945 #define vfms_lane_f32(__p0_413, __p1_413, __p2_413, __p3_413) __extension__ ({ \
46946   float32x2_t __ret_413; \
46947   float32x2_t __s0_413 = __p0_413; \
46948   float32x2_t __s1_413 = __p1_413; \
46949   float32x2_t __s2_413 = __p2_413; \
46950   __ret_413 = vfma_lane_f32(__s0_413, -__s1_413, __s2_413, __p3_413); \
46951   __ret_413; \
46952 })
46953 #else
46954 #define vfms_lane_f32(__p0_414, __p1_414, __p2_414, __p3_414) __extension__ ({ \
46955   float32x2_t __ret_414; \
46956   float32x2_t __s0_414 = __p0_414; \
46957   float32x2_t __s1_414 = __p1_414; \
46958   float32x2_t __s2_414 = __p2_414; \
46959   float32x2_t __rev0_414;  __rev0_414 = __builtin_shufflevector(__s0_414, __s0_414, 1, 0); \
46960   float32x2_t __rev1_414;  __rev1_414 = __builtin_shufflevector(__s1_414, __s1_414, 1, 0); \
46961   float32x2_t __rev2_414;  __rev2_414 = __builtin_shufflevector(__s2_414, __s2_414, 1, 0); \
46962   __ret_414 = __noswap_vfma_lane_f32(__rev0_414, -__rev1_414, __rev2_414, __p3_414); \
46963   __ret_414 = __builtin_shufflevector(__ret_414, __ret_414, 1, 0); \
46964   __ret_414; \
46965 })
46966 #endif
46967 
46968 #ifdef __LITTLE_ENDIAN__
46969 #define vfmsd_laneq_f64(__p0_415, __p1_415, __p2_415, __p3_415) __extension__ ({ \
46970   float64_t __ret_415; \
46971   float64_t __s0_415 = __p0_415; \
46972   float64_t __s1_415 = __p1_415; \
46973   float64x2_t __s2_415 = __p2_415; \
46974   __ret_415 = vfmad_laneq_f64(__s0_415, -__s1_415, __s2_415, __p3_415); \
46975   __ret_415; \
46976 })
46977 #else
46978 #define vfmsd_laneq_f64(__p0_416, __p1_416, __p2_416, __p3_416) __extension__ ({ \
46979   float64_t __ret_416; \
46980   float64_t __s0_416 = __p0_416; \
46981   float64_t __s1_416 = __p1_416; \
46982   float64x2_t __s2_416 = __p2_416; \
46983   float64x2_t __rev2_416;  __rev2_416 = __builtin_shufflevector(__s2_416, __s2_416, 1, 0); \
46984   __ret_416 = __noswap_vfmad_laneq_f64(__s0_416, -__s1_416, __rev2_416, __p3_416); \
46985   __ret_416; \
46986 })
46987 #endif
46988 
46989 #ifdef __LITTLE_ENDIAN__
46990 #define vfmss_laneq_f32(__p0_417, __p1_417, __p2_417, __p3_417) __extension__ ({ \
46991   float32_t __ret_417; \
46992   float32_t __s0_417 = __p0_417; \
46993   float32_t __s1_417 = __p1_417; \
46994   float32x4_t __s2_417 = __p2_417; \
46995   __ret_417 = vfmas_laneq_f32(__s0_417, -__s1_417, __s2_417, __p3_417); \
46996   __ret_417; \
46997 })
46998 #else
46999 #define vfmss_laneq_f32(__p0_418, __p1_418, __p2_418, __p3_418) __extension__ ({ \
47000   float32_t __ret_418; \
47001   float32_t __s0_418 = __p0_418; \
47002   float32_t __s1_418 = __p1_418; \
47003   float32x4_t __s2_418 = __p2_418; \
47004   float32x4_t __rev2_418;  __rev2_418 = __builtin_shufflevector(__s2_418, __s2_418, 3, 2, 1, 0); \
47005   __ret_418 = __noswap_vfmas_laneq_f32(__s0_418, -__s1_418, __rev2_418, __p3_418); \
47006   __ret_418; \
47007 })
47008 #endif
47009 
47010 #ifdef __LITTLE_ENDIAN__
47011 #define vfmsq_laneq_f64(__p0_419, __p1_419, __p2_419, __p3_419) __extension__ ({ \
47012   float64x2_t __ret_419; \
47013   float64x2_t __s0_419 = __p0_419; \
47014   float64x2_t __s1_419 = __p1_419; \
47015   float64x2_t __s2_419 = __p2_419; \
47016   __ret_419 = vfmaq_laneq_f64(__s0_419, -__s1_419, __s2_419, __p3_419); \
47017   __ret_419; \
47018 })
47019 #else
47020 #define vfmsq_laneq_f64(__p0_420, __p1_420, __p2_420, __p3_420) __extension__ ({ \
47021   float64x2_t __ret_420; \
47022   float64x2_t __s0_420 = __p0_420; \
47023   float64x2_t __s1_420 = __p1_420; \
47024   float64x2_t __s2_420 = __p2_420; \
47025   float64x2_t __rev0_420;  __rev0_420 = __builtin_shufflevector(__s0_420, __s0_420, 1, 0); \
47026   float64x2_t __rev1_420;  __rev1_420 = __builtin_shufflevector(__s1_420, __s1_420, 1, 0); \
47027   float64x2_t __rev2_420;  __rev2_420 = __builtin_shufflevector(__s2_420, __s2_420, 1, 0); \
47028   __ret_420 = __noswap_vfmaq_laneq_f64(__rev0_420, -__rev1_420, __rev2_420, __p3_420); \
47029   __ret_420 = __builtin_shufflevector(__ret_420, __ret_420, 1, 0); \
47030   __ret_420; \
47031 })
47032 #endif
47033 
47034 #ifdef __LITTLE_ENDIAN__
47035 #define vfmsq_laneq_f32(__p0_421, __p1_421, __p2_421, __p3_421) __extension__ ({ \
47036   float32x4_t __ret_421; \
47037   float32x4_t __s0_421 = __p0_421; \
47038   float32x4_t __s1_421 = __p1_421; \
47039   float32x4_t __s2_421 = __p2_421; \
47040   __ret_421 = vfmaq_laneq_f32(__s0_421, -__s1_421, __s2_421, __p3_421); \
47041   __ret_421; \
47042 })
47043 #else
47044 #define vfmsq_laneq_f32(__p0_422, __p1_422, __p2_422, __p3_422) __extension__ ({ \
47045   float32x4_t __ret_422; \
47046   float32x4_t __s0_422 = __p0_422; \
47047   float32x4_t __s1_422 = __p1_422; \
47048   float32x4_t __s2_422 = __p2_422; \
47049   float32x4_t __rev0_422;  __rev0_422 = __builtin_shufflevector(__s0_422, __s0_422, 3, 2, 1, 0); \
47050   float32x4_t __rev1_422;  __rev1_422 = __builtin_shufflevector(__s1_422, __s1_422, 3, 2, 1, 0); \
47051   float32x4_t __rev2_422;  __rev2_422 = __builtin_shufflevector(__s2_422, __s2_422, 3, 2, 1, 0); \
47052   __ret_422 = __noswap_vfmaq_laneq_f32(__rev0_422, -__rev1_422, __rev2_422, __p3_422); \
47053   __ret_422 = __builtin_shufflevector(__ret_422, __ret_422, 3, 2, 1, 0); \
47054   __ret_422; \
47055 })
47056 #endif
47057 
47058 #ifdef __LITTLE_ENDIAN__
47059 #define vfms_laneq_f64(__p0_423, __p1_423, __p2_423, __p3_423) __extension__ ({ \
47060   float64x1_t __ret_423; \
47061   float64x1_t __s0_423 = __p0_423; \
47062   float64x1_t __s1_423 = __p1_423; \
47063   float64x2_t __s2_423 = __p2_423; \
47064   __ret_423 = vfma_laneq_f64(__s0_423, -__s1_423, __s2_423, __p3_423); \
47065   __ret_423; \
47066 })
47067 #else
47068 #define vfms_laneq_f64(__p0_424, __p1_424, __p2_424, __p3_424) __extension__ ({ \
47069   float64x1_t __ret_424; \
47070   float64x1_t __s0_424 = __p0_424; \
47071   float64x1_t __s1_424 = __p1_424; \
47072   float64x2_t __s2_424 = __p2_424; \
47073   float64x2_t __rev2_424;  __rev2_424 = __builtin_shufflevector(__s2_424, __s2_424, 1, 0); \
47074   __ret_424 = __noswap_vfma_laneq_f64(__s0_424, -__s1_424, __rev2_424, __p3_424); \
47075   __ret_424; \
47076 })
47077 #endif
47078 
47079 #ifdef __LITTLE_ENDIAN__
47080 #define vfms_laneq_f32(__p0_425, __p1_425, __p2_425, __p3_425) __extension__ ({ \
47081   float32x2_t __ret_425; \
47082   float32x2_t __s0_425 = __p0_425; \
47083   float32x2_t __s1_425 = __p1_425; \
47084   float32x4_t __s2_425 = __p2_425; \
47085   __ret_425 = vfma_laneq_f32(__s0_425, -__s1_425, __s2_425, __p3_425); \
47086   __ret_425; \
47087 })
47088 #else
47089 #define vfms_laneq_f32(__p0_426, __p1_426, __p2_426, __p3_426) __extension__ ({ \
47090   float32x2_t __ret_426; \
47091   float32x2_t __s0_426 = __p0_426; \
47092   float32x2_t __s1_426 = __p1_426; \
47093   float32x4_t __s2_426 = __p2_426; \
47094   float32x2_t __rev0_426;  __rev0_426 = __builtin_shufflevector(__s0_426, __s0_426, 1, 0); \
47095   float32x2_t __rev1_426;  __rev1_426 = __builtin_shufflevector(__s1_426, __s1_426, 1, 0); \
47096   float32x4_t __rev2_426;  __rev2_426 = __builtin_shufflevector(__s2_426, __s2_426, 3, 2, 1, 0); \
47097   __ret_426 = __noswap_vfma_laneq_f32(__rev0_426, -__rev1_426, __rev2_426, __p3_426); \
47098   __ret_426 = __builtin_shufflevector(__ret_426, __ret_426, 1, 0); \
47099   __ret_426; \
47100 })
47101 #endif
47102 
47103 #ifdef __LITTLE_ENDIAN__
vfmsq_n_f64(float64x2_t __p0,float64x2_t __p1,float64_t __p2)47104 __ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
47105   float64x2_t __ret;
47106   __ret = vfmaq_f64(__p0, -__p1, (float64x2_t) {__p2, __p2});
47107   return __ret;
47108 }
47109 #else
vfmsq_n_f64(float64x2_t __p0,float64x2_t __p1,float64_t __p2)47110 __ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
47111   float64x2_t __ret;
47112   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47113   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47114   __ret = __noswap_vfmaq_f64(__rev0, -__rev1, (float64x2_t) {__p2, __p2});
47115   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47116   return __ret;
47117 }
47118 #endif
47119 
47120 #ifdef __LITTLE_ENDIAN__
vfmsq_n_f32(float32x4_t __p0,float32x4_t __p1,float32_t __p2)47121 __ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
47122   float32x4_t __ret;
47123   __ret = vfmaq_f32(__p0, -__p1, (float32x4_t) {__p2, __p2, __p2, __p2});
47124   return __ret;
47125 }
47126 #else
vfmsq_n_f32(float32x4_t __p0,float32x4_t __p1,float32_t __p2)47127 __ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
47128   float32x4_t __ret;
47129   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47130   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
47131   __ret = __noswap_vfmaq_f32(__rev0, -__rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
47132   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47133   return __ret;
47134 }
47135 #endif
47136 
vfms_n_f64(float64x1_t __p0,float64x1_t __p1,float64_t __p2)47137 __ai float64x1_t vfms_n_f64(float64x1_t __p0, float64x1_t __p1, float64_t __p2) {
47138   float64x1_t __ret;
47139   __ret = vfma_f64(__p0, -__p1, (float64x1_t) {__p2});
47140   return __ret;
47141 }
47142 #ifdef __LITTLE_ENDIAN__
vfms_n_f32(float32x2_t __p0,float32x2_t __p1,float32_t __p2)47143 __ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
47144   float32x2_t __ret;
47145   __ret = vfma_f32(__p0, -__p1, (float32x2_t) {__p2, __p2});
47146   return __ret;
47147 }
47148 #else
vfms_n_f32(float32x2_t __p0,float32x2_t __p1,float32_t __p2)47149 __ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
47150   float32x2_t __ret;
47151   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47152   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47153   __ret = __noswap_vfma_f32(__rev0, -__rev1, (float32x2_t) {__p2, __p2});
47154   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47155   return __ret;
47156 }
47157 #endif
47158 
47159 #ifdef __LITTLE_ENDIAN__
vget_high_p64(poly64x2_t __p0)47160 __ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
47161   poly64x1_t __ret;
47162   __ret = __builtin_shufflevector(__p0, __p0, 1);
47163   return __ret;
47164 }
47165 #else
vget_high_p64(poly64x2_t __p0)47166 __ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
47167   poly64x1_t __ret;
47168   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47169   __ret = __builtin_shufflevector(__rev0, __rev0, 1);
47170   return __ret;
47171 }
__noswap_vget_high_p64(poly64x2_t __p0)47172 __ai poly64x1_t __noswap_vget_high_p64(poly64x2_t __p0) {
47173   poly64x1_t __ret;
47174   __ret = __builtin_shufflevector(__p0, __p0, 1);
47175   return __ret;
47176 }
47177 #endif
47178 
47179 #ifdef __LITTLE_ENDIAN__
vget_high_f64(float64x2_t __p0)47180 __ai float64x1_t vget_high_f64(float64x2_t __p0) {
47181   float64x1_t __ret;
47182   __ret = __builtin_shufflevector(__p0, __p0, 1);
47183   return __ret;
47184 }
47185 #else
vget_high_f64(float64x2_t __p0)47186 __ai float64x1_t vget_high_f64(float64x2_t __p0) {
47187   float64x1_t __ret;
47188   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47189   __ret = __builtin_shufflevector(__rev0, __rev0, 1);
47190   return __ret;
47191 }
47192 #endif
47193 
47194 #define vget_lane_p64(__p0, __p1) __extension__ ({ \
47195   poly64_t __ret; \
47196   poly64x1_t __s0 = __p0; \
47197   __ret = (poly64_t) __builtin_neon_vget_lane_i64((poly64x1_t)__s0, __p1); \
47198   __ret; \
47199 })
47200 #ifdef __LITTLE_ENDIAN__
47201 #define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
47202   poly64_t __ret; \
47203   poly64x2_t __s0 = __p0; \
47204   __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((poly64x2_t)__s0, __p1); \
47205   __ret; \
47206 })
47207 #else
47208 #define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
47209   poly64_t __ret; \
47210   poly64x2_t __s0 = __p0; \
47211   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47212   __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((poly64x2_t)__rev0, __p1); \
47213   __ret; \
47214 })
47215 #define __noswap_vgetq_lane_p64(__p0, __p1) __extension__ ({ \
47216   poly64_t __ret; \
47217   poly64x2_t __s0 = __p0; \
47218   __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((poly64x2_t)__s0, __p1); \
47219   __ret; \
47220 })
47221 #endif
47222 
47223 #ifdef __LITTLE_ENDIAN__
47224 #define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
47225   float64_t __ret; \
47226   float64x2_t __s0 = __p0; \
47227   __ret = (float64_t) __builtin_neon_vgetq_lane_f64((float64x2_t)__s0, __p1); \
47228   __ret; \
47229 })
47230 #else
47231 #define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
47232   float64_t __ret; \
47233   float64x2_t __s0 = __p0; \
47234   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47235   __ret = (float64_t) __builtin_neon_vgetq_lane_f64((float64x2_t)__rev0, __p1); \
47236   __ret; \
47237 })
47238 #define __noswap_vgetq_lane_f64(__p0, __p1) __extension__ ({ \
47239   float64_t __ret; \
47240   float64x2_t __s0 = __p0; \
47241   __ret = (float64_t) __builtin_neon_vgetq_lane_f64((float64x2_t)__s0, __p1); \
47242   __ret; \
47243 })
47244 #endif
47245 
47246 #define vget_lane_f64(__p0, __p1) __extension__ ({ \
47247   float64_t __ret; \
47248   float64x1_t __s0 = __p0; \
47249   __ret = (float64_t) __builtin_neon_vget_lane_f64((float64x1_t)__s0, __p1); \
47250   __ret; \
47251 })
47252 #ifdef __LITTLE_ENDIAN__
vget_low_p64(poly64x2_t __p0)47253 __ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
47254   poly64x1_t __ret;
47255   __ret = __builtin_shufflevector(__p0, __p0, 0);
47256   return __ret;
47257 }
47258 #else
vget_low_p64(poly64x2_t __p0)47259 __ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
47260   poly64x1_t __ret;
47261   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47262   __ret = __builtin_shufflevector(__rev0, __rev0, 0);
47263   return __ret;
47264 }
47265 #endif
47266 
47267 #ifdef __LITTLE_ENDIAN__
vget_low_f64(float64x2_t __p0)47268 __ai float64x1_t vget_low_f64(float64x2_t __p0) {
47269   float64x1_t __ret;
47270   __ret = __builtin_shufflevector(__p0, __p0, 0);
47271   return __ret;
47272 }
47273 #else
vget_low_f64(float64x2_t __p0)47274 __ai float64x1_t vget_low_f64(float64x2_t __p0) {
47275   float64x1_t __ret;
47276   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47277   __ret = __builtin_shufflevector(__rev0, __rev0, 0);
47278   return __ret;
47279 }
47280 #endif
47281 
47282 #define vld1_p64(__p0) __extension__ ({ \
47283   poly64x1_t __ret; \
47284   __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
47285   __ret; \
47286 })
47287 #ifdef __LITTLE_ENDIAN__
47288 #define vld1q_p64(__p0) __extension__ ({ \
47289   poly64x2_t __ret; \
47290   __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
47291   __ret; \
47292 })
47293 #else
47294 #define vld1q_p64(__p0) __extension__ ({ \
47295   poly64x2_t __ret; \
47296   __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
47297   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47298   __ret; \
47299 })
47300 #endif
47301 
47302 #ifdef __LITTLE_ENDIAN__
47303 #define vld1q_f64(__p0) __extension__ ({ \
47304   float64x2_t __ret; \
47305   __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
47306   __ret; \
47307 })
47308 #else
47309 #define vld1q_f64(__p0) __extension__ ({ \
47310   float64x2_t __ret; \
47311   __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
47312   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47313   __ret; \
47314 })
47315 #endif
47316 
47317 #define vld1_f64(__p0) __extension__ ({ \
47318   float64x1_t __ret; \
47319   __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
47320   __ret; \
47321 })
47322 #define vld1_dup_p64(__p0) __extension__ ({ \
47323   poly64x1_t __ret; \
47324   __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
47325   __ret; \
47326 })
47327 #ifdef __LITTLE_ENDIAN__
47328 #define vld1q_dup_p64(__p0) __extension__ ({ \
47329   poly64x2_t __ret; \
47330   __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
47331   __ret; \
47332 })
47333 #else
47334 #define vld1q_dup_p64(__p0) __extension__ ({ \
47335   poly64x2_t __ret; \
47336   __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
47337   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47338   __ret; \
47339 })
47340 #endif
47341 
47342 #ifdef __LITTLE_ENDIAN__
47343 #define vld1q_dup_f64(__p0) __extension__ ({ \
47344   float64x2_t __ret; \
47345   __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
47346   __ret; \
47347 })
47348 #else
47349 #define vld1q_dup_f64(__p0) __extension__ ({ \
47350   float64x2_t __ret; \
47351   __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
47352   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47353   __ret; \
47354 })
47355 #endif
47356 
47357 #define vld1_dup_f64(__p0) __extension__ ({ \
47358   float64x1_t __ret; \
47359   __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
47360   __ret; \
47361 })
47362 #define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47363   poly64x1_t __ret; \
47364   poly64x1_t __s1 = __p1; \
47365   __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
47366   __ret; \
47367 })
47368 #ifdef __LITTLE_ENDIAN__
47369 #define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47370   poly64x2_t __ret; \
47371   poly64x2_t __s1 = __p1; \
47372   __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
47373   __ret; \
47374 })
47375 #else
47376 #define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47377   poly64x2_t __ret; \
47378   poly64x2_t __s1 = __p1; \
47379   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47380   __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
47381   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47382   __ret; \
47383 })
47384 #endif
47385 
47386 #ifdef __LITTLE_ENDIAN__
47387 #define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47388   float64x2_t __ret; \
47389   float64x2_t __s1 = __p1; \
47390   __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
47391   __ret; \
47392 })
47393 #else
47394 #define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47395   float64x2_t __ret; \
47396   float64x2_t __s1 = __p1; \
47397   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47398   __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
47399   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47400   __ret; \
47401 })
47402 #endif
47403 
47404 #define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47405   float64x1_t __ret; \
47406   float64x1_t __s1 = __p1; \
47407   __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
47408   __ret; \
47409 })
47410 #define vld1_p64_x2(__p0) __extension__ ({ \
47411   poly64x1x2_t __ret; \
47412   __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
47413   __ret; \
47414 })
47415 #ifdef __LITTLE_ENDIAN__
47416 #define vld1q_p64_x2(__p0) __extension__ ({ \
47417   poly64x2x2_t __ret; \
47418   __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
47419   __ret; \
47420 })
47421 #else
47422 #define vld1q_p64_x2(__p0) __extension__ ({ \
47423   poly64x2x2_t __ret; \
47424   __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
47425  \
47426   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47427   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47428   __ret; \
47429 })
47430 #endif
47431 
47432 #ifdef __LITTLE_ENDIAN__
47433 #define vld1q_f64_x2(__p0) __extension__ ({ \
47434   float64x2x2_t __ret; \
47435   __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
47436   __ret; \
47437 })
47438 #else
47439 #define vld1q_f64_x2(__p0) __extension__ ({ \
47440   float64x2x2_t __ret; \
47441   __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
47442  \
47443   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47444   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47445   __ret; \
47446 })
47447 #endif
47448 
47449 #define vld1_f64_x2(__p0) __extension__ ({ \
47450   float64x1x2_t __ret; \
47451   __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
47452   __ret; \
47453 })
47454 #define vld1_p64_x3(__p0) __extension__ ({ \
47455   poly64x1x3_t __ret; \
47456   __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
47457   __ret; \
47458 })
47459 #ifdef __LITTLE_ENDIAN__
47460 #define vld1q_p64_x3(__p0) __extension__ ({ \
47461   poly64x2x3_t __ret; \
47462   __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
47463   __ret; \
47464 })
47465 #else
47466 #define vld1q_p64_x3(__p0) __extension__ ({ \
47467   poly64x2x3_t __ret; \
47468   __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
47469  \
47470   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47471   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47472   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47473   __ret; \
47474 })
47475 #endif
47476 
47477 #ifdef __LITTLE_ENDIAN__
47478 #define vld1q_f64_x3(__p0) __extension__ ({ \
47479   float64x2x3_t __ret; \
47480   __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
47481   __ret; \
47482 })
47483 #else
47484 #define vld1q_f64_x3(__p0) __extension__ ({ \
47485   float64x2x3_t __ret; \
47486   __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
47487  \
47488   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47489   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47490   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47491   __ret; \
47492 })
47493 #endif
47494 
47495 #define vld1_f64_x3(__p0) __extension__ ({ \
47496   float64x1x3_t __ret; \
47497   __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
47498   __ret; \
47499 })
47500 #define vld1_p64_x4(__p0) __extension__ ({ \
47501   poly64x1x4_t __ret; \
47502   __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
47503   __ret; \
47504 })
47505 #ifdef __LITTLE_ENDIAN__
47506 #define vld1q_p64_x4(__p0) __extension__ ({ \
47507   poly64x2x4_t __ret; \
47508   __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
47509   __ret; \
47510 })
47511 #else
47512 #define vld1q_p64_x4(__p0) __extension__ ({ \
47513   poly64x2x4_t __ret; \
47514   __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
47515  \
47516   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47517   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47518   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47519   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
47520   __ret; \
47521 })
47522 #endif
47523 
47524 #ifdef __LITTLE_ENDIAN__
47525 #define vld1q_f64_x4(__p0) __extension__ ({ \
47526   float64x2x4_t __ret; \
47527   __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
47528   __ret; \
47529 })
47530 #else
47531 #define vld1q_f64_x4(__p0) __extension__ ({ \
47532   float64x2x4_t __ret; \
47533   __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
47534  \
47535   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47536   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47537   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47538   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
47539   __ret; \
47540 })
47541 #endif
47542 
47543 #define vld1_f64_x4(__p0) __extension__ ({ \
47544   float64x1x4_t __ret; \
47545   __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
47546   __ret; \
47547 })
47548 #define vld2_p64(__p0) __extension__ ({ \
47549   poly64x1x2_t __ret; \
47550   __builtin_neon_vld2_v(&__ret, __p0, 6); \
47551   __ret; \
47552 })
47553 #ifdef __LITTLE_ENDIAN__
47554 #define vld2q_p64(__p0) __extension__ ({ \
47555   poly64x2x2_t __ret; \
47556   __builtin_neon_vld2q_v(&__ret, __p0, 38); \
47557   __ret; \
47558 })
47559 #else
47560 #define vld2q_p64(__p0) __extension__ ({ \
47561   poly64x2x2_t __ret; \
47562   __builtin_neon_vld2q_v(&__ret, __p0, 38); \
47563  \
47564   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47565   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47566   __ret; \
47567 })
47568 #endif
47569 
47570 #ifdef __LITTLE_ENDIAN__
47571 #define vld2q_u64(__p0) __extension__ ({ \
47572   uint64x2x2_t __ret; \
47573   __builtin_neon_vld2q_v(&__ret, __p0, 51); \
47574   __ret; \
47575 })
47576 #else
47577 #define vld2q_u64(__p0) __extension__ ({ \
47578   uint64x2x2_t __ret; \
47579   __builtin_neon_vld2q_v(&__ret, __p0, 51); \
47580  \
47581   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47582   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47583   __ret; \
47584 })
47585 #endif
47586 
47587 #ifdef __LITTLE_ENDIAN__
47588 #define vld2q_f64(__p0) __extension__ ({ \
47589   float64x2x2_t __ret; \
47590   __builtin_neon_vld2q_v(&__ret, __p0, 42); \
47591   __ret; \
47592 })
47593 #else
47594 #define vld2q_f64(__p0) __extension__ ({ \
47595   float64x2x2_t __ret; \
47596   __builtin_neon_vld2q_v(&__ret, __p0, 42); \
47597  \
47598   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47599   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47600   __ret; \
47601 })
47602 #endif
47603 
47604 #ifdef __LITTLE_ENDIAN__
47605 #define vld2q_s64(__p0) __extension__ ({ \
47606   int64x2x2_t __ret; \
47607   __builtin_neon_vld2q_v(&__ret, __p0, 35); \
47608   __ret; \
47609 })
47610 #else
47611 #define vld2q_s64(__p0) __extension__ ({ \
47612   int64x2x2_t __ret; \
47613   __builtin_neon_vld2q_v(&__ret, __p0, 35); \
47614  \
47615   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47616   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47617   __ret; \
47618 })
47619 #endif
47620 
47621 #define vld2_f64(__p0) __extension__ ({ \
47622   float64x1x2_t __ret; \
47623   __builtin_neon_vld2_v(&__ret, __p0, 10); \
47624   __ret; \
47625 })
47626 #define vld2_dup_p64(__p0) __extension__ ({ \
47627   poly64x1x2_t __ret; \
47628   __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
47629   __ret; \
47630 })
47631 #ifdef __LITTLE_ENDIAN__
47632 #define vld2q_dup_p64(__p0) __extension__ ({ \
47633   poly64x2x2_t __ret; \
47634   __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
47635   __ret; \
47636 })
47637 #else
47638 #define vld2q_dup_p64(__p0) __extension__ ({ \
47639   poly64x2x2_t __ret; \
47640   __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
47641  \
47642   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47643   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47644   __ret; \
47645 })
47646 #endif
47647 
47648 #ifdef __LITTLE_ENDIAN__
47649 #define vld2q_dup_f64(__p0) __extension__ ({ \
47650   float64x2x2_t __ret; \
47651   __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
47652   __ret; \
47653 })
47654 #else
47655 #define vld2q_dup_f64(__p0) __extension__ ({ \
47656   float64x2x2_t __ret; \
47657   __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
47658  \
47659   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47660   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47661   __ret; \
47662 })
47663 #endif
47664 
47665 #define vld2_dup_f64(__p0) __extension__ ({ \
47666   float64x1x2_t __ret; \
47667   __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
47668   __ret; \
47669 })
47670 #define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47671   poly64x1x2_t __ret; \
47672   poly64x1x2_t __s1 = __p1; \
47673   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
47674   __ret; \
47675 })
47676 #ifdef __LITTLE_ENDIAN__
47677 #define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
47678   poly8x16x2_t __ret; \
47679   poly8x16x2_t __s1 = __p1; \
47680   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
47681   __ret; \
47682 })
47683 #else
47684 #define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
47685   poly8x16x2_t __ret; \
47686   poly8x16x2_t __s1 = __p1; \
47687   poly8x16x2_t __rev1; \
47688   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47689   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47690   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
47691  \
47692   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47693   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47694   __ret; \
47695 })
47696 #endif
47697 
47698 #ifdef __LITTLE_ENDIAN__
47699 #define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47700   poly64x2x2_t __ret; \
47701   poly64x2x2_t __s1 = __p1; \
47702   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
47703   __ret; \
47704 })
47705 #else
47706 #define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47707   poly64x2x2_t __ret; \
47708   poly64x2x2_t __s1 = __p1; \
47709   poly64x2x2_t __rev1; \
47710   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
47711   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
47712   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
47713  \
47714   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47715   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47716   __ret; \
47717 })
47718 #endif
47719 
47720 #ifdef __LITTLE_ENDIAN__
47721 #define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
47722   uint8x16x2_t __ret; \
47723   uint8x16x2_t __s1 = __p1; \
47724   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
47725   __ret; \
47726 })
47727 #else
47728 #define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
47729   uint8x16x2_t __ret; \
47730   uint8x16x2_t __s1 = __p1; \
47731   uint8x16x2_t __rev1; \
47732   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47733   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47734   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
47735  \
47736   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47737   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47738   __ret; \
47739 })
47740 #endif
47741 
47742 #ifdef __LITTLE_ENDIAN__
47743 #define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
47744   uint64x2x2_t __ret; \
47745   uint64x2x2_t __s1 = __p1; \
47746   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
47747   __ret; \
47748 })
47749 #else
47750 #define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
47751   uint64x2x2_t __ret; \
47752   uint64x2x2_t __s1 = __p1; \
47753   uint64x2x2_t __rev1; \
47754   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
47755   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
47756   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
47757  \
47758   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47759   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47760   __ret; \
47761 })
47762 #endif
47763 
47764 #ifdef __LITTLE_ENDIAN__
47765 #define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
47766   int8x16x2_t __ret; \
47767   int8x16x2_t __s1 = __p1; \
47768   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
47769   __ret; \
47770 })
47771 #else
47772 #define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
47773   int8x16x2_t __ret; \
47774   int8x16x2_t __s1 = __p1; \
47775   int8x16x2_t __rev1; \
47776   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47777   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47778   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
47779  \
47780   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47781   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47782   __ret; \
47783 })
47784 #endif
47785 
47786 #ifdef __LITTLE_ENDIAN__
47787 #define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47788   float64x2x2_t __ret; \
47789   float64x2x2_t __s1 = __p1; \
47790   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 42); \
47791   __ret; \
47792 })
47793 #else
47794 #define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47795   float64x2x2_t __ret; \
47796   float64x2x2_t __s1 = __p1; \
47797   float64x2x2_t __rev1; \
47798   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
47799   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
47800   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 42); \
47801  \
47802   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47803   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47804   __ret; \
47805 })
47806 #endif
47807 
47808 #ifdef __LITTLE_ENDIAN__
47809 #define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
47810   int64x2x2_t __ret; \
47811   int64x2x2_t __s1 = __p1; \
47812   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 35); \
47813   __ret; \
47814 })
47815 #else
47816 #define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
47817   int64x2x2_t __ret; \
47818   int64x2x2_t __s1 = __p1; \
47819   int64x2x2_t __rev1; \
47820   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
47821   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
47822   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 35); \
47823  \
47824   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47825   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47826   __ret; \
47827 })
47828 #endif
47829 
47830 #define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
47831   uint64x1x2_t __ret; \
47832   uint64x1x2_t __s1 = __p1; \
47833   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
47834   __ret; \
47835 })
47836 #define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47837   float64x1x2_t __ret; \
47838   float64x1x2_t __s1 = __p1; \
47839   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 10); \
47840   __ret; \
47841 })
47842 #define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
47843   int64x1x2_t __ret; \
47844   int64x1x2_t __s1 = __p1; \
47845   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 3); \
47846   __ret; \
47847 })
47848 #define vld3_p64(__p0) __extension__ ({ \
47849   poly64x1x3_t __ret; \
47850   __builtin_neon_vld3_v(&__ret, __p0, 6); \
47851   __ret; \
47852 })
47853 #ifdef __LITTLE_ENDIAN__
47854 #define vld3q_p64(__p0) __extension__ ({ \
47855   poly64x2x3_t __ret; \
47856   __builtin_neon_vld3q_v(&__ret, __p0, 38); \
47857   __ret; \
47858 })
47859 #else
47860 #define vld3q_p64(__p0) __extension__ ({ \
47861   poly64x2x3_t __ret; \
47862   __builtin_neon_vld3q_v(&__ret, __p0, 38); \
47863  \
47864   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47865   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47866   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47867   __ret; \
47868 })
47869 #endif
47870 
47871 #ifdef __LITTLE_ENDIAN__
47872 #define vld3q_u64(__p0) __extension__ ({ \
47873   uint64x2x3_t __ret; \
47874   __builtin_neon_vld3q_v(&__ret, __p0, 51); \
47875   __ret; \
47876 })
47877 #else
47878 #define vld3q_u64(__p0) __extension__ ({ \
47879   uint64x2x3_t __ret; \
47880   __builtin_neon_vld3q_v(&__ret, __p0, 51); \
47881  \
47882   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47883   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47884   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47885   __ret; \
47886 })
47887 #endif
47888 
47889 #ifdef __LITTLE_ENDIAN__
47890 #define vld3q_f64(__p0) __extension__ ({ \
47891   float64x2x3_t __ret; \
47892   __builtin_neon_vld3q_v(&__ret, __p0, 42); \
47893   __ret; \
47894 })
47895 #else
47896 #define vld3q_f64(__p0) __extension__ ({ \
47897   float64x2x3_t __ret; \
47898   __builtin_neon_vld3q_v(&__ret, __p0, 42); \
47899  \
47900   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47901   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47902   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47903   __ret; \
47904 })
47905 #endif
47906 
47907 #ifdef __LITTLE_ENDIAN__
47908 #define vld3q_s64(__p0) __extension__ ({ \
47909   int64x2x3_t __ret; \
47910   __builtin_neon_vld3q_v(&__ret, __p0, 35); \
47911   __ret; \
47912 })
47913 #else
47914 #define vld3q_s64(__p0) __extension__ ({ \
47915   int64x2x3_t __ret; \
47916   __builtin_neon_vld3q_v(&__ret, __p0, 35); \
47917  \
47918   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47919   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47920   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47921   __ret; \
47922 })
47923 #endif
47924 
47925 #define vld3_f64(__p0) __extension__ ({ \
47926   float64x1x3_t __ret; \
47927   __builtin_neon_vld3_v(&__ret, __p0, 10); \
47928   __ret; \
47929 })
47930 #define vld3_dup_p64(__p0) __extension__ ({ \
47931   poly64x1x3_t __ret; \
47932   __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
47933   __ret; \
47934 })
47935 #ifdef __LITTLE_ENDIAN__
47936 #define vld3q_dup_p64(__p0) __extension__ ({ \
47937   poly64x2x3_t __ret; \
47938   __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
47939   __ret; \
47940 })
47941 #else
47942 #define vld3q_dup_p64(__p0) __extension__ ({ \
47943   poly64x2x3_t __ret; \
47944   __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
47945  \
47946   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47947   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47948   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47949   __ret; \
47950 })
47951 #endif
47952 
47953 #ifdef __LITTLE_ENDIAN__
47954 #define vld3q_dup_f64(__p0) __extension__ ({ \
47955   float64x2x3_t __ret; \
47956   __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
47957   __ret; \
47958 })
47959 #else
47960 #define vld3q_dup_f64(__p0) __extension__ ({ \
47961   float64x2x3_t __ret; \
47962   __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
47963  \
47964   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47965   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47966   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47967   __ret; \
47968 })
47969 #endif
47970 
47971 #define vld3_dup_f64(__p0) __extension__ ({ \
47972   float64x1x3_t __ret; \
47973   __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
47974   __ret; \
47975 })
47976 #define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47977   poly64x1x3_t __ret; \
47978   poly64x1x3_t __s1 = __p1; \
47979   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
47980   __ret; \
47981 })
47982 #ifdef __LITTLE_ENDIAN__
47983 #define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
47984   poly8x16x3_t __ret; \
47985   poly8x16x3_t __s1 = __p1; \
47986   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
47987   __ret; \
47988 })
47989 #else
47990 #define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
47991   poly8x16x3_t __ret; \
47992   poly8x16x3_t __s1 = __p1; \
47993   poly8x16x3_t __rev1; \
47994   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47995   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47996   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47997   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
47998  \
47999   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48000   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48001   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48002   __ret; \
48003 })
48004 #endif
48005 
48006 #ifdef __LITTLE_ENDIAN__
48007 #define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48008   poly64x2x3_t __ret; \
48009   poly64x2x3_t __s1 = __p1; \
48010   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
48011   __ret; \
48012 })
48013 #else
48014 #define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48015   poly64x2x3_t __ret; \
48016   poly64x2x3_t __s1 = __p1; \
48017   poly64x2x3_t __rev1; \
48018   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
48019   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
48020   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
48021   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
48022  \
48023   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48024   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48025   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48026   __ret; \
48027 })
48028 #endif
48029 
48030 #ifdef __LITTLE_ENDIAN__
48031 #define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
48032   uint8x16x3_t __ret; \
48033   uint8x16x3_t __s1 = __p1; \
48034   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
48035   __ret; \
48036 })
48037 #else
48038 #define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
48039   uint8x16x3_t __ret; \
48040   uint8x16x3_t __s1 = __p1; \
48041   uint8x16x3_t __rev1; \
48042   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48043   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48044   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48045   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
48046  \
48047   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48048   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48049   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48050   __ret; \
48051 })
48052 #endif
48053 
48054 #ifdef __LITTLE_ENDIAN__
48055 #define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
48056   uint64x2x3_t __ret; \
48057   uint64x2x3_t __s1 = __p1; \
48058   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
48059   __ret; \
48060 })
48061 #else
48062 #define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
48063   uint64x2x3_t __ret; \
48064   uint64x2x3_t __s1 = __p1; \
48065   uint64x2x3_t __rev1; \
48066   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
48067   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
48068   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
48069   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
48070  \
48071   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48072   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48073   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48074   __ret; \
48075 })
48076 #endif
48077 
48078 #ifdef __LITTLE_ENDIAN__
48079 #define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
48080   int8x16x3_t __ret; \
48081   int8x16x3_t __s1 = __p1; \
48082   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
48083   __ret; \
48084 })
48085 #else
48086 #define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
48087   int8x16x3_t __ret; \
48088   int8x16x3_t __s1 = __p1; \
48089   int8x16x3_t __rev1; \
48090   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48091   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48092   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48093   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
48094  \
48095   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48096   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48097   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48098   __ret; \
48099 })
48100 #endif
48101 
48102 #ifdef __LITTLE_ENDIAN__
48103 #define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48104   float64x2x3_t __ret; \
48105   float64x2x3_t __s1 = __p1; \
48106   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 42); \
48107   __ret; \
48108 })
48109 #else
48110 #define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48111   float64x2x3_t __ret; \
48112   float64x2x3_t __s1 = __p1; \
48113   float64x2x3_t __rev1; \
48114   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
48115   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
48116   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
48117   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 42); \
48118  \
48119   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48120   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48121   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48122   __ret; \
48123 })
48124 #endif
48125 
48126 #ifdef __LITTLE_ENDIAN__
48127 #define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
48128   int64x2x3_t __ret; \
48129   int64x2x3_t __s1 = __p1; \
48130   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 35); \
48131   __ret; \
48132 })
48133 #else
48134 #define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
48135   int64x2x3_t __ret; \
48136   int64x2x3_t __s1 = __p1; \
48137   int64x2x3_t __rev1; \
48138   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
48139   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
48140   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
48141   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 35); \
48142  \
48143   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48144   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48145   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48146   __ret; \
48147 })
48148 #endif
48149 
48150 #define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
48151   uint64x1x3_t __ret; \
48152   uint64x1x3_t __s1 = __p1; \
48153   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
48154   __ret; \
48155 })
48156 #define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48157   float64x1x3_t __ret; \
48158   float64x1x3_t __s1 = __p1; \
48159   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 10); \
48160   __ret; \
48161 })
48162 #define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
48163   int64x1x3_t __ret; \
48164   int64x1x3_t __s1 = __p1; \
48165   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 3); \
48166   __ret; \
48167 })
48168 #define vld4_p64(__p0) __extension__ ({ \
48169   poly64x1x4_t __ret; \
48170   __builtin_neon_vld4_v(&__ret, __p0, 6); \
48171   __ret; \
48172 })
48173 #ifdef __LITTLE_ENDIAN__
48174 #define vld4q_p64(__p0) __extension__ ({ \
48175   poly64x2x4_t __ret; \
48176   __builtin_neon_vld4q_v(&__ret, __p0, 38); \
48177   __ret; \
48178 })
48179 #else
48180 #define vld4q_p64(__p0) __extension__ ({ \
48181   poly64x2x4_t __ret; \
48182   __builtin_neon_vld4q_v(&__ret, __p0, 38); \
48183  \
48184   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48185   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48186   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48187   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48188   __ret; \
48189 })
48190 #endif
48191 
48192 #ifdef __LITTLE_ENDIAN__
48193 #define vld4q_u64(__p0) __extension__ ({ \
48194   uint64x2x4_t __ret; \
48195   __builtin_neon_vld4q_v(&__ret, __p0, 51); \
48196   __ret; \
48197 })
48198 #else
48199 #define vld4q_u64(__p0) __extension__ ({ \
48200   uint64x2x4_t __ret; \
48201   __builtin_neon_vld4q_v(&__ret, __p0, 51); \
48202  \
48203   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48204   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48205   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48206   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48207   __ret; \
48208 })
48209 #endif
48210 
48211 #ifdef __LITTLE_ENDIAN__
48212 #define vld4q_f64(__p0) __extension__ ({ \
48213   float64x2x4_t __ret; \
48214   __builtin_neon_vld4q_v(&__ret, __p0, 42); \
48215   __ret; \
48216 })
48217 #else
48218 #define vld4q_f64(__p0) __extension__ ({ \
48219   float64x2x4_t __ret; \
48220   __builtin_neon_vld4q_v(&__ret, __p0, 42); \
48221  \
48222   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48223   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48224   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48225   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48226   __ret; \
48227 })
48228 #endif
48229 
48230 #ifdef __LITTLE_ENDIAN__
48231 #define vld4q_s64(__p0) __extension__ ({ \
48232   int64x2x4_t __ret; \
48233   __builtin_neon_vld4q_v(&__ret, __p0, 35); \
48234   __ret; \
48235 })
48236 #else
48237 #define vld4q_s64(__p0) __extension__ ({ \
48238   int64x2x4_t __ret; \
48239   __builtin_neon_vld4q_v(&__ret, __p0, 35); \
48240  \
48241   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48242   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48243   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48244   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48245   __ret; \
48246 })
48247 #endif
48248 
48249 #define vld4_f64(__p0) __extension__ ({ \
48250   float64x1x4_t __ret; \
48251   __builtin_neon_vld4_v(&__ret, __p0, 10); \
48252   __ret; \
48253 })
48254 #define vld4_dup_p64(__p0) __extension__ ({ \
48255   poly64x1x4_t __ret; \
48256   __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
48257   __ret; \
48258 })
48259 #ifdef __LITTLE_ENDIAN__
48260 #define vld4q_dup_p64(__p0) __extension__ ({ \
48261   poly64x2x4_t __ret; \
48262   __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
48263   __ret; \
48264 })
48265 #else
48266 #define vld4q_dup_p64(__p0) __extension__ ({ \
48267   poly64x2x4_t __ret; \
48268   __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
48269  \
48270   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48271   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48272   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48273   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48274   __ret; \
48275 })
48276 #endif
48277 
48278 #ifdef __LITTLE_ENDIAN__
48279 #define vld4q_dup_f64(__p0) __extension__ ({ \
48280   float64x2x4_t __ret; \
48281   __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
48282   __ret; \
48283 })
48284 #else
48285 #define vld4q_dup_f64(__p0) __extension__ ({ \
48286   float64x2x4_t __ret; \
48287   __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
48288  \
48289   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48290   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48291   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48292   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48293   __ret; \
48294 })
48295 #endif
48296 
48297 #define vld4_dup_f64(__p0) __extension__ ({ \
48298   float64x1x4_t __ret; \
48299   __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
48300   __ret; \
48301 })
48302 #define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48303   poly64x1x4_t __ret; \
48304   poly64x1x4_t __s1 = __p1; \
48305   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
48306   __ret; \
48307 })
48308 #ifdef __LITTLE_ENDIAN__
48309 #define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
48310   poly8x16x4_t __ret; \
48311   poly8x16x4_t __s1 = __p1; \
48312   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
48313   __ret; \
48314 })
48315 #else
48316 #define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
48317   poly8x16x4_t __ret; \
48318   poly8x16x4_t __s1 = __p1; \
48319   poly8x16x4_t __rev1; \
48320   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48321   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48322   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48323   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48324   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
48325  \
48326   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48327   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48328   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48329   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48330   __ret; \
48331 })
48332 #endif
48333 
48334 #ifdef __LITTLE_ENDIAN__
48335 #define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48336   poly64x2x4_t __ret; \
48337   poly64x2x4_t __s1 = __p1; \
48338   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
48339   __ret; \
48340 })
48341 #else
48342 #define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48343   poly64x2x4_t __ret; \
48344   poly64x2x4_t __s1 = __p1; \
48345   poly64x2x4_t __rev1; \
48346   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
48347   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
48348   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
48349   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
48350   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
48351  \
48352   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48353   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48354   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48355   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48356   __ret; \
48357 })
48358 #endif
48359 
48360 #ifdef __LITTLE_ENDIAN__
48361 #define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
48362   uint8x16x4_t __ret; \
48363   uint8x16x4_t __s1 = __p1; \
48364   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
48365   __ret; \
48366 })
48367 #else
48368 #define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
48369   uint8x16x4_t __ret; \
48370   uint8x16x4_t __s1 = __p1; \
48371   uint8x16x4_t __rev1; \
48372   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48373   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48374   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48375   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48376   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
48377  \
48378   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48379   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48380   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48381   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48382   __ret; \
48383 })
48384 #endif
48385 
48386 #ifdef __LITTLE_ENDIAN__
48387 #define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
48388   uint64x2x4_t __ret; \
48389   uint64x2x4_t __s1 = __p1; \
48390   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
48391   __ret; \
48392 })
48393 #else
48394 #define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
48395   uint64x2x4_t __ret; \
48396   uint64x2x4_t __s1 = __p1; \
48397   uint64x2x4_t __rev1; \
48398   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
48399   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
48400   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
48401   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
48402   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
48403  \
48404   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48405   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48406   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48407   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48408   __ret; \
48409 })
48410 #endif
48411 
48412 #ifdef __LITTLE_ENDIAN__
48413 #define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
48414   int8x16x4_t __ret; \
48415   int8x16x4_t __s1 = __p1; \
48416   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
48417   __ret; \
48418 })
48419 #else
48420 #define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
48421   int8x16x4_t __ret; \
48422   int8x16x4_t __s1 = __p1; \
48423   int8x16x4_t __rev1; \
48424   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48425   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48426   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48427   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48428   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
48429  \
48430   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48431   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48432   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48433   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48434   __ret; \
48435 })
48436 #endif
48437 
48438 #ifdef __LITTLE_ENDIAN__
48439 #define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48440   float64x2x4_t __ret; \
48441   float64x2x4_t __s1 = __p1; \
48442   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 42); \
48443   __ret; \
48444 })
48445 #else
48446 #define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48447   float64x2x4_t __ret; \
48448   float64x2x4_t __s1 = __p1; \
48449   float64x2x4_t __rev1; \
48450   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
48451   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
48452   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
48453   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
48454   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 42); \
48455  \
48456   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48457   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48458   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48459   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48460   __ret; \
48461 })
48462 #endif
48463 
48464 #ifdef __LITTLE_ENDIAN__
48465 #define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
48466   int64x2x4_t __ret; \
48467   int64x2x4_t __s1 = __p1; \
48468   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 35); \
48469   __ret; \
48470 })
48471 #else
48472 #define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
48473   int64x2x4_t __ret; \
48474   int64x2x4_t __s1 = __p1; \
48475   int64x2x4_t __rev1; \
48476   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
48477   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
48478   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
48479   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
48480   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 35); \
48481  \
48482   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48483   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48484   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48485   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48486   __ret; \
48487 })
48488 #endif
48489 
48490 #define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
48491   uint64x1x4_t __ret; \
48492   uint64x1x4_t __s1 = __p1; \
48493   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
48494   __ret; \
48495 })
48496 #define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48497   float64x1x4_t __ret; \
48498   float64x1x4_t __s1 = __p1; \
48499   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 10); \
48500   __ret; \
48501 })
48502 #define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
48503   int64x1x4_t __ret; \
48504   int64x1x4_t __s1 = __p1; \
48505   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 3); \
48506   __ret; \
48507 })
48508 #define vldrq_p128(__p0) __extension__ ({ \
48509   poly128_t __ret; \
48510   __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
48511   __ret; \
48512 })
48513 #ifdef __LITTLE_ENDIAN__
vmaxq_f64(float64x2_t __p0,float64x2_t __p1)48514 __ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
48515   float64x2_t __ret;
48516   __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
48517   return __ret;
48518 }
48519 #else
vmaxq_f64(float64x2_t __p0,float64x2_t __p1)48520 __ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
48521   float64x2_t __ret;
48522   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48523   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48524   __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
48525   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48526   return __ret;
48527 }
48528 #endif
48529 
vmax_f64(float64x1_t __p0,float64x1_t __p1)48530 __ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
48531   float64x1_t __ret;
48532   __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
48533   return __ret;
48534 }
48535 #ifdef __LITTLE_ENDIAN__
vmaxnmvq_f64(float64x2_t __p0)48536 __ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
48537   float64_t __ret;
48538   __ret = (float64_t) __builtin_neon_vmaxnmvq_f64(__p0);
48539   return __ret;
48540 }
48541 #else
vmaxnmvq_f64(float64x2_t __p0)48542 __ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
48543   float64_t __ret;
48544   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48545   __ret = (float64_t) __builtin_neon_vmaxnmvq_f64(__rev0);
48546   return __ret;
48547 }
48548 #endif
48549 
48550 #ifdef __LITTLE_ENDIAN__
vmaxnmvq_f32(float32x4_t __p0)48551 __ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
48552   float32_t __ret;
48553   __ret = (float32_t) __builtin_neon_vmaxnmvq_f32(__p0);
48554   return __ret;
48555 }
48556 #else
vmaxnmvq_f32(float32x4_t __p0)48557 __ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
48558   float32_t __ret;
48559   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48560   __ret = (float32_t) __builtin_neon_vmaxnmvq_f32(__rev0);
48561   return __ret;
48562 }
48563 #endif
48564 
48565 #ifdef __LITTLE_ENDIAN__
vmaxnmv_f32(float32x2_t __p0)48566 __ai float32_t vmaxnmv_f32(float32x2_t __p0) {
48567   float32_t __ret;
48568   __ret = (float32_t) __builtin_neon_vmaxnmv_f32(__p0);
48569   return __ret;
48570 }
48571 #else
vmaxnmv_f32(float32x2_t __p0)48572 __ai float32_t vmaxnmv_f32(float32x2_t __p0) {
48573   float32_t __ret;
48574   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48575   __ret = (float32_t) __builtin_neon_vmaxnmv_f32(__rev0);
48576   return __ret;
48577 }
48578 #endif
48579 
48580 #ifdef __LITTLE_ENDIAN__
vmaxvq_u8(uint8x16_t __p0)48581 __ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
48582   uint8_t __ret;
48583   __ret = (uint8_t) __builtin_neon_vmaxvq_u8(__p0);
48584   return __ret;
48585 }
48586 #else
vmaxvq_u8(uint8x16_t __p0)48587 __ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
48588   uint8_t __ret;
48589   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
48590   __ret = (uint8_t) __builtin_neon_vmaxvq_u8(__rev0);
48591   return __ret;
48592 }
48593 #endif
48594 
48595 #ifdef __LITTLE_ENDIAN__
vmaxvq_u32(uint32x4_t __p0)48596 __ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
48597   uint32_t __ret;
48598   __ret = (uint32_t) __builtin_neon_vmaxvq_u32(__p0);
48599   return __ret;
48600 }
48601 #else
vmaxvq_u32(uint32x4_t __p0)48602 __ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
48603   uint32_t __ret;
48604   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48605   __ret = (uint32_t) __builtin_neon_vmaxvq_u32(__rev0);
48606   return __ret;
48607 }
48608 #endif
48609 
48610 #ifdef __LITTLE_ENDIAN__
vmaxvq_u16(uint16x8_t __p0)48611 __ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
48612   uint16_t __ret;
48613   __ret = (uint16_t) __builtin_neon_vmaxvq_u16(__p0);
48614   return __ret;
48615 }
48616 #else
vmaxvq_u16(uint16x8_t __p0)48617 __ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
48618   uint16_t __ret;
48619   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
48620   __ret = (uint16_t) __builtin_neon_vmaxvq_u16(__rev0);
48621   return __ret;
48622 }
48623 #endif
48624 
48625 #ifdef __LITTLE_ENDIAN__
vmaxvq_s8(int8x16_t __p0)48626 __ai int8_t vmaxvq_s8(int8x16_t __p0) {
48627   int8_t __ret;
48628   __ret = (int8_t) __builtin_neon_vmaxvq_s8(__p0);
48629   return __ret;
48630 }
48631 #else
vmaxvq_s8(int8x16_t __p0)48632 __ai int8_t vmaxvq_s8(int8x16_t __p0) {
48633   int8_t __ret;
48634   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
48635   __ret = (int8_t) __builtin_neon_vmaxvq_s8(__rev0);
48636   return __ret;
48637 }
48638 #endif
48639 
48640 #ifdef __LITTLE_ENDIAN__
vmaxvq_f64(float64x2_t __p0)48641 __ai float64_t vmaxvq_f64(float64x2_t __p0) {
48642   float64_t __ret;
48643   __ret = (float64_t) __builtin_neon_vmaxvq_f64(__p0);
48644   return __ret;
48645 }
48646 #else
vmaxvq_f64(float64x2_t __p0)48647 __ai float64_t vmaxvq_f64(float64x2_t __p0) {
48648   float64_t __ret;
48649   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48650   __ret = (float64_t) __builtin_neon_vmaxvq_f64(__rev0);
48651   return __ret;
48652 }
48653 #endif
48654 
48655 #ifdef __LITTLE_ENDIAN__
vmaxvq_f32(float32x4_t __p0)48656 __ai float32_t vmaxvq_f32(float32x4_t __p0) {
48657   float32_t __ret;
48658   __ret = (float32_t) __builtin_neon_vmaxvq_f32(__p0);
48659   return __ret;
48660 }
48661 #else
vmaxvq_f32(float32x4_t __p0)48662 __ai float32_t vmaxvq_f32(float32x4_t __p0) {
48663   float32_t __ret;
48664   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48665   __ret = (float32_t) __builtin_neon_vmaxvq_f32(__rev0);
48666   return __ret;
48667 }
48668 #endif
48669 
48670 #ifdef __LITTLE_ENDIAN__
vmaxvq_s32(int32x4_t __p0)48671 __ai int32_t vmaxvq_s32(int32x4_t __p0) {
48672   int32_t __ret;
48673   __ret = (int32_t) __builtin_neon_vmaxvq_s32(__p0);
48674   return __ret;
48675 }
48676 #else
vmaxvq_s32(int32x4_t __p0)48677 __ai int32_t vmaxvq_s32(int32x4_t __p0) {
48678   int32_t __ret;
48679   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48680   __ret = (int32_t) __builtin_neon_vmaxvq_s32(__rev0);
48681   return __ret;
48682 }
48683 #endif
48684 
48685 #ifdef __LITTLE_ENDIAN__
vmaxvq_s16(int16x8_t __p0)48686 __ai int16_t vmaxvq_s16(int16x8_t __p0) {
48687   int16_t __ret;
48688   __ret = (int16_t) __builtin_neon_vmaxvq_s16(__p0);
48689   return __ret;
48690 }
48691 #else
vmaxvq_s16(int16x8_t __p0)48692 __ai int16_t vmaxvq_s16(int16x8_t __p0) {
48693   int16_t __ret;
48694   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
48695   __ret = (int16_t) __builtin_neon_vmaxvq_s16(__rev0);
48696   return __ret;
48697 }
48698 #endif
48699 
48700 #ifdef __LITTLE_ENDIAN__
vmaxv_u8(uint8x8_t __p0)48701 __ai uint8_t vmaxv_u8(uint8x8_t __p0) {
48702   uint8_t __ret;
48703   __ret = (uint8_t) __builtin_neon_vmaxv_u8(__p0);
48704   return __ret;
48705 }
48706 #else
vmaxv_u8(uint8x8_t __p0)48707 __ai uint8_t vmaxv_u8(uint8x8_t __p0) {
48708   uint8_t __ret;
48709   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
48710   __ret = (uint8_t) __builtin_neon_vmaxv_u8(__rev0);
48711   return __ret;
48712 }
48713 #endif
48714 
48715 #ifdef __LITTLE_ENDIAN__
vmaxv_u32(uint32x2_t __p0)48716 __ai uint32_t vmaxv_u32(uint32x2_t __p0) {
48717   uint32_t __ret;
48718   __ret = (uint32_t) __builtin_neon_vmaxv_u32(__p0);
48719   return __ret;
48720 }
48721 #else
vmaxv_u32(uint32x2_t __p0)48722 __ai uint32_t vmaxv_u32(uint32x2_t __p0) {
48723   uint32_t __ret;
48724   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48725   __ret = (uint32_t) __builtin_neon_vmaxv_u32(__rev0);
48726   return __ret;
48727 }
48728 #endif
48729 
48730 #ifdef __LITTLE_ENDIAN__
vmaxv_u16(uint16x4_t __p0)48731 __ai uint16_t vmaxv_u16(uint16x4_t __p0) {
48732   uint16_t __ret;
48733   __ret = (uint16_t) __builtin_neon_vmaxv_u16(__p0);
48734   return __ret;
48735 }
48736 #else
vmaxv_u16(uint16x4_t __p0)48737 __ai uint16_t vmaxv_u16(uint16x4_t __p0) {
48738   uint16_t __ret;
48739   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48740   __ret = (uint16_t) __builtin_neon_vmaxv_u16(__rev0);
48741   return __ret;
48742 }
48743 #endif
48744 
48745 #ifdef __LITTLE_ENDIAN__
vmaxv_s8(int8x8_t __p0)48746 __ai int8_t vmaxv_s8(int8x8_t __p0) {
48747   int8_t __ret;
48748   __ret = (int8_t) __builtin_neon_vmaxv_s8(__p0);
48749   return __ret;
48750 }
48751 #else
vmaxv_s8(int8x8_t __p0)48752 __ai int8_t vmaxv_s8(int8x8_t __p0) {
48753   int8_t __ret;
48754   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
48755   __ret = (int8_t) __builtin_neon_vmaxv_s8(__rev0);
48756   return __ret;
48757 }
48758 #endif
48759 
48760 #ifdef __LITTLE_ENDIAN__
vmaxv_f32(float32x2_t __p0)48761 __ai float32_t vmaxv_f32(float32x2_t __p0) {
48762   float32_t __ret;
48763   __ret = (float32_t) __builtin_neon_vmaxv_f32(__p0);
48764   return __ret;
48765 }
48766 #else
vmaxv_f32(float32x2_t __p0)48767 __ai float32_t vmaxv_f32(float32x2_t __p0) {
48768   float32_t __ret;
48769   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48770   __ret = (float32_t) __builtin_neon_vmaxv_f32(__rev0);
48771   return __ret;
48772 }
48773 #endif
48774 
48775 #ifdef __LITTLE_ENDIAN__
vmaxv_s32(int32x2_t __p0)48776 __ai int32_t vmaxv_s32(int32x2_t __p0) {
48777   int32_t __ret;
48778   __ret = (int32_t) __builtin_neon_vmaxv_s32(__p0);
48779   return __ret;
48780 }
48781 #else
vmaxv_s32(int32x2_t __p0)48782 __ai int32_t vmaxv_s32(int32x2_t __p0) {
48783   int32_t __ret;
48784   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48785   __ret = (int32_t) __builtin_neon_vmaxv_s32(__rev0);
48786   return __ret;
48787 }
48788 #endif
48789 
48790 #ifdef __LITTLE_ENDIAN__
vmaxv_s16(int16x4_t __p0)48791 __ai int16_t vmaxv_s16(int16x4_t __p0) {
48792   int16_t __ret;
48793   __ret = (int16_t) __builtin_neon_vmaxv_s16(__p0);
48794   return __ret;
48795 }
48796 #else
vmaxv_s16(int16x4_t __p0)48797 __ai int16_t vmaxv_s16(int16x4_t __p0) {
48798   int16_t __ret;
48799   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48800   __ret = (int16_t) __builtin_neon_vmaxv_s16(__rev0);
48801   return __ret;
48802 }
48803 #endif
48804 
48805 #ifdef __LITTLE_ENDIAN__
vminq_f64(float64x2_t __p0,float64x2_t __p1)48806 __ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
48807   float64x2_t __ret;
48808   __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
48809   return __ret;
48810 }
48811 #else
vminq_f64(float64x2_t __p0,float64x2_t __p1)48812 __ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
48813   float64x2_t __ret;
48814   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48815   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48816   __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
48817   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48818   return __ret;
48819 }
48820 #endif
48821 
vmin_f64(float64x1_t __p0,float64x1_t __p1)48822 __ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
48823   float64x1_t __ret;
48824   __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
48825   return __ret;
48826 }
48827 #ifdef __LITTLE_ENDIAN__
vminnmvq_f64(float64x2_t __p0)48828 __ai float64_t vminnmvq_f64(float64x2_t __p0) {
48829   float64_t __ret;
48830   __ret = (float64_t) __builtin_neon_vminnmvq_f64(__p0);
48831   return __ret;
48832 }
48833 #else
vminnmvq_f64(float64x2_t __p0)48834 __ai float64_t vminnmvq_f64(float64x2_t __p0) {
48835   float64_t __ret;
48836   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48837   __ret = (float64_t) __builtin_neon_vminnmvq_f64(__rev0);
48838   return __ret;
48839 }
48840 #endif
48841 
48842 #ifdef __LITTLE_ENDIAN__
vminnmvq_f32(float32x4_t __p0)48843 __ai float32_t vminnmvq_f32(float32x4_t __p0) {
48844   float32_t __ret;
48845   __ret = (float32_t) __builtin_neon_vminnmvq_f32(__p0);
48846   return __ret;
48847 }
48848 #else
vminnmvq_f32(float32x4_t __p0)48849 __ai float32_t vminnmvq_f32(float32x4_t __p0) {
48850   float32_t __ret;
48851   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48852   __ret = (float32_t) __builtin_neon_vminnmvq_f32(__rev0);
48853   return __ret;
48854 }
48855 #endif
48856 
48857 #ifdef __LITTLE_ENDIAN__
vminnmv_f32(float32x2_t __p0)48858 __ai float32_t vminnmv_f32(float32x2_t __p0) {
48859   float32_t __ret;
48860   __ret = (float32_t) __builtin_neon_vminnmv_f32(__p0);
48861   return __ret;
48862 }
48863 #else
vminnmv_f32(float32x2_t __p0)48864 __ai float32_t vminnmv_f32(float32x2_t __p0) {
48865   float32_t __ret;
48866   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48867   __ret = (float32_t) __builtin_neon_vminnmv_f32(__rev0);
48868   return __ret;
48869 }
48870 #endif
48871 
48872 #ifdef __LITTLE_ENDIAN__
vminvq_u8(uint8x16_t __p0)48873 __ai uint8_t vminvq_u8(uint8x16_t __p0) {
48874   uint8_t __ret;
48875   __ret = (uint8_t) __builtin_neon_vminvq_u8(__p0);
48876   return __ret;
48877 }
48878 #else
vminvq_u8(uint8x16_t __p0)48879 __ai uint8_t vminvq_u8(uint8x16_t __p0) {
48880   uint8_t __ret;
48881   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
48882   __ret = (uint8_t) __builtin_neon_vminvq_u8(__rev0);
48883   return __ret;
48884 }
48885 #endif
48886 
48887 #ifdef __LITTLE_ENDIAN__
vminvq_u32(uint32x4_t __p0)48888 __ai uint32_t vminvq_u32(uint32x4_t __p0) {
48889   uint32_t __ret;
48890   __ret = (uint32_t) __builtin_neon_vminvq_u32(__p0);
48891   return __ret;
48892 }
48893 #else
vminvq_u32(uint32x4_t __p0)48894 __ai uint32_t vminvq_u32(uint32x4_t __p0) {
48895   uint32_t __ret;
48896   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48897   __ret = (uint32_t) __builtin_neon_vminvq_u32(__rev0);
48898   return __ret;
48899 }
48900 #endif
48901 
48902 #ifdef __LITTLE_ENDIAN__
vminvq_u16(uint16x8_t __p0)48903 __ai uint16_t vminvq_u16(uint16x8_t __p0) {
48904   uint16_t __ret;
48905   __ret = (uint16_t) __builtin_neon_vminvq_u16(__p0);
48906   return __ret;
48907 }
48908 #else
vminvq_u16(uint16x8_t __p0)48909 __ai uint16_t vminvq_u16(uint16x8_t __p0) {
48910   uint16_t __ret;
48911   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
48912   __ret = (uint16_t) __builtin_neon_vminvq_u16(__rev0);
48913   return __ret;
48914 }
48915 #endif
48916 
48917 #ifdef __LITTLE_ENDIAN__
vminvq_s8(int8x16_t __p0)48918 __ai int8_t vminvq_s8(int8x16_t __p0) {
48919   int8_t __ret;
48920   __ret = (int8_t) __builtin_neon_vminvq_s8(__p0);
48921   return __ret;
48922 }
48923 #else
vminvq_s8(int8x16_t __p0)48924 __ai int8_t vminvq_s8(int8x16_t __p0) {
48925   int8_t __ret;
48926   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
48927   __ret = (int8_t) __builtin_neon_vminvq_s8(__rev0);
48928   return __ret;
48929 }
48930 #endif
48931 
48932 #ifdef __LITTLE_ENDIAN__
vminvq_f64(float64x2_t __p0)48933 __ai float64_t vminvq_f64(float64x2_t __p0) {
48934   float64_t __ret;
48935   __ret = (float64_t) __builtin_neon_vminvq_f64(__p0);
48936   return __ret;
48937 }
48938 #else
vminvq_f64(float64x2_t __p0)48939 __ai float64_t vminvq_f64(float64x2_t __p0) {
48940   float64_t __ret;
48941   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48942   __ret = (float64_t) __builtin_neon_vminvq_f64(__rev0);
48943   return __ret;
48944 }
48945 #endif
48946 
48947 #ifdef __LITTLE_ENDIAN__
vminvq_f32(float32x4_t __p0)48948 __ai float32_t vminvq_f32(float32x4_t __p0) {
48949   float32_t __ret;
48950   __ret = (float32_t) __builtin_neon_vminvq_f32(__p0);
48951   return __ret;
48952 }
48953 #else
vminvq_f32(float32x4_t __p0)48954 __ai float32_t vminvq_f32(float32x4_t __p0) {
48955   float32_t __ret;
48956   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48957   __ret = (float32_t) __builtin_neon_vminvq_f32(__rev0);
48958   return __ret;
48959 }
48960 #endif
48961 
48962 #ifdef __LITTLE_ENDIAN__
vminvq_s32(int32x4_t __p0)48963 __ai int32_t vminvq_s32(int32x4_t __p0) {
48964   int32_t __ret;
48965   __ret = (int32_t) __builtin_neon_vminvq_s32(__p0);
48966   return __ret;
48967 }
48968 #else
vminvq_s32(int32x4_t __p0)48969 __ai int32_t vminvq_s32(int32x4_t __p0) {
48970   int32_t __ret;
48971   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48972   __ret = (int32_t) __builtin_neon_vminvq_s32(__rev0);
48973   return __ret;
48974 }
48975 #endif
48976 
48977 #ifdef __LITTLE_ENDIAN__
vminvq_s16(int16x8_t __p0)48978 __ai int16_t vminvq_s16(int16x8_t __p0) {
48979   int16_t __ret;
48980   __ret = (int16_t) __builtin_neon_vminvq_s16(__p0);
48981   return __ret;
48982 }
48983 #else
vminvq_s16(int16x8_t __p0)48984 __ai int16_t vminvq_s16(int16x8_t __p0) {
48985   int16_t __ret;
48986   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
48987   __ret = (int16_t) __builtin_neon_vminvq_s16(__rev0);
48988   return __ret;
48989 }
48990 #endif
48991 
48992 #ifdef __LITTLE_ENDIAN__
vminv_u8(uint8x8_t __p0)48993 __ai uint8_t vminv_u8(uint8x8_t __p0) {
48994   uint8_t __ret;
48995   __ret = (uint8_t) __builtin_neon_vminv_u8(__p0);
48996   return __ret;
48997 }
48998 #else
vminv_u8(uint8x8_t __p0)48999 __ai uint8_t vminv_u8(uint8x8_t __p0) {
49000   uint8_t __ret;
49001   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
49002   __ret = (uint8_t) __builtin_neon_vminv_u8(__rev0);
49003   return __ret;
49004 }
49005 #endif
49006 
49007 #ifdef __LITTLE_ENDIAN__
vminv_u32(uint32x2_t __p0)49008 __ai uint32_t vminv_u32(uint32x2_t __p0) {
49009   uint32_t __ret;
49010   __ret = (uint32_t) __builtin_neon_vminv_u32(__p0);
49011   return __ret;
49012 }
49013 #else
vminv_u32(uint32x2_t __p0)49014 __ai uint32_t vminv_u32(uint32x2_t __p0) {
49015   uint32_t __ret;
49016   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
49017   __ret = (uint32_t) __builtin_neon_vminv_u32(__rev0);
49018   return __ret;
49019 }
49020 #endif
49021 
49022 #ifdef __LITTLE_ENDIAN__
vminv_u16(uint16x4_t __p0)49023 __ai uint16_t vminv_u16(uint16x4_t __p0) {
49024   uint16_t __ret;
49025   __ret = (uint16_t) __builtin_neon_vminv_u16(__p0);
49026   return __ret;
49027 }
49028 #else
vminv_u16(uint16x4_t __p0)49029 __ai uint16_t vminv_u16(uint16x4_t __p0) {
49030   uint16_t __ret;
49031   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
49032   __ret = (uint16_t) __builtin_neon_vminv_u16(__rev0);
49033   return __ret;
49034 }
49035 #endif
49036 
49037 #ifdef __LITTLE_ENDIAN__
vminv_s8(int8x8_t __p0)49038 __ai int8_t vminv_s8(int8x8_t __p0) {
49039   int8_t __ret;
49040   __ret = (int8_t) __builtin_neon_vminv_s8(__p0);
49041   return __ret;
49042 }
49043 #else
vminv_s8(int8x8_t __p0)49044 __ai int8_t vminv_s8(int8x8_t __p0) {
49045   int8_t __ret;
49046   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
49047   __ret = (int8_t) __builtin_neon_vminv_s8(__rev0);
49048   return __ret;
49049 }
49050 #endif
49051 
49052 #ifdef __LITTLE_ENDIAN__
vminv_f32(float32x2_t __p0)49053 __ai float32_t vminv_f32(float32x2_t __p0) {
49054   float32_t __ret;
49055   __ret = (float32_t) __builtin_neon_vminv_f32(__p0);
49056   return __ret;
49057 }
49058 #else
vminv_f32(float32x2_t __p0)49059 __ai float32_t vminv_f32(float32x2_t __p0) {
49060   float32_t __ret;
49061   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
49062   __ret = (float32_t) __builtin_neon_vminv_f32(__rev0);
49063   return __ret;
49064 }
49065 #endif
49066 
49067 #ifdef __LITTLE_ENDIAN__
vminv_s32(int32x2_t __p0)49068 __ai int32_t vminv_s32(int32x2_t __p0) {
49069   int32_t __ret;
49070   __ret = (int32_t) __builtin_neon_vminv_s32(__p0);
49071   return __ret;
49072 }
49073 #else
vminv_s32(int32x2_t __p0)49074 __ai int32_t vminv_s32(int32x2_t __p0) {
49075   int32_t __ret;
49076   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
49077   __ret = (int32_t) __builtin_neon_vminv_s32(__rev0);
49078   return __ret;
49079 }
49080 #endif
49081 
49082 #ifdef __LITTLE_ENDIAN__
vminv_s16(int16x4_t __p0)49083 __ai int16_t vminv_s16(int16x4_t __p0) {
49084   int16_t __ret;
49085   __ret = (int16_t) __builtin_neon_vminv_s16(__p0);
49086   return __ret;
49087 }
49088 #else
vminv_s16(int16x4_t __p0)49089 __ai int16_t vminv_s16(int16x4_t __p0) {
49090   int16_t __ret;
49091   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
49092   __ret = (int16_t) __builtin_neon_vminv_s16(__rev0);
49093   return __ret;
49094 }
49095 #endif
49096 
49097 #ifdef __LITTLE_ENDIAN__
vmlaq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)49098 __ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
49099   float64x2_t __ret;
49100   __ret = __p0 + __p1 * __p2;
49101   return __ret;
49102 }
49103 #else
vmlaq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)49104 __ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
49105   float64x2_t __ret;
49106   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
49107   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
49108   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
49109   __ret = __rev0 + __rev1 * __rev2;
49110   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
49111   return __ret;
49112 }
49113 #endif
49114 
vmla_f64(float64x1_t __p0,float64x1_t __p1,float64x1_t __p2)49115 __ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
49116   float64x1_t __ret;
49117   __ret = __p0 + __p1 * __p2;
49118   return __ret;
49119 }
49120 #ifdef __LITTLE_ENDIAN__
49121 #define vmlaq_laneq_u32(__p0_427, __p1_427, __p2_427, __p3_427) __extension__ ({ \
49122   uint32x4_t __ret_427; \
49123   uint32x4_t __s0_427 = __p0_427; \
49124   uint32x4_t __s1_427 = __p1_427; \
49125   uint32x4_t __s2_427 = __p2_427; \
49126   __ret_427 = __s0_427 + __s1_427 * splatq_laneq_u32(__s2_427, __p3_427); \
49127   __ret_427; \
49128 })
49129 #else
49130 #define vmlaq_laneq_u32(__p0_428, __p1_428, __p2_428, __p3_428) __extension__ ({ \
49131   uint32x4_t __ret_428; \
49132   uint32x4_t __s0_428 = __p0_428; \
49133   uint32x4_t __s1_428 = __p1_428; \
49134   uint32x4_t __s2_428 = __p2_428; \
49135   uint32x4_t __rev0_428;  __rev0_428 = __builtin_shufflevector(__s0_428, __s0_428, 3, 2, 1, 0); \
49136   uint32x4_t __rev1_428;  __rev1_428 = __builtin_shufflevector(__s1_428, __s1_428, 3, 2, 1, 0); \
49137   uint32x4_t __rev2_428;  __rev2_428 = __builtin_shufflevector(__s2_428, __s2_428, 3, 2, 1, 0); \
49138   __ret_428 = __rev0_428 + __rev1_428 * __noswap_splatq_laneq_u32(__rev2_428, __p3_428); \
49139   __ret_428 = __builtin_shufflevector(__ret_428, __ret_428, 3, 2, 1, 0); \
49140   __ret_428; \
49141 })
49142 #endif
49143 
49144 #ifdef __LITTLE_ENDIAN__
49145 #define vmlaq_laneq_u16(__p0_429, __p1_429, __p2_429, __p3_429) __extension__ ({ \
49146   uint16x8_t __ret_429; \
49147   uint16x8_t __s0_429 = __p0_429; \
49148   uint16x8_t __s1_429 = __p1_429; \
49149   uint16x8_t __s2_429 = __p2_429; \
49150   __ret_429 = __s0_429 + __s1_429 * splatq_laneq_u16(__s2_429, __p3_429); \
49151   __ret_429; \
49152 })
49153 #else
49154 #define vmlaq_laneq_u16(__p0_430, __p1_430, __p2_430, __p3_430) __extension__ ({ \
49155   uint16x8_t __ret_430; \
49156   uint16x8_t __s0_430 = __p0_430; \
49157   uint16x8_t __s1_430 = __p1_430; \
49158   uint16x8_t __s2_430 = __p2_430; \
49159   uint16x8_t __rev0_430;  __rev0_430 = __builtin_shufflevector(__s0_430, __s0_430, 7, 6, 5, 4, 3, 2, 1, 0); \
49160   uint16x8_t __rev1_430;  __rev1_430 = __builtin_shufflevector(__s1_430, __s1_430, 7, 6, 5, 4, 3, 2, 1, 0); \
49161   uint16x8_t __rev2_430;  __rev2_430 = __builtin_shufflevector(__s2_430, __s2_430, 7, 6, 5, 4, 3, 2, 1, 0); \
49162   __ret_430 = __rev0_430 + __rev1_430 * __noswap_splatq_laneq_u16(__rev2_430, __p3_430); \
49163   __ret_430 = __builtin_shufflevector(__ret_430, __ret_430, 7, 6, 5, 4, 3, 2, 1, 0); \
49164   __ret_430; \
49165 })
49166 #endif
49167 
49168 #ifdef __LITTLE_ENDIAN__
49169 #define vmlaq_laneq_f32(__p0_431, __p1_431, __p2_431, __p3_431) __extension__ ({ \
49170   float32x4_t __ret_431; \
49171   float32x4_t __s0_431 = __p0_431; \
49172   float32x4_t __s1_431 = __p1_431; \
49173   float32x4_t __s2_431 = __p2_431; \
49174   __ret_431 = __s0_431 + __s1_431 * splatq_laneq_f32(__s2_431, __p3_431); \
49175   __ret_431; \
49176 })
49177 #else
49178 #define vmlaq_laneq_f32(__p0_432, __p1_432, __p2_432, __p3_432) __extension__ ({ \
49179   float32x4_t __ret_432; \
49180   float32x4_t __s0_432 = __p0_432; \
49181   float32x4_t __s1_432 = __p1_432; \
49182   float32x4_t __s2_432 = __p2_432; \
49183   float32x4_t __rev0_432;  __rev0_432 = __builtin_shufflevector(__s0_432, __s0_432, 3, 2, 1, 0); \
49184   float32x4_t __rev1_432;  __rev1_432 = __builtin_shufflevector(__s1_432, __s1_432, 3, 2, 1, 0); \
49185   float32x4_t __rev2_432;  __rev2_432 = __builtin_shufflevector(__s2_432, __s2_432, 3, 2, 1, 0); \
49186   __ret_432 = __rev0_432 + __rev1_432 * __noswap_splatq_laneq_f32(__rev2_432, __p3_432); \
49187   __ret_432 = __builtin_shufflevector(__ret_432, __ret_432, 3, 2, 1, 0); \
49188   __ret_432; \
49189 })
49190 #endif
49191 
49192 #ifdef __LITTLE_ENDIAN__
49193 #define vmlaq_laneq_s32(__p0_433, __p1_433, __p2_433, __p3_433) __extension__ ({ \
49194   int32x4_t __ret_433; \
49195   int32x4_t __s0_433 = __p0_433; \
49196   int32x4_t __s1_433 = __p1_433; \
49197   int32x4_t __s2_433 = __p2_433; \
49198   __ret_433 = __s0_433 + __s1_433 * splatq_laneq_s32(__s2_433, __p3_433); \
49199   __ret_433; \
49200 })
49201 #else
49202 #define vmlaq_laneq_s32(__p0_434, __p1_434, __p2_434, __p3_434) __extension__ ({ \
49203   int32x4_t __ret_434; \
49204   int32x4_t __s0_434 = __p0_434; \
49205   int32x4_t __s1_434 = __p1_434; \
49206   int32x4_t __s2_434 = __p2_434; \
49207   int32x4_t __rev0_434;  __rev0_434 = __builtin_shufflevector(__s0_434, __s0_434, 3, 2, 1, 0); \
49208   int32x4_t __rev1_434;  __rev1_434 = __builtin_shufflevector(__s1_434, __s1_434, 3, 2, 1, 0); \
49209   int32x4_t __rev2_434;  __rev2_434 = __builtin_shufflevector(__s2_434, __s2_434, 3, 2, 1, 0); \
49210   __ret_434 = __rev0_434 + __rev1_434 * __noswap_splatq_laneq_s32(__rev2_434, __p3_434); \
49211   __ret_434 = __builtin_shufflevector(__ret_434, __ret_434, 3, 2, 1, 0); \
49212   __ret_434; \
49213 })
49214 #endif
49215 
49216 #ifdef __LITTLE_ENDIAN__
49217 #define vmlaq_laneq_s16(__p0_435, __p1_435, __p2_435, __p3_435) __extension__ ({ \
49218   int16x8_t __ret_435; \
49219   int16x8_t __s0_435 = __p0_435; \
49220   int16x8_t __s1_435 = __p1_435; \
49221   int16x8_t __s2_435 = __p2_435; \
49222   __ret_435 = __s0_435 + __s1_435 * splatq_laneq_s16(__s2_435, __p3_435); \
49223   __ret_435; \
49224 })
49225 #else
49226 #define vmlaq_laneq_s16(__p0_436, __p1_436, __p2_436, __p3_436) __extension__ ({ \
49227   int16x8_t __ret_436; \
49228   int16x8_t __s0_436 = __p0_436; \
49229   int16x8_t __s1_436 = __p1_436; \
49230   int16x8_t __s2_436 = __p2_436; \
49231   int16x8_t __rev0_436;  __rev0_436 = __builtin_shufflevector(__s0_436, __s0_436, 7, 6, 5, 4, 3, 2, 1, 0); \
49232   int16x8_t __rev1_436;  __rev1_436 = __builtin_shufflevector(__s1_436, __s1_436, 7, 6, 5, 4, 3, 2, 1, 0); \
49233   int16x8_t __rev2_436;  __rev2_436 = __builtin_shufflevector(__s2_436, __s2_436, 7, 6, 5, 4, 3, 2, 1, 0); \
49234   __ret_436 = __rev0_436 + __rev1_436 * __noswap_splatq_laneq_s16(__rev2_436, __p3_436); \
49235   __ret_436 = __builtin_shufflevector(__ret_436, __ret_436, 7, 6, 5, 4, 3, 2, 1, 0); \
49236   __ret_436; \
49237 })
49238 #endif
49239 
49240 #ifdef __LITTLE_ENDIAN__
49241 #define vmla_laneq_u32(__p0_437, __p1_437, __p2_437, __p3_437) __extension__ ({ \
49242   uint32x2_t __ret_437; \
49243   uint32x2_t __s0_437 = __p0_437; \
49244   uint32x2_t __s1_437 = __p1_437; \
49245   uint32x4_t __s2_437 = __p2_437; \
49246   __ret_437 = __s0_437 + __s1_437 * splat_laneq_u32(__s2_437, __p3_437); \
49247   __ret_437; \
49248 })
49249 #else
49250 #define vmla_laneq_u32(__p0_438, __p1_438, __p2_438, __p3_438) __extension__ ({ \
49251   uint32x2_t __ret_438; \
49252   uint32x2_t __s0_438 = __p0_438; \
49253   uint32x2_t __s1_438 = __p1_438; \
49254   uint32x4_t __s2_438 = __p2_438; \
49255   uint32x2_t __rev0_438;  __rev0_438 = __builtin_shufflevector(__s0_438, __s0_438, 1, 0); \
49256   uint32x2_t __rev1_438;  __rev1_438 = __builtin_shufflevector(__s1_438, __s1_438, 1, 0); \
49257   uint32x4_t __rev2_438;  __rev2_438 = __builtin_shufflevector(__s2_438, __s2_438, 3, 2, 1, 0); \
49258   __ret_438 = __rev0_438 + __rev1_438 * __noswap_splat_laneq_u32(__rev2_438, __p3_438); \
49259   __ret_438 = __builtin_shufflevector(__ret_438, __ret_438, 1, 0); \
49260   __ret_438; \
49261 })
49262 #endif
49263 
49264 #ifdef __LITTLE_ENDIAN__
49265 #define vmla_laneq_u16(__p0_439, __p1_439, __p2_439, __p3_439) __extension__ ({ \
49266   uint16x4_t __ret_439; \
49267   uint16x4_t __s0_439 = __p0_439; \
49268   uint16x4_t __s1_439 = __p1_439; \
49269   uint16x8_t __s2_439 = __p2_439; \
49270   __ret_439 = __s0_439 + __s1_439 * splat_laneq_u16(__s2_439, __p3_439); \
49271   __ret_439; \
49272 })
49273 #else
49274 #define vmla_laneq_u16(__p0_440, __p1_440, __p2_440, __p3_440) __extension__ ({ \
49275   uint16x4_t __ret_440; \
49276   uint16x4_t __s0_440 = __p0_440; \
49277   uint16x4_t __s1_440 = __p1_440; \
49278   uint16x8_t __s2_440 = __p2_440; \
49279   uint16x4_t __rev0_440;  __rev0_440 = __builtin_shufflevector(__s0_440, __s0_440, 3, 2, 1, 0); \
49280   uint16x4_t __rev1_440;  __rev1_440 = __builtin_shufflevector(__s1_440, __s1_440, 3, 2, 1, 0); \
49281   uint16x8_t __rev2_440;  __rev2_440 = __builtin_shufflevector(__s2_440, __s2_440, 7, 6, 5, 4, 3, 2, 1, 0); \
49282   __ret_440 = __rev0_440 + __rev1_440 * __noswap_splat_laneq_u16(__rev2_440, __p3_440); \
49283   __ret_440 = __builtin_shufflevector(__ret_440, __ret_440, 3, 2, 1, 0); \
49284   __ret_440; \
49285 })
49286 #endif
49287 
49288 #ifdef __LITTLE_ENDIAN__
49289 #define vmla_laneq_f32(__p0_441, __p1_441, __p2_441, __p3_441) __extension__ ({ \
49290   float32x2_t __ret_441; \
49291   float32x2_t __s0_441 = __p0_441; \
49292   float32x2_t __s1_441 = __p1_441; \
49293   float32x4_t __s2_441 = __p2_441; \
49294   __ret_441 = __s0_441 + __s1_441 * splat_laneq_f32(__s2_441, __p3_441); \
49295   __ret_441; \
49296 })
49297 #else
49298 #define vmla_laneq_f32(__p0_442, __p1_442, __p2_442, __p3_442) __extension__ ({ \
49299   float32x2_t __ret_442; \
49300   float32x2_t __s0_442 = __p0_442; \
49301   float32x2_t __s1_442 = __p1_442; \
49302   float32x4_t __s2_442 = __p2_442; \
49303   float32x2_t __rev0_442;  __rev0_442 = __builtin_shufflevector(__s0_442, __s0_442, 1, 0); \
49304   float32x2_t __rev1_442;  __rev1_442 = __builtin_shufflevector(__s1_442, __s1_442, 1, 0); \
49305   float32x4_t __rev2_442;  __rev2_442 = __builtin_shufflevector(__s2_442, __s2_442, 3, 2, 1, 0); \
49306   __ret_442 = __rev0_442 + __rev1_442 * __noswap_splat_laneq_f32(__rev2_442, __p3_442); \
49307   __ret_442 = __builtin_shufflevector(__ret_442, __ret_442, 1, 0); \
49308   __ret_442; \
49309 })
49310 #endif
49311 
49312 #ifdef __LITTLE_ENDIAN__
49313 #define vmla_laneq_s32(__p0_443, __p1_443, __p2_443, __p3_443) __extension__ ({ \
49314   int32x2_t __ret_443; \
49315   int32x2_t __s0_443 = __p0_443; \
49316   int32x2_t __s1_443 = __p1_443; \
49317   int32x4_t __s2_443 = __p2_443; \
49318   __ret_443 = __s0_443 + __s1_443 * splat_laneq_s32(__s2_443, __p3_443); \
49319   __ret_443; \
49320 })
49321 #else
49322 #define vmla_laneq_s32(__p0_444, __p1_444, __p2_444, __p3_444) __extension__ ({ \
49323   int32x2_t __ret_444; \
49324   int32x2_t __s0_444 = __p0_444; \
49325   int32x2_t __s1_444 = __p1_444; \
49326   int32x4_t __s2_444 = __p2_444; \
49327   int32x2_t __rev0_444;  __rev0_444 = __builtin_shufflevector(__s0_444, __s0_444, 1, 0); \
49328   int32x2_t __rev1_444;  __rev1_444 = __builtin_shufflevector(__s1_444, __s1_444, 1, 0); \
49329   int32x4_t __rev2_444;  __rev2_444 = __builtin_shufflevector(__s2_444, __s2_444, 3, 2, 1, 0); \
49330   __ret_444 = __rev0_444 + __rev1_444 * __noswap_splat_laneq_s32(__rev2_444, __p3_444); \
49331   __ret_444 = __builtin_shufflevector(__ret_444, __ret_444, 1, 0); \
49332   __ret_444; \
49333 })
49334 #endif
49335 
49336 #ifdef __LITTLE_ENDIAN__
49337 #define vmla_laneq_s16(__p0_445, __p1_445, __p2_445, __p3_445) __extension__ ({ \
49338   int16x4_t __ret_445; \
49339   int16x4_t __s0_445 = __p0_445; \
49340   int16x4_t __s1_445 = __p1_445; \
49341   int16x8_t __s2_445 = __p2_445; \
49342   __ret_445 = __s0_445 + __s1_445 * splat_laneq_s16(__s2_445, __p3_445); \
49343   __ret_445; \
49344 })
49345 #else
49346 #define vmla_laneq_s16(__p0_446, __p1_446, __p2_446, __p3_446) __extension__ ({ \
49347   int16x4_t __ret_446; \
49348   int16x4_t __s0_446 = __p0_446; \
49349   int16x4_t __s1_446 = __p1_446; \
49350   int16x8_t __s2_446 = __p2_446; \
49351   int16x4_t __rev0_446;  __rev0_446 = __builtin_shufflevector(__s0_446, __s0_446, 3, 2, 1, 0); \
49352   int16x4_t __rev1_446;  __rev1_446 = __builtin_shufflevector(__s1_446, __s1_446, 3, 2, 1, 0); \
49353   int16x8_t __rev2_446;  __rev2_446 = __builtin_shufflevector(__s2_446, __s2_446, 7, 6, 5, 4, 3, 2, 1, 0); \
49354   __ret_446 = __rev0_446 + __rev1_446 * __noswap_splat_laneq_s16(__rev2_446, __p3_446); \
49355   __ret_446 = __builtin_shufflevector(__ret_446, __ret_446, 3, 2, 1, 0); \
49356   __ret_446; \
49357 })
49358 #endif
49359 
49360 #ifdef __LITTLE_ENDIAN__
49361 #define vmlal_high_lane_u32(__p0_447, __p1_447, __p2_447, __p3_447) __extension__ ({ \
49362   uint64x2_t __ret_447; \
49363   uint64x2_t __s0_447 = __p0_447; \
49364   uint32x4_t __s1_447 = __p1_447; \
49365   uint32x2_t __s2_447 = __p2_447; \
49366   __ret_447 = __s0_447 + vmull_u32(vget_high_u32(__s1_447), splat_lane_u32(__s2_447, __p3_447)); \
49367   __ret_447; \
49368 })
49369 #else
49370 #define vmlal_high_lane_u32(__p0_448, __p1_448, __p2_448, __p3_448) __extension__ ({ \
49371   uint64x2_t __ret_448; \
49372   uint64x2_t __s0_448 = __p0_448; \
49373   uint32x4_t __s1_448 = __p1_448; \
49374   uint32x2_t __s2_448 = __p2_448; \
49375   uint64x2_t __rev0_448;  __rev0_448 = __builtin_shufflevector(__s0_448, __s0_448, 1, 0); \
49376   uint32x4_t __rev1_448;  __rev1_448 = __builtin_shufflevector(__s1_448, __s1_448, 3, 2, 1, 0); \
49377   uint32x2_t __rev2_448;  __rev2_448 = __builtin_shufflevector(__s2_448, __s2_448, 1, 0); \
49378   __ret_448 = __rev0_448 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1_448), __noswap_splat_lane_u32(__rev2_448, __p3_448)); \
49379   __ret_448 = __builtin_shufflevector(__ret_448, __ret_448, 1, 0); \
49380   __ret_448; \
49381 })
49382 #endif
49383 
49384 #ifdef __LITTLE_ENDIAN__
49385 #define vmlal_high_lane_u16(__p0_449, __p1_449, __p2_449, __p3_449) __extension__ ({ \
49386   uint32x4_t __ret_449; \
49387   uint32x4_t __s0_449 = __p0_449; \
49388   uint16x8_t __s1_449 = __p1_449; \
49389   uint16x4_t __s2_449 = __p2_449; \
49390   __ret_449 = __s0_449 + vmull_u16(vget_high_u16(__s1_449), splat_lane_u16(__s2_449, __p3_449)); \
49391   __ret_449; \
49392 })
49393 #else
49394 #define vmlal_high_lane_u16(__p0_450, __p1_450, __p2_450, __p3_450) __extension__ ({ \
49395   uint32x4_t __ret_450; \
49396   uint32x4_t __s0_450 = __p0_450; \
49397   uint16x8_t __s1_450 = __p1_450; \
49398   uint16x4_t __s2_450 = __p2_450; \
49399   uint32x4_t __rev0_450;  __rev0_450 = __builtin_shufflevector(__s0_450, __s0_450, 3, 2, 1, 0); \
49400   uint16x8_t __rev1_450;  __rev1_450 = __builtin_shufflevector(__s1_450, __s1_450, 7, 6, 5, 4, 3, 2, 1, 0); \
49401   uint16x4_t __rev2_450;  __rev2_450 = __builtin_shufflevector(__s2_450, __s2_450, 3, 2, 1, 0); \
49402   __ret_450 = __rev0_450 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1_450), __noswap_splat_lane_u16(__rev2_450, __p3_450)); \
49403   __ret_450 = __builtin_shufflevector(__ret_450, __ret_450, 3, 2, 1, 0); \
49404   __ret_450; \
49405 })
49406 #endif
49407 
49408 #ifdef __LITTLE_ENDIAN__
49409 #define vmlal_high_lane_s32(__p0_451, __p1_451, __p2_451, __p3_451) __extension__ ({ \
49410   int64x2_t __ret_451; \
49411   int64x2_t __s0_451 = __p0_451; \
49412   int32x4_t __s1_451 = __p1_451; \
49413   int32x2_t __s2_451 = __p2_451; \
49414   __ret_451 = __s0_451 + vmull_s32(vget_high_s32(__s1_451), splat_lane_s32(__s2_451, __p3_451)); \
49415   __ret_451; \
49416 })
49417 #else
49418 #define vmlal_high_lane_s32(__p0_452, __p1_452, __p2_452, __p3_452) __extension__ ({ \
49419   int64x2_t __ret_452; \
49420   int64x2_t __s0_452 = __p0_452; \
49421   int32x4_t __s1_452 = __p1_452; \
49422   int32x2_t __s2_452 = __p2_452; \
49423   int64x2_t __rev0_452;  __rev0_452 = __builtin_shufflevector(__s0_452, __s0_452, 1, 0); \
49424   int32x4_t __rev1_452;  __rev1_452 = __builtin_shufflevector(__s1_452, __s1_452, 3, 2, 1, 0); \
49425   int32x2_t __rev2_452;  __rev2_452 = __builtin_shufflevector(__s2_452, __s2_452, 1, 0); \
49426   __ret_452 = __rev0_452 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1_452), __noswap_splat_lane_s32(__rev2_452, __p3_452)); \
49427   __ret_452 = __builtin_shufflevector(__ret_452, __ret_452, 1, 0); \
49428   __ret_452; \
49429 })
49430 #endif
49431 
49432 #ifdef __LITTLE_ENDIAN__
49433 #define vmlal_high_lane_s16(__p0_453, __p1_453, __p2_453, __p3_453) __extension__ ({ \
49434   int32x4_t __ret_453; \
49435   int32x4_t __s0_453 = __p0_453; \
49436   int16x8_t __s1_453 = __p1_453; \
49437   int16x4_t __s2_453 = __p2_453; \
49438   __ret_453 = __s0_453 + vmull_s16(vget_high_s16(__s1_453), splat_lane_s16(__s2_453, __p3_453)); \
49439   __ret_453; \
49440 })
49441 #else
49442 #define vmlal_high_lane_s16(__p0_454, __p1_454, __p2_454, __p3_454) __extension__ ({ \
49443   int32x4_t __ret_454; \
49444   int32x4_t __s0_454 = __p0_454; \
49445   int16x8_t __s1_454 = __p1_454; \
49446   int16x4_t __s2_454 = __p2_454; \
49447   int32x4_t __rev0_454;  __rev0_454 = __builtin_shufflevector(__s0_454, __s0_454, 3, 2, 1, 0); \
49448   int16x8_t __rev1_454;  __rev1_454 = __builtin_shufflevector(__s1_454, __s1_454, 7, 6, 5, 4, 3, 2, 1, 0); \
49449   int16x4_t __rev2_454;  __rev2_454 = __builtin_shufflevector(__s2_454, __s2_454, 3, 2, 1, 0); \
49450   __ret_454 = __rev0_454 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1_454), __noswap_splat_lane_s16(__rev2_454, __p3_454)); \
49451   __ret_454 = __builtin_shufflevector(__ret_454, __ret_454, 3, 2, 1, 0); \
49452   __ret_454; \
49453 })
49454 #endif
49455 
49456 #ifdef __LITTLE_ENDIAN__
49457 #define vmlal_high_laneq_u32(__p0_455, __p1_455, __p2_455, __p3_455) __extension__ ({ \
49458   uint64x2_t __ret_455; \
49459   uint64x2_t __s0_455 = __p0_455; \
49460   uint32x4_t __s1_455 = __p1_455; \
49461   uint32x4_t __s2_455 = __p2_455; \
49462   __ret_455 = __s0_455 + vmull_u32(vget_high_u32(__s1_455), splat_laneq_u32(__s2_455, __p3_455)); \
49463   __ret_455; \
49464 })
49465 #else
49466 #define vmlal_high_laneq_u32(__p0_456, __p1_456, __p2_456, __p3_456) __extension__ ({ \
49467   uint64x2_t __ret_456; \
49468   uint64x2_t __s0_456 = __p0_456; \
49469   uint32x4_t __s1_456 = __p1_456; \
49470   uint32x4_t __s2_456 = __p2_456; \
49471   uint64x2_t __rev0_456;  __rev0_456 = __builtin_shufflevector(__s0_456, __s0_456, 1, 0); \
49472   uint32x4_t __rev1_456;  __rev1_456 = __builtin_shufflevector(__s1_456, __s1_456, 3, 2, 1, 0); \
49473   uint32x4_t __rev2_456;  __rev2_456 = __builtin_shufflevector(__s2_456, __s2_456, 3, 2, 1, 0); \
49474   __ret_456 = __rev0_456 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1_456), __noswap_splat_laneq_u32(__rev2_456, __p3_456)); \
49475   __ret_456 = __builtin_shufflevector(__ret_456, __ret_456, 1, 0); \
49476   __ret_456; \
49477 })
49478 #endif
49479 
49480 #ifdef __LITTLE_ENDIAN__
49481 #define vmlal_high_laneq_u16(__p0_457, __p1_457, __p2_457, __p3_457) __extension__ ({ \
49482   uint32x4_t __ret_457; \
49483   uint32x4_t __s0_457 = __p0_457; \
49484   uint16x8_t __s1_457 = __p1_457; \
49485   uint16x8_t __s2_457 = __p2_457; \
49486   __ret_457 = __s0_457 + vmull_u16(vget_high_u16(__s1_457), splat_laneq_u16(__s2_457, __p3_457)); \
49487   __ret_457; \
49488 })
49489 #else
49490 #define vmlal_high_laneq_u16(__p0_458, __p1_458, __p2_458, __p3_458) __extension__ ({ \
49491   uint32x4_t __ret_458; \
49492   uint32x4_t __s0_458 = __p0_458; \
49493   uint16x8_t __s1_458 = __p1_458; \
49494   uint16x8_t __s2_458 = __p2_458; \
49495   uint32x4_t __rev0_458;  __rev0_458 = __builtin_shufflevector(__s0_458, __s0_458, 3, 2, 1, 0); \
49496   uint16x8_t __rev1_458;  __rev1_458 = __builtin_shufflevector(__s1_458, __s1_458, 7, 6, 5, 4, 3, 2, 1, 0); \
49497   uint16x8_t __rev2_458;  __rev2_458 = __builtin_shufflevector(__s2_458, __s2_458, 7, 6, 5, 4, 3, 2, 1, 0); \
49498   __ret_458 = __rev0_458 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1_458), __noswap_splat_laneq_u16(__rev2_458, __p3_458)); \
49499   __ret_458 = __builtin_shufflevector(__ret_458, __ret_458, 3, 2, 1, 0); \
49500   __ret_458; \
49501 })
49502 #endif
49503 
49504 #ifdef __LITTLE_ENDIAN__
49505 #define vmlal_high_laneq_s32(__p0_459, __p1_459, __p2_459, __p3_459) __extension__ ({ \
49506   int64x2_t __ret_459; \
49507   int64x2_t __s0_459 = __p0_459; \
49508   int32x4_t __s1_459 = __p1_459; \
49509   int32x4_t __s2_459 = __p2_459; \
49510   __ret_459 = __s0_459 + vmull_s32(vget_high_s32(__s1_459), splat_laneq_s32(__s2_459, __p3_459)); \
49511   __ret_459; \
49512 })
49513 #else
49514 #define vmlal_high_laneq_s32(__p0_460, __p1_460, __p2_460, __p3_460) __extension__ ({ \
49515   int64x2_t __ret_460; \
49516   int64x2_t __s0_460 = __p0_460; \
49517   int32x4_t __s1_460 = __p1_460; \
49518   int32x4_t __s2_460 = __p2_460; \
49519   int64x2_t __rev0_460;  __rev0_460 = __builtin_shufflevector(__s0_460, __s0_460, 1, 0); \
49520   int32x4_t __rev1_460;  __rev1_460 = __builtin_shufflevector(__s1_460, __s1_460, 3, 2, 1, 0); \
49521   int32x4_t __rev2_460;  __rev2_460 = __builtin_shufflevector(__s2_460, __s2_460, 3, 2, 1, 0); \
49522   __ret_460 = __rev0_460 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1_460), __noswap_splat_laneq_s32(__rev2_460, __p3_460)); \
49523   __ret_460 = __builtin_shufflevector(__ret_460, __ret_460, 1, 0); \
49524   __ret_460; \
49525 })
49526 #endif
49527 
49528 #ifdef __LITTLE_ENDIAN__
49529 #define vmlal_high_laneq_s16(__p0_461, __p1_461, __p2_461, __p3_461) __extension__ ({ \
49530   int32x4_t __ret_461; \
49531   int32x4_t __s0_461 = __p0_461; \
49532   int16x8_t __s1_461 = __p1_461; \
49533   int16x8_t __s2_461 = __p2_461; \
49534   __ret_461 = __s0_461 + vmull_s16(vget_high_s16(__s1_461), splat_laneq_s16(__s2_461, __p3_461)); \
49535   __ret_461; \
49536 })
49537 #else
49538 #define vmlal_high_laneq_s16(__p0_462, __p1_462, __p2_462, __p3_462) __extension__ ({ \
49539   int32x4_t __ret_462; \
49540   int32x4_t __s0_462 = __p0_462; \
49541   int16x8_t __s1_462 = __p1_462; \
49542   int16x8_t __s2_462 = __p2_462; \
49543   int32x4_t __rev0_462;  __rev0_462 = __builtin_shufflevector(__s0_462, __s0_462, 3, 2, 1, 0); \
49544   int16x8_t __rev1_462;  __rev1_462 = __builtin_shufflevector(__s1_462, __s1_462, 7, 6, 5, 4, 3, 2, 1, 0); \
49545   int16x8_t __rev2_462;  __rev2_462 = __builtin_shufflevector(__s2_462, __s2_462, 7, 6, 5, 4, 3, 2, 1, 0); \
49546   __ret_462 = __rev0_462 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1_462), __noswap_splat_laneq_s16(__rev2_462, __p3_462)); \
49547   __ret_462 = __builtin_shufflevector(__ret_462, __ret_462, 3, 2, 1, 0); \
49548   __ret_462; \
49549 })
49550 #endif
49551 
49552 #ifdef __LITTLE_ENDIAN__
49553 #define vmlal_laneq_u32(__p0_463, __p1_463, __p2_463, __p3_463) __extension__ ({ \
49554   uint64x2_t __ret_463; \
49555   uint64x2_t __s0_463 = __p0_463; \
49556   uint32x2_t __s1_463 = __p1_463; \
49557   uint32x4_t __s2_463 = __p2_463; \
49558   __ret_463 = __s0_463 + vmull_u32(__s1_463, splat_laneq_u32(__s2_463, __p3_463)); \
49559   __ret_463; \
49560 })
49561 #else
49562 #define vmlal_laneq_u32(__p0_464, __p1_464, __p2_464, __p3_464) __extension__ ({ \
49563   uint64x2_t __ret_464; \
49564   uint64x2_t __s0_464 = __p0_464; \
49565   uint32x2_t __s1_464 = __p1_464; \
49566   uint32x4_t __s2_464 = __p2_464; \
49567   uint64x2_t __rev0_464;  __rev0_464 = __builtin_shufflevector(__s0_464, __s0_464, 1, 0); \
49568   uint32x2_t __rev1_464;  __rev1_464 = __builtin_shufflevector(__s1_464, __s1_464, 1, 0); \
49569   uint32x4_t __rev2_464;  __rev2_464 = __builtin_shufflevector(__s2_464, __s2_464, 3, 2, 1, 0); \
49570   __ret_464 = __rev0_464 + __noswap_vmull_u32(__rev1_464, __noswap_splat_laneq_u32(__rev2_464, __p3_464)); \
49571   __ret_464 = __builtin_shufflevector(__ret_464, __ret_464, 1, 0); \
49572   __ret_464; \
49573 })
49574 #endif
49575 
49576 #ifdef __LITTLE_ENDIAN__
49577 #define vmlal_laneq_u16(__p0_465, __p1_465, __p2_465, __p3_465) __extension__ ({ \
49578   uint32x4_t __ret_465; \
49579   uint32x4_t __s0_465 = __p0_465; \
49580   uint16x4_t __s1_465 = __p1_465; \
49581   uint16x8_t __s2_465 = __p2_465; \
49582   __ret_465 = __s0_465 + vmull_u16(__s1_465, splat_laneq_u16(__s2_465, __p3_465)); \
49583   __ret_465; \
49584 })
49585 #else
49586 #define vmlal_laneq_u16(__p0_466, __p1_466, __p2_466, __p3_466) __extension__ ({ \
49587   uint32x4_t __ret_466; \
49588   uint32x4_t __s0_466 = __p0_466; \
49589   uint16x4_t __s1_466 = __p1_466; \
49590   uint16x8_t __s2_466 = __p2_466; \
49591   uint32x4_t __rev0_466;  __rev0_466 = __builtin_shufflevector(__s0_466, __s0_466, 3, 2, 1, 0); \
49592   uint16x4_t __rev1_466;  __rev1_466 = __builtin_shufflevector(__s1_466, __s1_466, 3, 2, 1, 0); \
49593   uint16x8_t __rev2_466;  __rev2_466 = __builtin_shufflevector(__s2_466, __s2_466, 7, 6, 5, 4, 3, 2, 1, 0); \
49594   __ret_466 = __rev0_466 + __noswap_vmull_u16(__rev1_466, __noswap_splat_laneq_u16(__rev2_466, __p3_466)); \
49595   __ret_466 = __builtin_shufflevector(__ret_466, __ret_466, 3, 2, 1, 0); \
49596   __ret_466; \
49597 })
49598 #endif
49599 
49600 #ifdef __LITTLE_ENDIAN__
49601 #define vmlal_laneq_s32(__p0_467, __p1_467, __p2_467, __p3_467) __extension__ ({ \
49602   int64x2_t __ret_467; \
49603   int64x2_t __s0_467 = __p0_467; \
49604   int32x2_t __s1_467 = __p1_467; \
49605   int32x4_t __s2_467 = __p2_467; \
49606   __ret_467 = __s0_467 + vmull_s32(__s1_467, splat_laneq_s32(__s2_467, __p3_467)); \
49607   __ret_467; \
49608 })
49609 #else
49610 #define vmlal_laneq_s32(__p0_468, __p1_468, __p2_468, __p3_468) __extension__ ({ \
49611   int64x2_t __ret_468; \
49612   int64x2_t __s0_468 = __p0_468; \
49613   int32x2_t __s1_468 = __p1_468; \
49614   int32x4_t __s2_468 = __p2_468; \
49615   int64x2_t __rev0_468;  __rev0_468 = __builtin_shufflevector(__s0_468, __s0_468, 1, 0); \
49616   int32x2_t __rev1_468;  __rev1_468 = __builtin_shufflevector(__s1_468, __s1_468, 1, 0); \
49617   int32x4_t __rev2_468;  __rev2_468 = __builtin_shufflevector(__s2_468, __s2_468, 3, 2, 1, 0); \
49618   __ret_468 = __rev0_468 + __noswap_vmull_s32(__rev1_468, __noswap_splat_laneq_s32(__rev2_468, __p3_468)); \
49619   __ret_468 = __builtin_shufflevector(__ret_468, __ret_468, 1, 0); \
49620   __ret_468; \
49621 })
49622 #endif
49623 
49624 #ifdef __LITTLE_ENDIAN__
49625 #define vmlal_laneq_s16(__p0_469, __p1_469, __p2_469, __p3_469) __extension__ ({ \
49626   int32x4_t __ret_469; \
49627   int32x4_t __s0_469 = __p0_469; \
49628   int16x4_t __s1_469 = __p1_469; \
49629   int16x8_t __s2_469 = __p2_469; \
49630   __ret_469 = __s0_469 + vmull_s16(__s1_469, splat_laneq_s16(__s2_469, __p3_469)); \
49631   __ret_469; \
49632 })
49633 #else
49634 #define vmlal_laneq_s16(__p0_470, __p1_470, __p2_470, __p3_470) __extension__ ({ \
49635   int32x4_t __ret_470; \
49636   int32x4_t __s0_470 = __p0_470; \
49637   int16x4_t __s1_470 = __p1_470; \
49638   int16x8_t __s2_470 = __p2_470; \
49639   int32x4_t __rev0_470;  __rev0_470 = __builtin_shufflevector(__s0_470, __s0_470, 3, 2, 1, 0); \
49640   int16x4_t __rev1_470;  __rev1_470 = __builtin_shufflevector(__s1_470, __s1_470, 3, 2, 1, 0); \
49641   int16x8_t __rev2_470;  __rev2_470 = __builtin_shufflevector(__s2_470, __s2_470, 7, 6, 5, 4, 3, 2, 1, 0); \
49642   __ret_470 = __rev0_470 + __noswap_vmull_s16(__rev1_470, __noswap_splat_laneq_s16(__rev2_470, __p3_470)); \
49643   __ret_470 = __builtin_shufflevector(__ret_470, __ret_470, 3, 2, 1, 0); \
49644   __ret_470; \
49645 })
49646 #endif
49647 
49648 #ifdef __LITTLE_ENDIAN__
vmlsq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)49649 __ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
49650   float64x2_t __ret;
49651   __ret = __p0 - __p1 * __p2;
49652   return __ret;
49653 }
49654 #else
vmlsq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)49655 __ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
49656   float64x2_t __ret;
49657   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
49658   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
49659   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
49660   __ret = __rev0 - __rev1 * __rev2;
49661   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
49662   return __ret;
49663 }
49664 #endif
49665 
vmls_f64(float64x1_t __p0,float64x1_t __p1,float64x1_t __p2)49666 __ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
49667   float64x1_t __ret;
49668   __ret = __p0 - __p1 * __p2;
49669   return __ret;
49670 }
49671 #ifdef __LITTLE_ENDIAN__
49672 #define vmlsq_laneq_u32(__p0_471, __p1_471, __p2_471, __p3_471) __extension__ ({ \
49673   uint32x4_t __ret_471; \
49674   uint32x4_t __s0_471 = __p0_471; \
49675   uint32x4_t __s1_471 = __p1_471; \
49676   uint32x4_t __s2_471 = __p2_471; \
49677   __ret_471 = __s0_471 - __s1_471 * splatq_laneq_u32(__s2_471, __p3_471); \
49678   __ret_471; \
49679 })
49680 #else
49681 #define vmlsq_laneq_u32(__p0_472, __p1_472, __p2_472, __p3_472) __extension__ ({ \
49682   uint32x4_t __ret_472; \
49683   uint32x4_t __s0_472 = __p0_472; \
49684   uint32x4_t __s1_472 = __p1_472; \
49685   uint32x4_t __s2_472 = __p2_472; \
49686   uint32x4_t __rev0_472;  __rev0_472 = __builtin_shufflevector(__s0_472, __s0_472, 3, 2, 1, 0); \
49687   uint32x4_t __rev1_472;  __rev1_472 = __builtin_shufflevector(__s1_472, __s1_472, 3, 2, 1, 0); \
49688   uint32x4_t __rev2_472;  __rev2_472 = __builtin_shufflevector(__s2_472, __s2_472, 3, 2, 1, 0); \
49689   __ret_472 = __rev0_472 - __rev1_472 * __noswap_splatq_laneq_u32(__rev2_472, __p3_472); \
49690   __ret_472 = __builtin_shufflevector(__ret_472, __ret_472, 3, 2, 1, 0); \
49691   __ret_472; \
49692 })
49693 #endif
49694 
49695 #ifdef __LITTLE_ENDIAN__
49696 #define vmlsq_laneq_u16(__p0_473, __p1_473, __p2_473, __p3_473) __extension__ ({ \
49697   uint16x8_t __ret_473; \
49698   uint16x8_t __s0_473 = __p0_473; \
49699   uint16x8_t __s1_473 = __p1_473; \
49700   uint16x8_t __s2_473 = __p2_473; \
49701   __ret_473 = __s0_473 - __s1_473 * splatq_laneq_u16(__s2_473, __p3_473); \
49702   __ret_473; \
49703 })
49704 #else
49705 #define vmlsq_laneq_u16(__p0_474, __p1_474, __p2_474, __p3_474) __extension__ ({ \
49706   uint16x8_t __ret_474; \
49707   uint16x8_t __s0_474 = __p0_474; \
49708   uint16x8_t __s1_474 = __p1_474; \
49709   uint16x8_t __s2_474 = __p2_474; \
49710   uint16x8_t __rev0_474;  __rev0_474 = __builtin_shufflevector(__s0_474, __s0_474, 7, 6, 5, 4, 3, 2, 1, 0); \
49711   uint16x8_t __rev1_474;  __rev1_474 = __builtin_shufflevector(__s1_474, __s1_474, 7, 6, 5, 4, 3, 2, 1, 0); \
49712   uint16x8_t __rev2_474;  __rev2_474 = __builtin_shufflevector(__s2_474, __s2_474, 7, 6, 5, 4, 3, 2, 1, 0); \
49713   __ret_474 = __rev0_474 - __rev1_474 * __noswap_splatq_laneq_u16(__rev2_474, __p3_474); \
49714   __ret_474 = __builtin_shufflevector(__ret_474, __ret_474, 7, 6, 5, 4, 3, 2, 1, 0); \
49715   __ret_474; \
49716 })
49717 #endif
49718 
49719 #ifdef __LITTLE_ENDIAN__
49720 #define vmlsq_laneq_f32(__p0_475, __p1_475, __p2_475, __p3_475) __extension__ ({ \
49721   float32x4_t __ret_475; \
49722   float32x4_t __s0_475 = __p0_475; \
49723   float32x4_t __s1_475 = __p1_475; \
49724   float32x4_t __s2_475 = __p2_475; \
49725   __ret_475 = __s0_475 - __s1_475 * splatq_laneq_f32(__s2_475, __p3_475); \
49726   __ret_475; \
49727 })
49728 #else
49729 #define vmlsq_laneq_f32(__p0_476, __p1_476, __p2_476, __p3_476) __extension__ ({ \
49730   float32x4_t __ret_476; \
49731   float32x4_t __s0_476 = __p0_476; \
49732   float32x4_t __s1_476 = __p1_476; \
49733   float32x4_t __s2_476 = __p2_476; \
49734   float32x4_t __rev0_476;  __rev0_476 = __builtin_shufflevector(__s0_476, __s0_476, 3, 2, 1, 0); \
49735   float32x4_t __rev1_476;  __rev1_476 = __builtin_shufflevector(__s1_476, __s1_476, 3, 2, 1, 0); \
49736   float32x4_t __rev2_476;  __rev2_476 = __builtin_shufflevector(__s2_476, __s2_476, 3, 2, 1, 0); \
49737   __ret_476 = __rev0_476 - __rev1_476 * __noswap_splatq_laneq_f32(__rev2_476, __p3_476); \
49738   __ret_476 = __builtin_shufflevector(__ret_476, __ret_476, 3, 2, 1, 0); \
49739   __ret_476; \
49740 })
49741 #endif
49742 
49743 #ifdef __LITTLE_ENDIAN__
49744 #define vmlsq_laneq_s32(__p0_477, __p1_477, __p2_477, __p3_477) __extension__ ({ \
49745   int32x4_t __ret_477; \
49746   int32x4_t __s0_477 = __p0_477; \
49747   int32x4_t __s1_477 = __p1_477; \
49748   int32x4_t __s2_477 = __p2_477; \
49749   __ret_477 = __s0_477 - __s1_477 * splatq_laneq_s32(__s2_477, __p3_477); \
49750   __ret_477; \
49751 })
49752 #else
49753 #define vmlsq_laneq_s32(__p0_478, __p1_478, __p2_478, __p3_478) __extension__ ({ \
49754   int32x4_t __ret_478; \
49755   int32x4_t __s0_478 = __p0_478; \
49756   int32x4_t __s1_478 = __p1_478; \
49757   int32x4_t __s2_478 = __p2_478; \
49758   int32x4_t __rev0_478;  __rev0_478 = __builtin_shufflevector(__s0_478, __s0_478, 3, 2, 1, 0); \
49759   int32x4_t __rev1_478;  __rev1_478 = __builtin_shufflevector(__s1_478, __s1_478, 3, 2, 1, 0); \
49760   int32x4_t __rev2_478;  __rev2_478 = __builtin_shufflevector(__s2_478, __s2_478, 3, 2, 1, 0); \
49761   __ret_478 = __rev0_478 - __rev1_478 * __noswap_splatq_laneq_s32(__rev2_478, __p3_478); \
49762   __ret_478 = __builtin_shufflevector(__ret_478, __ret_478, 3, 2, 1, 0); \
49763   __ret_478; \
49764 })
49765 #endif
49766 
49767 #ifdef __LITTLE_ENDIAN__
49768 #define vmlsq_laneq_s16(__p0_479, __p1_479, __p2_479, __p3_479) __extension__ ({ \
49769   int16x8_t __ret_479; \
49770   int16x8_t __s0_479 = __p0_479; \
49771   int16x8_t __s1_479 = __p1_479; \
49772   int16x8_t __s2_479 = __p2_479; \
49773   __ret_479 = __s0_479 - __s1_479 * splatq_laneq_s16(__s2_479, __p3_479); \
49774   __ret_479; \
49775 })
49776 #else
49777 #define vmlsq_laneq_s16(__p0_480, __p1_480, __p2_480, __p3_480) __extension__ ({ \
49778   int16x8_t __ret_480; \
49779   int16x8_t __s0_480 = __p0_480; \
49780   int16x8_t __s1_480 = __p1_480; \
49781   int16x8_t __s2_480 = __p2_480; \
49782   int16x8_t __rev0_480;  __rev0_480 = __builtin_shufflevector(__s0_480, __s0_480, 7, 6, 5, 4, 3, 2, 1, 0); \
49783   int16x8_t __rev1_480;  __rev1_480 = __builtin_shufflevector(__s1_480, __s1_480, 7, 6, 5, 4, 3, 2, 1, 0); \
49784   int16x8_t __rev2_480;  __rev2_480 = __builtin_shufflevector(__s2_480, __s2_480, 7, 6, 5, 4, 3, 2, 1, 0); \
49785   __ret_480 = __rev0_480 - __rev1_480 * __noswap_splatq_laneq_s16(__rev2_480, __p3_480); \
49786   __ret_480 = __builtin_shufflevector(__ret_480, __ret_480, 7, 6, 5, 4, 3, 2, 1, 0); \
49787   __ret_480; \
49788 })
49789 #endif
49790 
49791 #ifdef __LITTLE_ENDIAN__
49792 #define vmls_laneq_u32(__p0_481, __p1_481, __p2_481, __p3_481) __extension__ ({ \
49793   uint32x2_t __ret_481; \
49794   uint32x2_t __s0_481 = __p0_481; \
49795   uint32x2_t __s1_481 = __p1_481; \
49796   uint32x4_t __s2_481 = __p2_481; \
49797   __ret_481 = __s0_481 - __s1_481 * splat_laneq_u32(__s2_481, __p3_481); \
49798   __ret_481; \
49799 })
49800 #else
49801 #define vmls_laneq_u32(__p0_482, __p1_482, __p2_482, __p3_482) __extension__ ({ \
49802   uint32x2_t __ret_482; \
49803   uint32x2_t __s0_482 = __p0_482; \
49804   uint32x2_t __s1_482 = __p1_482; \
49805   uint32x4_t __s2_482 = __p2_482; \
49806   uint32x2_t __rev0_482;  __rev0_482 = __builtin_shufflevector(__s0_482, __s0_482, 1, 0); \
49807   uint32x2_t __rev1_482;  __rev1_482 = __builtin_shufflevector(__s1_482, __s1_482, 1, 0); \
49808   uint32x4_t __rev2_482;  __rev2_482 = __builtin_shufflevector(__s2_482, __s2_482, 3, 2, 1, 0); \
49809   __ret_482 = __rev0_482 - __rev1_482 * __noswap_splat_laneq_u32(__rev2_482, __p3_482); \
49810   __ret_482 = __builtin_shufflevector(__ret_482, __ret_482, 1, 0); \
49811   __ret_482; \
49812 })
49813 #endif
49814 
49815 #ifdef __LITTLE_ENDIAN__
49816 #define vmls_laneq_u16(__p0_483, __p1_483, __p2_483, __p3_483) __extension__ ({ \
49817   uint16x4_t __ret_483; \
49818   uint16x4_t __s0_483 = __p0_483; \
49819   uint16x4_t __s1_483 = __p1_483; \
49820   uint16x8_t __s2_483 = __p2_483; \
49821   __ret_483 = __s0_483 - __s1_483 * splat_laneq_u16(__s2_483, __p3_483); \
49822   __ret_483; \
49823 })
49824 #else
49825 #define vmls_laneq_u16(__p0_484, __p1_484, __p2_484, __p3_484) __extension__ ({ \
49826   uint16x4_t __ret_484; \
49827   uint16x4_t __s0_484 = __p0_484; \
49828   uint16x4_t __s1_484 = __p1_484; \
49829   uint16x8_t __s2_484 = __p2_484; \
49830   uint16x4_t __rev0_484;  __rev0_484 = __builtin_shufflevector(__s0_484, __s0_484, 3, 2, 1, 0); \
49831   uint16x4_t __rev1_484;  __rev1_484 = __builtin_shufflevector(__s1_484, __s1_484, 3, 2, 1, 0); \
49832   uint16x8_t __rev2_484;  __rev2_484 = __builtin_shufflevector(__s2_484, __s2_484, 7, 6, 5, 4, 3, 2, 1, 0); \
49833   __ret_484 = __rev0_484 - __rev1_484 * __noswap_splat_laneq_u16(__rev2_484, __p3_484); \
49834   __ret_484 = __builtin_shufflevector(__ret_484, __ret_484, 3, 2, 1, 0); \
49835   __ret_484; \
49836 })
49837 #endif
49838 
49839 #ifdef __LITTLE_ENDIAN__
49840 #define vmls_laneq_f32(__p0_485, __p1_485, __p2_485, __p3_485) __extension__ ({ \
49841   float32x2_t __ret_485; \
49842   float32x2_t __s0_485 = __p0_485; \
49843   float32x2_t __s1_485 = __p1_485; \
49844   float32x4_t __s2_485 = __p2_485; \
49845   __ret_485 = __s0_485 - __s1_485 * splat_laneq_f32(__s2_485, __p3_485); \
49846   __ret_485; \
49847 })
49848 #else
49849 #define vmls_laneq_f32(__p0_486, __p1_486, __p2_486, __p3_486) __extension__ ({ \
49850   float32x2_t __ret_486; \
49851   float32x2_t __s0_486 = __p0_486; \
49852   float32x2_t __s1_486 = __p1_486; \
49853   float32x4_t __s2_486 = __p2_486; \
49854   float32x2_t __rev0_486;  __rev0_486 = __builtin_shufflevector(__s0_486, __s0_486, 1, 0); \
49855   float32x2_t __rev1_486;  __rev1_486 = __builtin_shufflevector(__s1_486, __s1_486, 1, 0); \
49856   float32x4_t __rev2_486;  __rev2_486 = __builtin_shufflevector(__s2_486, __s2_486, 3, 2, 1, 0); \
49857   __ret_486 = __rev0_486 - __rev1_486 * __noswap_splat_laneq_f32(__rev2_486, __p3_486); \
49858   __ret_486 = __builtin_shufflevector(__ret_486, __ret_486, 1, 0); \
49859   __ret_486; \
49860 })
49861 #endif
49862 
49863 #ifdef __LITTLE_ENDIAN__
49864 #define vmls_laneq_s32(__p0_487, __p1_487, __p2_487, __p3_487) __extension__ ({ \
49865   int32x2_t __ret_487; \
49866   int32x2_t __s0_487 = __p0_487; \
49867   int32x2_t __s1_487 = __p1_487; \
49868   int32x4_t __s2_487 = __p2_487; \
49869   __ret_487 = __s0_487 - __s1_487 * splat_laneq_s32(__s2_487, __p3_487); \
49870   __ret_487; \
49871 })
49872 #else
49873 #define vmls_laneq_s32(__p0_488, __p1_488, __p2_488, __p3_488) __extension__ ({ \
49874   int32x2_t __ret_488; \
49875   int32x2_t __s0_488 = __p0_488; \
49876   int32x2_t __s1_488 = __p1_488; \
49877   int32x4_t __s2_488 = __p2_488; \
49878   int32x2_t __rev0_488;  __rev0_488 = __builtin_shufflevector(__s0_488, __s0_488, 1, 0); \
49879   int32x2_t __rev1_488;  __rev1_488 = __builtin_shufflevector(__s1_488, __s1_488, 1, 0); \
49880   int32x4_t __rev2_488;  __rev2_488 = __builtin_shufflevector(__s2_488, __s2_488, 3, 2, 1, 0); \
49881   __ret_488 = __rev0_488 - __rev1_488 * __noswap_splat_laneq_s32(__rev2_488, __p3_488); \
49882   __ret_488 = __builtin_shufflevector(__ret_488, __ret_488, 1, 0); \
49883   __ret_488; \
49884 })
49885 #endif
49886 
49887 #ifdef __LITTLE_ENDIAN__
49888 #define vmls_laneq_s16(__p0_489, __p1_489, __p2_489, __p3_489) __extension__ ({ \
49889   int16x4_t __ret_489; \
49890   int16x4_t __s0_489 = __p0_489; \
49891   int16x4_t __s1_489 = __p1_489; \
49892   int16x8_t __s2_489 = __p2_489; \
49893   __ret_489 = __s0_489 - __s1_489 * splat_laneq_s16(__s2_489, __p3_489); \
49894   __ret_489; \
49895 })
49896 #else
49897 #define vmls_laneq_s16(__p0_490, __p1_490, __p2_490, __p3_490) __extension__ ({ \
49898   int16x4_t __ret_490; \
49899   int16x4_t __s0_490 = __p0_490; \
49900   int16x4_t __s1_490 = __p1_490; \
49901   int16x8_t __s2_490 = __p2_490; \
49902   int16x4_t __rev0_490;  __rev0_490 = __builtin_shufflevector(__s0_490, __s0_490, 3, 2, 1, 0); \
49903   int16x4_t __rev1_490;  __rev1_490 = __builtin_shufflevector(__s1_490, __s1_490, 3, 2, 1, 0); \
49904   int16x8_t __rev2_490;  __rev2_490 = __builtin_shufflevector(__s2_490, __s2_490, 7, 6, 5, 4, 3, 2, 1, 0); \
49905   __ret_490 = __rev0_490 - __rev1_490 * __noswap_splat_laneq_s16(__rev2_490, __p3_490); \
49906   __ret_490 = __builtin_shufflevector(__ret_490, __ret_490, 3, 2, 1, 0); \
49907   __ret_490; \
49908 })
49909 #endif
49910 
49911 #ifdef __LITTLE_ENDIAN__
49912 #define vmlsl_high_lane_u32(__p0_491, __p1_491, __p2_491, __p3_491) __extension__ ({ \
49913   uint64x2_t __ret_491; \
49914   uint64x2_t __s0_491 = __p0_491; \
49915   uint32x4_t __s1_491 = __p1_491; \
49916   uint32x2_t __s2_491 = __p2_491; \
49917   __ret_491 = __s0_491 - vmull_u32(vget_high_u32(__s1_491), splat_lane_u32(__s2_491, __p3_491)); \
49918   __ret_491; \
49919 })
49920 #else
49921 #define vmlsl_high_lane_u32(__p0_492, __p1_492, __p2_492, __p3_492) __extension__ ({ \
49922   uint64x2_t __ret_492; \
49923   uint64x2_t __s0_492 = __p0_492; \
49924   uint32x4_t __s1_492 = __p1_492; \
49925   uint32x2_t __s2_492 = __p2_492; \
49926   uint64x2_t __rev0_492;  __rev0_492 = __builtin_shufflevector(__s0_492, __s0_492, 1, 0); \
49927   uint32x4_t __rev1_492;  __rev1_492 = __builtin_shufflevector(__s1_492, __s1_492, 3, 2, 1, 0); \
49928   uint32x2_t __rev2_492;  __rev2_492 = __builtin_shufflevector(__s2_492, __s2_492, 1, 0); \
49929   __ret_492 = __rev0_492 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1_492), __noswap_splat_lane_u32(__rev2_492, __p3_492)); \
49930   __ret_492 = __builtin_shufflevector(__ret_492, __ret_492, 1, 0); \
49931   __ret_492; \
49932 })
49933 #endif
49934 
49935 #ifdef __LITTLE_ENDIAN__
49936 #define vmlsl_high_lane_u16(__p0_493, __p1_493, __p2_493, __p3_493) __extension__ ({ \
49937   uint32x4_t __ret_493; \
49938   uint32x4_t __s0_493 = __p0_493; \
49939   uint16x8_t __s1_493 = __p1_493; \
49940   uint16x4_t __s2_493 = __p2_493; \
49941   __ret_493 = __s0_493 - vmull_u16(vget_high_u16(__s1_493), splat_lane_u16(__s2_493, __p3_493)); \
49942   __ret_493; \
49943 })
49944 #else
49945 #define vmlsl_high_lane_u16(__p0_494, __p1_494, __p2_494, __p3_494) __extension__ ({ \
49946   uint32x4_t __ret_494; \
49947   uint32x4_t __s0_494 = __p0_494; \
49948   uint16x8_t __s1_494 = __p1_494; \
49949   uint16x4_t __s2_494 = __p2_494; \
49950   uint32x4_t __rev0_494;  __rev0_494 = __builtin_shufflevector(__s0_494, __s0_494, 3, 2, 1, 0); \
49951   uint16x8_t __rev1_494;  __rev1_494 = __builtin_shufflevector(__s1_494, __s1_494, 7, 6, 5, 4, 3, 2, 1, 0); \
49952   uint16x4_t __rev2_494;  __rev2_494 = __builtin_shufflevector(__s2_494, __s2_494, 3, 2, 1, 0); \
49953   __ret_494 = __rev0_494 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1_494), __noswap_splat_lane_u16(__rev2_494, __p3_494)); \
49954   __ret_494 = __builtin_shufflevector(__ret_494, __ret_494, 3, 2, 1, 0); \
49955   __ret_494; \
49956 })
49957 #endif
49958 
49959 #ifdef __LITTLE_ENDIAN__
49960 #define vmlsl_high_lane_s32(__p0_495, __p1_495, __p2_495, __p3_495) __extension__ ({ \
49961   int64x2_t __ret_495; \
49962   int64x2_t __s0_495 = __p0_495; \
49963   int32x4_t __s1_495 = __p1_495; \
49964   int32x2_t __s2_495 = __p2_495; \
49965   __ret_495 = __s0_495 - vmull_s32(vget_high_s32(__s1_495), splat_lane_s32(__s2_495, __p3_495)); \
49966   __ret_495; \
49967 })
49968 #else
49969 #define vmlsl_high_lane_s32(__p0_496, __p1_496, __p2_496, __p3_496) __extension__ ({ \
49970   int64x2_t __ret_496; \
49971   int64x2_t __s0_496 = __p0_496; \
49972   int32x4_t __s1_496 = __p1_496; \
49973   int32x2_t __s2_496 = __p2_496; \
49974   int64x2_t __rev0_496;  __rev0_496 = __builtin_shufflevector(__s0_496, __s0_496, 1, 0); \
49975   int32x4_t __rev1_496;  __rev1_496 = __builtin_shufflevector(__s1_496, __s1_496, 3, 2, 1, 0); \
49976   int32x2_t __rev2_496;  __rev2_496 = __builtin_shufflevector(__s2_496, __s2_496, 1, 0); \
49977   __ret_496 = __rev0_496 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1_496), __noswap_splat_lane_s32(__rev2_496, __p3_496)); \
49978   __ret_496 = __builtin_shufflevector(__ret_496, __ret_496, 1, 0); \
49979   __ret_496; \
49980 })
49981 #endif
49982 
49983 #ifdef __LITTLE_ENDIAN__
49984 #define vmlsl_high_lane_s16(__p0_497, __p1_497, __p2_497, __p3_497) __extension__ ({ \
49985   int32x4_t __ret_497; \
49986   int32x4_t __s0_497 = __p0_497; \
49987   int16x8_t __s1_497 = __p1_497; \
49988   int16x4_t __s2_497 = __p2_497; \
49989   __ret_497 = __s0_497 - vmull_s16(vget_high_s16(__s1_497), splat_lane_s16(__s2_497, __p3_497)); \
49990   __ret_497; \
49991 })
49992 #else
49993 #define vmlsl_high_lane_s16(__p0_498, __p1_498, __p2_498, __p3_498) __extension__ ({ \
49994   int32x4_t __ret_498; \
49995   int32x4_t __s0_498 = __p0_498; \
49996   int16x8_t __s1_498 = __p1_498; \
49997   int16x4_t __s2_498 = __p2_498; \
49998   int32x4_t __rev0_498;  __rev0_498 = __builtin_shufflevector(__s0_498, __s0_498, 3, 2, 1, 0); \
49999   int16x8_t __rev1_498;  __rev1_498 = __builtin_shufflevector(__s1_498, __s1_498, 7, 6, 5, 4, 3, 2, 1, 0); \
50000   int16x4_t __rev2_498;  __rev2_498 = __builtin_shufflevector(__s2_498, __s2_498, 3, 2, 1, 0); \
50001   __ret_498 = __rev0_498 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1_498), __noswap_splat_lane_s16(__rev2_498, __p3_498)); \
50002   __ret_498 = __builtin_shufflevector(__ret_498, __ret_498, 3, 2, 1, 0); \
50003   __ret_498; \
50004 })
50005 #endif
50006 
50007 #ifdef __LITTLE_ENDIAN__
50008 #define vmlsl_high_laneq_u32(__p0_499, __p1_499, __p2_499, __p3_499) __extension__ ({ \
50009   uint64x2_t __ret_499; \
50010   uint64x2_t __s0_499 = __p0_499; \
50011   uint32x4_t __s1_499 = __p1_499; \
50012   uint32x4_t __s2_499 = __p2_499; \
50013   __ret_499 = __s0_499 - vmull_u32(vget_high_u32(__s1_499), splat_laneq_u32(__s2_499, __p3_499)); \
50014   __ret_499; \
50015 })
50016 #else
50017 #define vmlsl_high_laneq_u32(__p0_500, __p1_500, __p2_500, __p3_500) __extension__ ({ \
50018   uint64x2_t __ret_500; \
50019   uint64x2_t __s0_500 = __p0_500; \
50020   uint32x4_t __s1_500 = __p1_500; \
50021   uint32x4_t __s2_500 = __p2_500; \
50022   uint64x2_t __rev0_500;  __rev0_500 = __builtin_shufflevector(__s0_500, __s0_500, 1, 0); \
50023   uint32x4_t __rev1_500;  __rev1_500 = __builtin_shufflevector(__s1_500, __s1_500, 3, 2, 1, 0); \
50024   uint32x4_t __rev2_500;  __rev2_500 = __builtin_shufflevector(__s2_500, __s2_500, 3, 2, 1, 0); \
50025   __ret_500 = __rev0_500 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1_500), __noswap_splat_laneq_u32(__rev2_500, __p3_500)); \
50026   __ret_500 = __builtin_shufflevector(__ret_500, __ret_500, 1, 0); \
50027   __ret_500; \
50028 })
50029 #endif
50030 
50031 #ifdef __LITTLE_ENDIAN__
50032 #define vmlsl_high_laneq_u16(__p0_501, __p1_501, __p2_501, __p3_501) __extension__ ({ \
50033   uint32x4_t __ret_501; \
50034   uint32x4_t __s0_501 = __p0_501; \
50035   uint16x8_t __s1_501 = __p1_501; \
50036   uint16x8_t __s2_501 = __p2_501; \
50037   __ret_501 = __s0_501 - vmull_u16(vget_high_u16(__s1_501), splat_laneq_u16(__s2_501, __p3_501)); \
50038   __ret_501; \
50039 })
50040 #else
50041 #define vmlsl_high_laneq_u16(__p0_502, __p1_502, __p2_502, __p3_502) __extension__ ({ \
50042   uint32x4_t __ret_502; \
50043   uint32x4_t __s0_502 = __p0_502; \
50044   uint16x8_t __s1_502 = __p1_502; \
50045   uint16x8_t __s2_502 = __p2_502; \
50046   uint32x4_t __rev0_502;  __rev0_502 = __builtin_shufflevector(__s0_502, __s0_502, 3, 2, 1, 0); \
50047   uint16x8_t __rev1_502;  __rev1_502 = __builtin_shufflevector(__s1_502, __s1_502, 7, 6, 5, 4, 3, 2, 1, 0); \
50048   uint16x8_t __rev2_502;  __rev2_502 = __builtin_shufflevector(__s2_502, __s2_502, 7, 6, 5, 4, 3, 2, 1, 0); \
50049   __ret_502 = __rev0_502 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1_502), __noswap_splat_laneq_u16(__rev2_502, __p3_502)); \
50050   __ret_502 = __builtin_shufflevector(__ret_502, __ret_502, 3, 2, 1, 0); \
50051   __ret_502; \
50052 })
50053 #endif
50054 
50055 #ifdef __LITTLE_ENDIAN__
50056 #define vmlsl_high_laneq_s32(__p0_503, __p1_503, __p2_503, __p3_503) __extension__ ({ \
50057   int64x2_t __ret_503; \
50058   int64x2_t __s0_503 = __p0_503; \
50059   int32x4_t __s1_503 = __p1_503; \
50060   int32x4_t __s2_503 = __p2_503; \
50061   __ret_503 = __s0_503 - vmull_s32(vget_high_s32(__s1_503), splat_laneq_s32(__s2_503, __p3_503)); \
50062   __ret_503; \
50063 })
50064 #else
50065 #define vmlsl_high_laneq_s32(__p0_504, __p1_504, __p2_504, __p3_504) __extension__ ({ \
50066   int64x2_t __ret_504; \
50067   int64x2_t __s0_504 = __p0_504; \
50068   int32x4_t __s1_504 = __p1_504; \
50069   int32x4_t __s2_504 = __p2_504; \
50070   int64x2_t __rev0_504;  __rev0_504 = __builtin_shufflevector(__s0_504, __s0_504, 1, 0); \
50071   int32x4_t __rev1_504;  __rev1_504 = __builtin_shufflevector(__s1_504, __s1_504, 3, 2, 1, 0); \
50072   int32x4_t __rev2_504;  __rev2_504 = __builtin_shufflevector(__s2_504, __s2_504, 3, 2, 1, 0); \
50073   __ret_504 = __rev0_504 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1_504), __noswap_splat_laneq_s32(__rev2_504, __p3_504)); \
50074   __ret_504 = __builtin_shufflevector(__ret_504, __ret_504, 1, 0); \
50075   __ret_504; \
50076 })
50077 #endif
50078 
50079 #ifdef __LITTLE_ENDIAN__
50080 #define vmlsl_high_laneq_s16(__p0_505, __p1_505, __p2_505, __p3_505) __extension__ ({ \
50081   int32x4_t __ret_505; \
50082   int32x4_t __s0_505 = __p0_505; \
50083   int16x8_t __s1_505 = __p1_505; \
50084   int16x8_t __s2_505 = __p2_505; \
50085   __ret_505 = __s0_505 - vmull_s16(vget_high_s16(__s1_505), splat_laneq_s16(__s2_505, __p3_505)); \
50086   __ret_505; \
50087 })
50088 #else
50089 #define vmlsl_high_laneq_s16(__p0_506, __p1_506, __p2_506, __p3_506) __extension__ ({ \
50090   int32x4_t __ret_506; \
50091   int32x4_t __s0_506 = __p0_506; \
50092   int16x8_t __s1_506 = __p1_506; \
50093   int16x8_t __s2_506 = __p2_506; \
50094   int32x4_t __rev0_506;  __rev0_506 = __builtin_shufflevector(__s0_506, __s0_506, 3, 2, 1, 0); \
50095   int16x8_t __rev1_506;  __rev1_506 = __builtin_shufflevector(__s1_506, __s1_506, 7, 6, 5, 4, 3, 2, 1, 0); \
50096   int16x8_t __rev2_506;  __rev2_506 = __builtin_shufflevector(__s2_506, __s2_506, 7, 6, 5, 4, 3, 2, 1, 0); \
50097   __ret_506 = __rev0_506 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1_506), __noswap_splat_laneq_s16(__rev2_506, __p3_506)); \
50098   __ret_506 = __builtin_shufflevector(__ret_506, __ret_506, 3, 2, 1, 0); \
50099   __ret_506; \
50100 })
50101 #endif
50102 
50103 #ifdef __LITTLE_ENDIAN__
50104 #define vmlsl_laneq_u32(__p0_507, __p1_507, __p2_507, __p3_507) __extension__ ({ \
50105   uint64x2_t __ret_507; \
50106   uint64x2_t __s0_507 = __p0_507; \
50107   uint32x2_t __s1_507 = __p1_507; \
50108   uint32x4_t __s2_507 = __p2_507; \
50109   __ret_507 = __s0_507 - vmull_u32(__s1_507, splat_laneq_u32(__s2_507, __p3_507)); \
50110   __ret_507; \
50111 })
50112 #else
50113 #define vmlsl_laneq_u32(__p0_508, __p1_508, __p2_508, __p3_508) __extension__ ({ \
50114   uint64x2_t __ret_508; \
50115   uint64x2_t __s0_508 = __p0_508; \
50116   uint32x2_t __s1_508 = __p1_508; \
50117   uint32x4_t __s2_508 = __p2_508; \
50118   uint64x2_t __rev0_508;  __rev0_508 = __builtin_shufflevector(__s0_508, __s0_508, 1, 0); \
50119   uint32x2_t __rev1_508;  __rev1_508 = __builtin_shufflevector(__s1_508, __s1_508, 1, 0); \
50120   uint32x4_t __rev2_508;  __rev2_508 = __builtin_shufflevector(__s2_508, __s2_508, 3, 2, 1, 0); \
50121   __ret_508 = __rev0_508 - __noswap_vmull_u32(__rev1_508, __noswap_splat_laneq_u32(__rev2_508, __p3_508)); \
50122   __ret_508 = __builtin_shufflevector(__ret_508, __ret_508, 1, 0); \
50123   __ret_508; \
50124 })
50125 #endif
50126 
50127 #ifdef __LITTLE_ENDIAN__
50128 #define vmlsl_laneq_u16(__p0_509, __p1_509, __p2_509, __p3_509) __extension__ ({ \
50129   uint32x4_t __ret_509; \
50130   uint32x4_t __s0_509 = __p0_509; \
50131   uint16x4_t __s1_509 = __p1_509; \
50132   uint16x8_t __s2_509 = __p2_509; \
50133   __ret_509 = __s0_509 - vmull_u16(__s1_509, splat_laneq_u16(__s2_509, __p3_509)); \
50134   __ret_509; \
50135 })
50136 #else
50137 #define vmlsl_laneq_u16(__p0_510, __p1_510, __p2_510, __p3_510) __extension__ ({ \
50138   uint32x4_t __ret_510; \
50139   uint32x4_t __s0_510 = __p0_510; \
50140   uint16x4_t __s1_510 = __p1_510; \
50141   uint16x8_t __s2_510 = __p2_510; \
50142   uint32x4_t __rev0_510;  __rev0_510 = __builtin_shufflevector(__s0_510, __s0_510, 3, 2, 1, 0); \
50143   uint16x4_t __rev1_510;  __rev1_510 = __builtin_shufflevector(__s1_510, __s1_510, 3, 2, 1, 0); \
50144   uint16x8_t __rev2_510;  __rev2_510 = __builtin_shufflevector(__s2_510, __s2_510, 7, 6, 5, 4, 3, 2, 1, 0); \
50145   __ret_510 = __rev0_510 - __noswap_vmull_u16(__rev1_510, __noswap_splat_laneq_u16(__rev2_510, __p3_510)); \
50146   __ret_510 = __builtin_shufflevector(__ret_510, __ret_510, 3, 2, 1, 0); \
50147   __ret_510; \
50148 })
50149 #endif
50150 
50151 #ifdef __LITTLE_ENDIAN__
50152 #define vmlsl_laneq_s32(__p0_511, __p1_511, __p2_511, __p3_511) __extension__ ({ \
50153   int64x2_t __ret_511; \
50154   int64x2_t __s0_511 = __p0_511; \
50155   int32x2_t __s1_511 = __p1_511; \
50156   int32x4_t __s2_511 = __p2_511; \
50157   __ret_511 = __s0_511 - vmull_s32(__s1_511, splat_laneq_s32(__s2_511, __p3_511)); \
50158   __ret_511; \
50159 })
50160 #else
50161 #define vmlsl_laneq_s32(__p0_512, __p1_512, __p2_512, __p3_512) __extension__ ({ \
50162   int64x2_t __ret_512; \
50163   int64x2_t __s0_512 = __p0_512; \
50164   int32x2_t __s1_512 = __p1_512; \
50165   int32x4_t __s2_512 = __p2_512; \
50166   int64x2_t __rev0_512;  __rev0_512 = __builtin_shufflevector(__s0_512, __s0_512, 1, 0); \
50167   int32x2_t __rev1_512;  __rev1_512 = __builtin_shufflevector(__s1_512, __s1_512, 1, 0); \
50168   int32x4_t __rev2_512;  __rev2_512 = __builtin_shufflevector(__s2_512, __s2_512, 3, 2, 1, 0); \
50169   __ret_512 = __rev0_512 - __noswap_vmull_s32(__rev1_512, __noswap_splat_laneq_s32(__rev2_512, __p3_512)); \
50170   __ret_512 = __builtin_shufflevector(__ret_512, __ret_512, 1, 0); \
50171   __ret_512; \
50172 })
50173 #endif
50174 
50175 #ifdef __LITTLE_ENDIAN__
50176 #define vmlsl_laneq_s16(__p0_513, __p1_513, __p2_513, __p3_513) __extension__ ({ \
50177   int32x4_t __ret_513; \
50178   int32x4_t __s0_513 = __p0_513; \
50179   int16x4_t __s1_513 = __p1_513; \
50180   int16x8_t __s2_513 = __p2_513; \
50181   __ret_513 = __s0_513 - vmull_s16(__s1_513, splat_laneq_s16(__s2_513, __p3_513)); \
50182   __ret_513; \
50183 })
50184 #else
50185 #define vmlsl_laneq_s16(__p0_514, __p1_514, __p2_514, __p3_514) __extension__ ({ \
50186   int32x4_t __ret_514; \
50187   int32x4_t __s0_514 = __p0_514; \
50188   int16x4_t __s1_514 = __p1_514; \
50189   int16x8_t __s2_514 = __p2_514; \
50190   int32x4_t __rev0_514;  __rev0_514 = __builtin_shufflevector(__s0_514, __s0_514, 3, 2, 1, 0); \
50191   int16x4_t __rev1_514;  __rev1_514 = __builtin_shufflevector(__s1_514, __s1_514, 3, 2, 1, 0); \
50192   int16x8_t __rev2_514;  __rev2_514 = __builtin_shufflevector(__s2_514, __s2_514, 7, 6, 5, 4, 3, 2, 1, 0); \
50193   __ret_514 = __rev0_514 - __noswap_vmull_s16(__rev1_514, __noswap_splat_laneq_s16(__rev2_514, __p3_514)); \
50194   __ret_514 = __builtin_shufflevector(__ret_514, __ret_514, 3, 2, 1, 0); \
50195   __ret_514; \
50196 })
50197 #endif
50198 
vmov_n_p64(poly64_t __p0)50199 __ai poly64x1_t vmov_n_p64(poly64_t __p0) {
50200   poly64x1_t __ret;
50201   __ret = (poly64x1_t) {__p0};
50202   return __ret;
50203 }
50204 #ifdef __LITTLE_ENDIAN__
vmovq_n_p64(poly64_t __p0)50205 __ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
50206   poly64x2_t __ret;
50207   __ret = (poly64x2_t) {__p0, __p0};
50208   return __ret;
50209 }
50210 #else
vmovq_n_p64(poly64_t __p0)50211 __ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
50212   poly64x2_t __ret;
50213   __ret = (poly64x2_t) {__p0, __p0};
50214   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50215   return __ret;
50216 }
50217 #endif
50218 
50219 #ifdef __LITTLE_ENDIAN__
vmovq_n_f64(float64_t __p0)50220 __ai float64x2_t vmovq_n_f64(float64_t __p0) {
50221   float64x2_t __ret;
50222   __ret = (float64x2_t) {__p0, __p0};
50223   return __ret;
50224 }
50225 #else
vmovq_n_f64(float64_t __p0)50226 __ai float64x2_t vmovq_n_f64(float64_t __p0) {
50227   float64x2_t __ret;
50228   __ret = (float64x2_t) {__p0, __p0};
50229   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50230   return __ret;
50231 }
50232 #endif
50233 
vmov_n_f64(float64_t __p0)50234 __ai float64x1_t vmov_n_f64(float64_t __p0) {
50235   float64x1_t __ret;
50236   __ret = (float64x1_t) {__p0};
50237   return __ret;
50238 }
50239 #ifdef __LITTLE_ENDIAN__
vmovl_high_u8(uint8x16_t __p0_515)50240 __ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_515) {
50241   uint16x8_t __ret_515;
50242   uint8x8_t __a1_515 = vget_high_u8(__p0_515);
50243   __ret_515 = (uint16x8_t)(vshll_n_u8(__a1_515, 0));
50244   return __ret_515;
50245 }
50246 #else
vmovl_high_u8(uint8x16_t __p0_516)50247 __ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_516) {
50248   uint16x8_t __ret_516;
50249   uint8x16_t __rev0_516;  __rev0_516 = __builtin_shufflevector(__p0_516, __p0_516, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50250   uint8x8_t __a1_516 = __noswap_vget_high_u8(__rev0_516);
50251   __ret_516 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_516, 0));
50252   __ret_516 = __builtin_shufflevector(__ret_516, __ret_516, 7, 6, 5, 4, 3, 2, 1, 0);
50253   return __ret_516;
50254 }
__noswap_vmovl_high_u8(uint8x16_t __p0_517)50255 __ai uint16x8_t __noswap_vmovl_high_u8(uint8x16_t __p0_517) {
50256   uint16x8_t __ret_517;
50257   uint8x8_t __a1_517 = __noswap_vget_high_u8(__p0_517);
50258   __ret_517 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_517, 0));
50259   return __ret_517;
50260 }
50261 #endif
50262 
50263 #ifdef __LITTLE_ENDIAN__
vmovl_high_u32(uint32x4_t __p0_518)50264 __ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_518) {
50265   uint64x2_t __ret_518;
50266   uint32x2_t __a1_518 = vget_high_u32(__p0_518);
50267   __ret_518 = (uint64x2_t)(vshll_n_u32(__a1_518, 0));
50268   return __ret_518;
50269 }
50270 #else
vmovl_high_u32(uint32x4_t __p0_519)50271 __ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_519) {
50272   uint64x2_t __ret_519;
50273   uint32x4_t __rev0_519;  __rev0_519 = __builtin_shufflevector(__p0_519, __p0_519, 3, 2, 1, 0);
50274   uint32x2_t __a1_519 = __noswap_vget_high_u32(__rev0_519);
50275   __ret_519 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_519, 0));
50276   __ret_519 = __builtin_shufflevector(__ret_519, __ret_519, 1, 0);
50277   return __ret_519;
50278 }
__noswap_vmovl_high_u32(uint32x4_t __p0_520)50279 __ai uint64x2_t __noswap_vmovl_high_u32(uint32x4_t __p0_520) {
50280   uint64x2_t __ret_520;
50281   uint32x2_t __a1_520 = __noswap_vget_high_u32(__p0_520);
50282   __ret_520 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_520, 0));
50283   return __ret_520;
50284 }
50285 #endif
50286 
50287 #ifdef __LITTLE_ENDIAN__
vmovl_high_u16(uint16x8_t __p0_521)50288 __ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_521) {
50289   uint32x4_t __ret_521;
50290   uint16x4_t __a1_521 = vget_high_u16(__p0_521);
50291   __ret_521 = (uint32x4_t)(vshll_n_u16(__a1_521, 0));
50292   return __ret_521;
50293 }
50294 #else
vmovl_high_u16(uint16x8_t __p0_522)50295 __ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_522) {
50296   uint32x4_t __ret_522;
50297   uint16x8_t __rev0_522;  __rev0_522 = __builtin_shufflevector(__p0_522, __p0_522, 7, 6, 5, 4, 3, 2, 1, 0);
50298   uint16x4_t __a1_522 = __noswap_vget_high_u16(__rev0_522);
50299   __ret_522 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_522, 0));
50300   __ret_522 = __builtin_shufflevector(__ret_522, __ret_522, 3, 2, 1, 0);
50301   return __ret_522;
50302 }
__noswap_vmovl_high_u16(uint16x8_t __p0_523)50303 __ai uint32x4_t __noswap_vmovl_high_u16(uint16x8_t __p0_523) {
50304   uint32x4_t __ret_523;
50305   uint16x4_t __a1_523 = __noswap_vget_high_u16(__p0_523);
50306   __ret_523 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_523, 0));
50307   return __ret_523;
50308 }
50309 #endif
50310 
50311 #ifdef __LITTLE_ENDIAN__
vmovl_high_s8(int8x16_t __p0_524)50312 __ai int16x8_t vmovl_high_s8(int8x16_t __p0_524) {
50313   int16x8_t __ret_524;
50314   int8x8_t __a1_524 = vget_high_s8(__p0_524);
50315   __ret_524 = (int16x8_t)(vshll_n_s8(__a1_524, 0));
50316   return __ret_524;
50317 }
50318 #else
vmovl_high_s8(int8x16_t __p0_525)50319 __ai int16x8_t vmovl_high_s8(int8x16_t __p0_525) {
50320   int16x8_t __ret_525;
50321   int8x16_t __rev0_525;  __rev0_525 = __builtin_shufflevector(__p0_525, __p0_525, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50322   int8x8_t __a1_525 = __noswap_vget_high_s8(__rev0_525);
50323   __ret_525 = (int16x8_t)(__noswap_vshll_n_s8(__a1_525, 0));
50324   __ret_525 = __builtin_shufflevector(__ret_525, __ret_525, 7, 6, 5, 4, 3, 2, 1, 0);
50325   return __ret_525;
50326 }
__noswap_vmovl_high_s8(int8x16_t __p0_526)50327 __ai int16x8_t __noswap_vmovl_high_s8(int8x16_t __p0_526) {
50328   int16x8_t __ret_526;
50329   int8x8_t __a1_526 = __noswap_vget_high_s8(__p0_526);
50330   __ret_526 = (int16x8_t)(__noswap_vshll_n_s8(__a1_526, 0));
50331   return __ret_526;
50332 }
50333 #endif
50334 
50335 #ifdef __LITTLE_ENDIAN__
vmovl_high_s32(int32x4_t __p0_527)50336 __ai int64x2_t vmovl_high_s32(int32x4_t __p0_527) {
50337   int64x2_t __ret_527;
50338   int32x2_t __a1_527 = vget_high_s32(__p0_527);
50339   __ret_527 = (int64x2_t)(vshll_n_s32(__a1_527, 0));
50340   return __ret_527;
50341 }
50342 #else
vmovl_high_s32(int32x4_t __p0_528)50343 __ai int64x2_t vmovl_high_s32(int32x4_t __p0_528) {
50344   int64x2_t __ret_528;
50345   int32x4_t __rev0_528;  __rev0_528 = __builtin_shufflevector(__p0_528, __p0_528, 3, 2, 1, 0);
50346   int32x2_t __a1_528 = __noswap_vget_high_s32(__rev0_528);
50347   __ret_528 = (int64x2_t)(__noswap_vshll_n_s32(__a1_528, 0));
50348   __ret_528 = __builtin_shufflevector(__ret_528, __ret_528, 1, 0);
50349   return __ret_528;
50350 }
__noswap_vmovl_high_s32(int32x4_t __p0_529)50351 __ai int64x2_t __noswap_vmovl_high_s32(int32x4_t __p0_529) {
50352   int64x2_t __ret_529;
50353   int32x2_t __a1_529 = __noswap_vget_high_s32(__p0_529);
50354   __ret_529 = (int64x2_t)(__noswap_vshll_n_s32(__a1_529, 0));
50355   return __ret_529;
50356 }
50357 #endif
50358 
50359 #ifdef __LITTLE_ENDIAN__
vmovl_high_s16(int16x8_t __p0_530)50360 __ai int32x4_t vmovl_high_s16(int16x8_t __p0_530) {
50361   int32x4_t __ret_530;
50362   int16x4_t __a1_530 = vget_high_s16(__p0_530);
50363   __ret_530 = (int32x4_t)(vshll_n_s16(__a1_530, 0));
50364   return __ret_530;
50365 }
50366 #else
vmovl_high_s16(int16x8_t __p0_531)50367 __ai int32x4_t vmovl_high_s16(int16x8_t __p0_531) {
50368   int32x4_t __ret_531;
50369   int16x8_t __rev0_531;  __rev0_531 = __builtin_shufflevector(__p0_531, __p0_531, 7, 6, 5, 4, 3, 2, 1, 0);
50370   int16x4_t __a1_531 = __noswap_vget_high_s16(__rev0_531);
50371   __ret_531 = (int32x4_t)(__noswap_vshll_n_s16(__a1_531, 0));
50372   __ret_531 = __builtin_shufflevector(__ret_531, __ret_531, 3, 2, 1, 0);
50373   return __ret_531;
50374 }
__noswap_vmovl_high_s16(int16x8_t __p0_532)50375 __ai int32x4_t __noswap_vmovl_high_s16(int16x8_t __p0_532) {
50376   int32x4_t __ret_532;
50377   int16x4_t __a1_532 = __noswap_vget_high_s16(__p0_532);
50378   __ret_532 = (int32x4_t)(__noswap_vshll_n_s16(__a1_532, 0));
50379   return __ret_532;
50380 }
50381 #endif
50382 
50383 #ifdef __LITTLE_ENDIAN__
vmovn_high_u32(uint16x4_t __p0,uint32x4_t __p1)50384 __ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
50385   uint16x8_t __ret;
50386   __ret = vcombine_u16(__p0, vmovn_u32(__p1));
50387   return __ret;
50388 }
50389 #else
vmovn_high_u32(uint16x4_t __p0,uint32x4_t __p1)50390 __ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
50391   uint16x8_t __ret;
50392   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50393   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
50394   __ret = __noswap_vcombine_u16(__rev0, __noswap_vmovn_u32(__rev1));
50395   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
50396   return __ret;
50397 }
50398 #endif
50399 
50400 #ifdef __LITTLE_ENDIAN__
vmovn_high_u64(uint32x2_t __p0,uint64x2_t __p1)50401 __ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
50402   uint32x4_t __ret;
50403   __ret = vcombine_u32(__p0, vmovn_u64(__p1));
50404   return __ret;
50405 }
50406 #else
vmovn_high_u64(uint32x2_t __p0,uint64x2_t __p1)50407 __ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
50408   uint32x4_t __ret;
50409   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50410   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50411   __ret = __noswap_vcombine_u32(__rev0, __noswap_vmovn_u64(__rev1));
50412   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
50413   return __ret;
50414 }
50415 #endif
50416 
50417 #ifdef __LITTLE_ENDIAN__
vmovn_high_u16(uint8x8_t __p0,uint16x8_t __p1)50418 __ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
50419   uint8x16_t __ret;
50420   __ret = vcombine_u8(__p0, vmovn_u16(__p1));
50421   return __ret;
50422 }
50423 #else
vmovn_high_u16(uint8x8_t __p0,uint16x8_t __p1)50424 __ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
50425   uint8x16_t __ret;
50426   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
50427   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
50428   __ret = __noswap_vcombine_u8(__rev0, __noswap_vmovn_u16(__rev1));
50429   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50430   return __ret;
50431 }
50432 #endif
50433 
50434 #ifdef __LITTLE_ENDIAN__
vmovn_high_s32(int16x4_t __p0,int32x4_t __p1)50435 __ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
50436   int16x8_t __ret;
50437   __ret = vcombine_s16(__p0, vmovn_s32(__p1));
50438   return __ret;
50439 }
50440 #else
vmovn_high_s32(int16x4_t __p0,int32x4_t __p1)50441 __ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
50442   int16x8_t __ret;
50443   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50444   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
50445   __ret = __noswap_vcombine_s16(__rev0, __noswap_vmovn_s32(__rev1));
50446   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
50447   return __ret;
50448 }
50449 #endif
50450 
50451 #ifdef __LITTLE_ENDIAN__
vmovn_high_s64(int32x2_t __p0,int64x2_t __p1)50452 __ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
50453   int32x4_t __ret;
50454   __ret = vcombine_s32(__p0, vmovn_s64(__p1));
50455   return __ret;
50456 }
50457 #else
vmovn_high_s64(int32x2_t __p0,int64x2_t __p1)50458 __ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
50459   int32x4_t __ret;
50460   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50461   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50462   __ret = __noswap_vcombine_s32(__rev0, __noswap_vmovn_s64(__rev1));
50463   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
50464   return __ret;
50465 }
50466 #endif
50467 
50468 #ifdef __LITTLE_ENDIAN__
vmovn_high_s16(int8x8_t __p0,int16x8_t __p1)50469 __ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
50470   int8x16_t __ret;
50471   __ret = vcombine_s8(__p0, vmovn_s16(__p1));
50472   return __ret;
50473 }
50474 #else
vmovn_high_s16(int8x8_t __p0,int16x8_t __p1)50475 __ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
50476   int8x16_t __ret;
50477   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
50478   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
50479   __ret = __noswap_vcombine_s8(__rev0, __noswap_vmovn_s16(__rev1));
50480   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50481   return __ret;
50482 }
50483 #endif
50484 
50485 #ifdef __LITTLE_ENDIAN__
vmulq_f64(float64x2_t __p0,float64x2_t __p1)50486 __ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
50487   float64x2_t __ret;
50488   __ret = __p0 * __p1;
50489   return __ret;
50490 }
50491 #else
vmulq_f64(float64x2_t __p0,float64x2_t __p1)50492 __ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
50493   float64x2_t __ret;
50494   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50495   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50496   __ret = __rev0 * __rev1;
50497   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50498   return __ret;
50499 }
50500 #endif
50501 
vmul_f64(float64x1_t __p0,float64x1_t __p1)50502 __ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
50503   float64x1_t __ret;
50504   __ret = __p0 * __p1;
50505   return __ret;
50506 }
50507 #define vmuld_lane_f64(__p0_533, __p1_533, __p2_533) __extension__ ({ \
50508   float64_t __ret_533; \
50509   float64_t __s0_533 = __p0_533; \
50510   float64x1_t __s1_533 = __p1_533; \
50511   __ret_533 = __s0_533 * vget_lane_f64(__s1_533, __p2_533); \
50512   __ret_533; \
50513 })
50514 #ifdef __LITTLE_ENDIAN__
50515 #define vmuls_lane_f32(__p0_534, __p1_534, __p2_534) __extension__ ({ \
50516   float32_t __ret_534; \
50517   float32_t __s0_534 = __p0_534; \
50518   float32x2_t __s1_534 = __p1_534; \
50519   __ret_534 = __s0_534 * vget_lane_f32(__s1_534, __p2_534); \
50520   __ret_534; \
50521 })
50522 #else
50523 #define vmuls_lane_f32(__p0_535, __p1_535, __p2_535) __extension__ ({ \
50524   float32_t __ret_535; \
50525   float32_t __s0_535 = __p0_535; \
50526   float32x2_t __s1_535 = __p1_535; \
50527   float32x2_t __rev1_535;  __rev1_535 = __builtin_shufflevector(__s1_535, __s1_535, 1, 0); \
50528   __ret_535 = __s0_535 * __noswap_vget_lane_f32(__rev1_535, __p2_535); \
50529   __ret_535; \
50530 })
50531 #endif
50532 
50533 #define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50534   float64x1_t __ret; \
50535   float64x1_t __s0 = __p0; \
50536   float64x1_t __s1 = __p1; \
50537   __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
50538   __ret; \
50539 })
50540 #ifdef __LITTLE_ENDIAN__
50541 #define vmulq_lane_f64(__p0_536, __p1_536, __p2_536) __extension__ ({ \
50542   float64x2_t __ret_536; \
50543   float64x2_t __s0_536 = __p0_536; \
50544   float64x1_t __s1_536 = __p1_536; \
50545   __ret_536 = __s0_536 * splatq_lane_f64(__s1_536, __p2_536); \
50546   __ret_536; \
50547 })
50548 #else
50549 #define vmulq_lane_f64(__p0_537, __p1_537, __p2_537) __extension__ ({ \
50550   float64x2_t __ret_537; \
50551   float64x2_t __s0_537 = __p0_537; \
50552   float64x1_t __s1_537 = __p1_537; \
50553   float64x2_t __rev0_537;  __rev0_537 = __builtin_shufflevector(__s0_537, __s0_537, 1, 0); \
50554   __ret_537 = __rev0_537 * __noswap_splatq_lane_f64(__s1_537, __p2_537); \
50555   __ret_537 = __builtin_shufflevector(__ret_537, __ret_537, 1, 0); \
50556   __ret_537; \
50557 })
50558 #endif
50559 
50560 #ifdef __LITTLE_ENDIAN__
50561 #define vmuld_laneq_f64(__p0_538, __p1_538, __p2_538) __extension__ ({ \
50562   float64_t __ret_538; \
50563   float64_t __s0_538 = __p0_538; \
50564   float64x2_t __s1_538 = __p1_538; \
50565   __ret_538 = __s0_538 * vgetq_lane_f64(__s1_538, __p2_538); \
50566   __ret_538; \
50567 })
50568 #else
50569 #define vmuld_laneq_f64(__p0_539, __p1_539, __p2_539) __extension__ ({ \
50570   float64_t __ret_539; \
50571   float64_t __s0_539 = __p0_539; \
50572   float64x2_t __s1_539 = __p1_539; \
50573   float64x2_t __rev1_539;  __rev1_539 = __builtin_shufflevector(__s1_539, __s1_539, 1, 0); \
50574   __ret_539 = __s0_539 * __noswap_vgetq_lane_f64(__rev1_539, __p2_539); \
50575   __ret_539; \
50576 })
50577 #endif
50578 
50579 #ifdef __LITTLE_ENDIAN__
50580 #define vmuls_laneq_f32(__p0_540, __p1_540, __p2_540) __extension__ ({ \
50581   float32_t __ret_540; \
50582   float32_t __s0_540 = __p0_540; \
50583   float32x4_t __s1_540 = __p1_540; \
50584   __ret_540 = __s0_540 * vgetq_lane_f32(__s1_540, __p2_540); \
50585   __ret_540; \
50586 })
50587 #else
50588 #define vmuls_laneq_f32(__p0_541, __p1_541, __p2_541) __extension__ ({ \
50589   float32_t __ret_541; \
50590   float32_t __s0_541 = __p0_541; \
50591   float32x4_t __s1_541 = __p1_541; \
50592   float32x4_t __rev1_541;  __rev1_541 = __builtin_shufflevector(__s1_541, __s1_541, 3, 2, 1, 0); \
50593   __ret_541 = __s0_541 * __noswap_vgetq_lane_f32(__rev1_541, __p2_541); \
50594   __ret_541; \
50595 })
50596 #endif
50597 
50598 #ifdef __LITTLE_ENDIAN__
50599 #define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
50600   float64x1_t __ret; \
50601   float64x1_t __s0 = __p0; \
50602   float64x2_t __s1 = __p1; \
50603   __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 10); \
50604   __ret; \
50605 })
50606 #else
50607 #define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
50608   float64x1_t __ret; \
50609   float64x1_t __s0 = __p0; \
50610   float64x2_t __s1 = __p1; \
50611   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
50612   __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__rev1, __p2, 10); \
50613   __ret; \
50614 })
50615 #endif
50616 
50617 #ifdef __LITTLE_ENDIAN__
50618 #define vmulq_laneq_u32(__p0_542, __p1_542, __p2_542) __extension__ ({ \
50619   uint32x4_t __ret_542; \
50620   uint32x4_t __s0_542 = __p0_542; \
50621   uint32x4_t __s1_542 = __p1_542; \
50622   __ret_542 = __s0_542 * splatq_laneq_u32(__s1_542, __p2_542); \
50623   __ret_542; \
50624 })
50625 #else
50626 #define vmulq_laneq_u32(__p0_543, __p1_543, __p2_543) __extension__ ({ \
50627   uint32x4_t __ret_543; \
50628   uint32x4_t __s0_543 = __p0_543; \
50629   uint32x4_t __s1_543 = __p1_543; \
50630   uint32x4_t __rev0_543;  __rev0_543 = __builtin_shufflevector(__s0_543, __s0_543, 3, 2, 1, 0); \
50631   uint32x4_t __rev1_543;  __rev1_543 = __builtin_shufflevector(__s1_543, __s1_543, 3, 2, 1, 0); \
50632   __ret_543 = __rev0_543 * __noswap_splatq_laneq_u32(__rev1_543, __p2_543); \
50633   __ret_543 = __builtin_shufflevector(__ret_543, __ret_543, 3, 2, 1, 0); \
50634   __ret_543; \
50635 })
50636 #endif
50637 
50638 #ifdef __LITTLE_ENDIAN__
50639 #define vmulq_laneq_u16(__p0_544, __p1_544, __p2_544) __extension__ ({ \
50640   uint16x8_t __ret_544; \
50641   uint16x8_t __s0_544 = __p0_544; \
50642   uint16x8_t __s1_544 = __p1_544; \
50643   __ret_544 = __s0_544 * splatq_laneq_u16(__s1_544, __p2_544); \
50644   __ret_544; \
50645 })
50646 #else
50647 #define vmulq_laneq_u16(__p0_545, __p1_545, __p2_545) __extension__ ({ \
50648   uint16x8_t __ret_545; \
50649   uint16x8_t __s0_545 = __p0_545; \
50650   uint16x8_t __s1_545 = __p1_545; \
50651   uint16x8_t __rev0_545;  __rev0_545 = __builtin_shufflevector(__s0_545, __s0_545, 7, 6, 5, 4, 3, 2, 1, 0); \
50652   uint16x8_t __rev1_545;  __rev1_545 = __builtin_shufflevector(__s1_545, __s1_545, 7, 6, 5, 4, 3, 2, 1, 0); \
50653   __ret_545 = __rev0_545 * __noswap_splatq_laneq_u16(__rev1_545, __p2_545); \
50654   __ret_545 = __builtin_shufflevector(__ret_545, __ret_545, 7, 6, 5, 4, 3, 2, 1, 0); \
50655   __ret_545; \
50656 })
50657 #endif
50658 
50659 #ifdef __LITTLE_ENDIAN__
50660 #define vmulq_laneq_f64(__p0_546, __p1_546, __p2_546) __extension__ ({ \
50661   float64x2_t __ret_546; \
50662   float64x2_t __s0_546 = __p0_546; \
50663   float64x2_t __s1_546 = __p1_546; \
50664   __ret_546 = __s0_546 * splatq_laneq_f64(__s1_546, __p2_546); \
50665   __ret_546; \
50666 })
50667 #else
50668 #define vmulq_laneq_f64(__p0_547, __p1_547, __p2_547) __extension__ ({ \
50669   float64x2_t __ret_547; \
50670   float64x2_t __s0_547 = __p0_547; \
50671   float64x2_t __s1_547 = __p1_547; \
50672   float64x2_t __rev0_547;  __rev0_547 = __builtin_shufflevector(__s0_547, __s0_547, 1, 0); \
50673   float64x2_t __rev1_547;  __rev1_547 = __builtin_shufflevector(__s1_547, __s1_547, 1, 0); \
50674   __ret_547 = __rev0_547 * __noswap_splatq_laneq_f64(__rev1_547, __p2_547); \
50675   __ret_547 = __builtin_shufflevector(__ret_547, __ret_547, 1, 0); \
50676   __ret_547; \
50677 })
50678 #endif
50679 
50680 #ifdef __LITTLE_ENDIAN__
50681 #define vmulq_laneq_f32(__p0_548, __p1_548, __p2_548) __extension__ ({ \
50682   float32x4_t __ret_548; \
50683   float32x4_t __s0_548 = __p0_548; \
50684   float32x4_t __s1_548 = __p1_548; \
50685   __ret_548 = __s0_548 * splatq_laneq_f32(__s1_548, __p2_548); \
50686   __ret_548; \
50687 })
50688 #else
50689 #define vmulq_laneq_f32(__p0_549, __p1_549, __p2_549) __extension__ ({ \
50690   float32x4_t __ret_549; \
50691   float32x4_t __s0_549 = __p0_549; \
50692   float32x4_t __s1_549 = __p1_549; \
50693   float32x4_t __rev0_549;  __rev0_549 = __builtin_shufflevector(__s0_549, __s0_549, 3, 2, 1, 0); \
50694   float32x4_t __rev1_549;  __rev1_549 = __builtin_shufflevector(__s1_549, __s1_549, 3, 2, 1, 0); \
50695   __ret_549 = __rev0_549 * __noswap_splatq_laneq_f32(__rev1_549, __p2_549); \
50696   __ret_549 = __builtin_shufflevector(__ret_549, __ret_549, 3, 2, 1, 0); \
50697   __ret_549; \
50698 })
50699 #endif
50700 
50701 #ifdef __LITTLE_ENDIAN__
50702 #define vmulq_laneq_s32(__p0_550, __p1_550, __p2_550) __extension__ ({ \
50703   int32x4_t __ret_550; \
50704   int32x4_t __s0_550 = __p0_550; \
50705   int32x4_t __s1_550 = __p1_550; \
50706   __ret_550 = __s0_550 * splatq_laneq_s32(__s1_550, __p2_550); \
50707   __ret_550; \
50708 })
50709 #else
50710 #define vmulq_laneq_s32(__p0_551, __p1_551, __p2_551) __extension__ ({ \
50711   int32x4_t __ret_551; \
50712   int32x4_t __s0_551 = __p0_551; \
50713   int32x4_t __s1_551 = __p1_551; \
50714   int32x4_t __rev0_551;  __rev0_551 = __builtin_shufflevector(__s0_551, __s0_551, 3, 2, 1, 0); \
50715   int32x4_t __rev1_551;  __rev1_551 = __builtin_shufflevector(__s1_551, __s1_551, 3, 2, 1, 0); \
50716   __ret_551 = __rev0_551 * __noswap_splatq_laneq_s32(__rev1_551, __p2_551); \
50717   __ret_551 = __builtin_shufflevector(__ret_551, __ret_551, 3, 2, 1, 0); \
50718   __ret_551; \
50719 })
50720 #endif
50721 
50722 #ifdef __LITTLE_ENDIAN__
50723 #define vmulq_laneq_s16(__p0_552, __p1_552, __p2_552) __extension__ ({ \
50724   int16x8_t __ret_552; \
50725   int16x8_t __s0_552 = __p0_552; \
50726   int16x8_t __s1_552 = __p1_552; \
50727   __ret_552 = __s0_552 * splatq_laneq_s16(__s1_552, __p2_552); \
50728   __ret_552; \
50729 })
50730 #else
50731 #define vmulq_laneq_s16(__p0_553, __p1_553, __p2_553) __extension__ ({ \
50732   int16x8_t __ret_553; \
50733   int16x8_t __s0_553 = __p0_553; \
50734   int16x8_t __s1_553 = __p1_553; \
50735   int16x8_t __rev0_553;  __rev0_553 = __builtin_shufflevector(__s0_553, __s0_553, 7, 6, 5, 4, 3, 2, 1, 0); \
50736   int16x8_t __rev1_553;  __rev1_553 = __builtin_shufflevector(__s1_553, __s1_553, 7, 6, 5, 4, 3, 2, 1, 0); \
50737   __ret_553 = __rev0_553 * __noswap_splatq_laneq_s16(__rev1_553, __p2_553); \
50738   __ret_553 = __builtin_shufflevector(__ret_553, __ret_553, 7, 6, 5, 4, 3, 2, 1, 0); \
50739   __ret_553; \
50740 })
50741 #endif
50742 
50743 #ifdef __LITTLE_ENDIAN__
50744 #define vmul_laneq_u32(__p0_554, __p1_554, __p2_554) __extension__ ({ \
50745   uint32x2_t __ret_554; \
50746   uint32x2_t __s0_554 = __p0_554; \
50747   uint32x4_t __s1_554 = __p1_554; \
50748   __ret_554 = __s0_554 * splat_laneq_u32(__s1_554, __p2_554); \
50749   __ret_554; \
50750 })
50751 #else
50752 #define vmul_laneq_u32(__p0_555, __p1_555, __p2_555) __extension__ ({ \
50753   uint32x2_t __ret_555; \
50754   uint32x2_t __s0_555 = __p0_555; \
50755   uint32x4_t __s1_555 = __p1_555; \
50756   uint32x2_t __rev0_555;  __rev0_555 = __builtin_shufflevector(__s0_555, __s0_555, 1, 0); \
50757   uint32x4_t __rev1_555;  __rev1_555 = __builtin_shufflevector(__s1_555, __s1_555, 3, 2, 1, 0); \
50758   __ret_555 = __rev0_555 * __noswap_splat_laneq_u32(__rev1_555, __p2_555); \
50759   __ret_555 = __builtin_shufflevector(__ret_555, __ret_555, 1, 0); \
50760   __ret_555; \
50761 })
50762 #endif
50763 
50764 #ifdef __LITTLE_ENDIAN__
50765 #define vmul_laneq_u16(__p0_556, __p1_556, __p2_556) __extension__ ({ \
50766   uint16x4_t __ret_556; \
50767   uint16x4_t __s0_556 = __p0_556; \
50768   uint16x8_t __s1_556 = __p1_556; \
50769   __ret_556 = __s0_556 * splat_laneq_u16(__s1_556, __p2_556); \
50770   __ret_556; \
50771 })
50772 #else
50773 #define vmul_laneq_u16(__p0_557, __p1_557, __p2_557) __extension__ ({ \
50774   uint16x4_t __ret_557; \
50775   uint16x4_t __s0_557 = __p0_557; \
50776   uint16x8_t __s1_557 = __p1_557; \
50777   uint16x4_t __rev0_557;  __rev0_557 = __builtin_shufflevector(__s0_557, __s0_557, 3, 2, 1, 0); \
50778   uint16x8_t __rev1_557;  __rev1_557 = __builtin_shufflevector(__s1_557, __s1_557, 7, 6, 5, 4, 3, 2, 1, 0); \
50779   __ret_557 = __rev0_557 * __noswap_splat_laneq_u16(__rev1_557, __p2_557); \
50780   __ret_557 = __builtin_shufflevector(__ret_557, __ret_557, 3, 2, 1, 0); \
50781   __ret_557; \
50782 })
50783 #endif
50784 
50785 #ifdef __LITTLE_ENDIAN__
50786 #define vmul_laneq_f32(__p0_558, __p1_558, __p2_558) __extension__ ({ \
50787   float32x2_t __ret_558; \
50788   float32x2_t __s0_558 = __p0_558; \
50789   float32x4_t __s1_558 = __p1_558; \
50790   __ret_558 = __s0_558 * splat_laneq_f32(__s1_558, __p2_558); \
50791   __ret_558; \
50792 })
50793 #else
50794 #define vmul_laneq_f32(__p0_559, __p1_559, __p2_559) __extension__ ({ \
50795   float32x2_t __ret_559; \
50796   float32x2_t __s0_559 = __p0_559; \
50797   float32x4_t __s1_559 = __p1_559; \
50798   float32x2_t __rev0_559;  __rev0_559 = __builtin_shufflevector(__s0_559, __s0_559, 1, 0); \
50799   float32x4_t __rev1_559;  __rev1_559 = __builtin_shufflevector(__s1_559, __s1_559, 3, 2, 1, 0); \
50800   __ret_559 = __rev0_559 * __noswap_splat_laneq_f32(__rev1_559, __p2_559); \
50801   __ret_559 = __builtin_shufflevector(__ret_559, __ret_559, 1, 0); \
50802   __ret_559; \
50803 })
50804 #endif
50805 
50806 #ifdef __LITTLE_ENDIAN__
50807 #define vmul_laneq_s32(__p0_560, __p1_560, __p2_560) __extension__ ({ \
50808   int32x2_t __ret_560; \
50809   int32x2_t __s0_560 = __p0_560; \
50810   int32x4_t __s1_560 = __p1_560; \
50811   __ret_560 = __s0_560 * splat_laneq_s32(__s1_560, __p2_560); \
50812   __ret_560; \
50813 })
50814 #else
50815 #define vmul_laneq_s32(__p0_561, __p1_561, __p2_561) __extension__ ({ \
50816   int32x2_t __ret_561; \
50817   int32x2_t __s0_561 = __p0_561; \
50818   int32x4_t __s1_561 = __p1_561; \
50819   int32x2_t __rev0_561;  __rev0_561 = __builtin_shufflevector(__s0_561, __s0_561, 1, 0); \
50820   int32x4_t __rev1_561;  __rev1_561 = __builtin_shufflevector(__s1_561, __s1_561, 3, 2, 1, 0); \
50821   __ret_561 = __rev0_561 * __noswap_splat_laneq_s32(__rev1_561, __p2_561); \
50822   __ret_561 = __builtin_shufflevector(__ret_561, __ret_561, 1, 0); \
50823   __ret_561; \
50824 })
50825 #endif
50826 
50827 #ifdef __LITTLE_ENDIAN__
50828 #define vmul_laneq_s16(__p0_562, __p1_562, __p2_562) __extension__ ({ \
50829   int16x4_t __ret_562; \
50830   int16x4_t __s0_562 = __p0_562; \
50831   int16x8_t __s1_562 = __p1_562; \
50832   __ret_562 = __s0_562 * splat_laneq_s16(__s1_562, __p2_562); \
50833   __ret_562; \
50834 })
50835 #else
50836 #define vmul_laneq_s16(__p0_563, __p1_563, __p2_563) __extension__ ({ \
50837   int16x4_t __ret_563; \
50838   int16x4_t __s0_563 = __p0_563; \
50839   int16x8_t __s1_563 = __p1_563; \
50840   int16x4_t __rev0_563;  __rev0_563 = __builtin_shufflevector(__s0_563, __s0_563, 3, 2, 1, 0); \
50841   int16x8_t __rev1_563;  __rev1_563 = __builtin_shufflevector(__s1_563, __s1_563, 7, 6, 5, 4, 3, 2, 1, 0); \
50842   __ret_563 = __rev0_563 * __noswap_splat_laneq_s16(__rev1_563, __p2_563); \
50843   __ret_563 = __builtin_shufflevector(__ret_563, __ret_563, 3, 2, 1, 0); \
50844   __ret_563; \
50845 })
50846 #endif
50847 
vmul_n_f64(float64x1_t __p0,float64_t __p1)50848 __ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
50849   float64x1_t __ret;
50850   __ret = (float64x1_t) __builtin_neon_vmul_n_f64((float64x1_t)__p0, __p1);
50851   return __ret;
50852 }
50853 #ifdef __LITTLE_ENDIAN__
vmulq_n_f64(float64x2_t __p0,float64_t __p1)50854 __ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
50855   float64x2_t __ret;
50856   __ret = __p0 * (float64x2_t) {__p1, __p1};
50857   return __ret;
50858 }
50859 #else
vmulq_n_f64(float64x2_t __p0,float64_t __p1)50860 __ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
50861   float64x2_t __ret;
50862   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50863   __ret = __rev0 * (float64x2_t) {__p1, __p1};
50864   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50865   return __ret;
50866 }
50867 #endif
50868 
50869 #ifdef __LITTLE_ENDIAN__
vmull_high_p8(poly8x16_t __p0,poly8x16_t __p1)50870 __ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
50871   poly16x8_t __ret;
50872   __ret = vmull_p8(vget_high_p8(__p0), vget_high_p8(__p1));
50873   return __ret;
50874 }
50875 #else
vmull_high_p8(poly8x16_t __p0,poly8x16_t __p1)50876 __ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
50877   poly16x8_t __ret;
50878   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50879   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50880   __ret = __noswap_vmull_p8(__noswap_vget_high_p8(__rev0), __noswap_vget_high_p8(__rev1));
50881   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
50882   return __ret;
50883 }
50884 #endif
50885 
50886 #ifdef __LITTLE_ENDIAN__
vmull_high_u8(uint8x16_t __p0,uint8x16_t __p1)50887 __ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
50888   uint16x8_t __ret;
50889   __ret = vmull_u8(vget_high_u8(__p0), vget_high_u8(__p1));
50890   return __ret;
50891 }
50892 #else
vmull_high_u8(uint8x16_t __p0,uint8x16_t __p1)50893 __ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
50894   uint16x8_t __ret;
50895   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50896   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50897   __ret = __noswap_vmull_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
50898   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
50899   return __ret;
50900 }
50901 #endif
50902 
50903 #ifdef __LITTLE_ENDIAN__
vmull_high_u32(uint32x4_t __p0,uint32x4_t __p1)50904 __ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
50905   uint64x2_t __ret;
50906   __ret = vmull_u32(vget_high_u32(__p0), vget_high_u32(__p1));
50907   return __ret;
50908 }
50909 #else
vmull_high_u32(uint32x4_t __p0,uint32x4_t __p1)50910 __ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
50911   uint64x2_t __ret;
50912   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50913   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
50914   __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
50915   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50916   return __ret;
50917 }
50918 #endif
50919 
50920 #ifdef __LITTLE_ENDIAN__
vmull_high_u16(uint16x8_t __p0,uint16x8_t __p1)50921 __ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
50922   uint32x4_t __ret;
50923   __ret = vmull_u16(vget_high_u16(__p0), vget_high_u16(__p1));
50924   return __ret;
50925 }
50926 #else
vmull_high_u16(uint16x8_t __p0,uint16x8_t __p1)50927 __ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
50928   uint32x4_t __ret;
50929   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
50930   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
50931   __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
50932   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
50933   return __ret;
50934 }
50935 #endif
50936 
50937 #ifdef __LITTLE_ENDIAN__
vmull_high_s8(int8x16_t __p0,int8x16_t __p1)50938 __ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
50939   int16x8_t __ret;
50940   __ret = vmull_s8(vget_high_s8(__p0), vget_high_s8(__p1));
50941   return __ret;
50942 }
50943 #else
vmull_high_s8(int8x16_t __p0,int8x16_t __p1)50944 __ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
50945   int16x8_t __ret;
50946   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50947   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50948   __ret = __noswap_vmull_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
50949   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
50950   return __ret;
50951 }
50952 #endif
50953 
50954 #ifdef __LITTLE_ENDIAN__
vmull_high_s32(int32x4_t __p0,int32x4_t __p1)50955 __ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
50956   int64x2_t __ret;
50957   __ret = vmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
50958   return __ret;
50959 }
50960 #else
vmull_high_s32(int32x4_t __p0,int32x4_t __p1)50961 __ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
50962   int64x2_t __ret;
50963   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50964   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
50965   __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
50966   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50967   return __ret;
50968 }
50969 #endif
50970 
50971 #ifdef __LITTLE_ENDIAN__
vmull_high_s16(int16x8_t __p0,int16x8_t __p1)50972 __ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
50973   int32x4_t __ret;
50974   __ret = vmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
50975   return __ret;
50976 }
50977 #else
vmull_high_s16(int16x8_t __p0,int16x8_t __p1)50978 __ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
50979   int32x4_t __ret;
50980   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
50981   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
50982   __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
50983   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
50984   return __ret;
50985 }
50986 #endif
50987 
50988 #ifdef __LITTLE_ENDIAN__
50989 #define vmull_high_lane_u32(__p0_564, __p1_564, __p2_564) __extension__ ({ \
50990   uint64x2_t __ret_564; \
50991   uint32x4_t __s0_564 = __p0_564; \
50992   uint32x2_t __s1_564 = __p1_564; \
50993   __ret_564 = vmull_u32(vget_high_u32(__s0_564), splat_lane_u32(__s1_564, __p2_564)); \
50994   __ret_564; \
50995 })
50996 #else
50997 #define vmull_high_lane_u32(__p0_565, __p1_565, __p2_565) __extension__ ({ \
50998   uint64x2_t __ret_565; \
50999   uint32x4_t __s0_565 = __p0_565; \
51000   uint32x2_t __s1_565 = __p1_565; \
51001   uint32x4_t __rev0_565;  __rev0_565 = __builtin_shufflevector(__s0_565, __s0_565, 3, 2, 1, 0); \
51002   uint32x2_t __rev1_565;  __rev1_565 = __builtin_shufflevector(__s1_565, __s1_565, 1, 0); \
51003   __ret_565 = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0_565), __noswap_splat_lane_u32(__rev1_565, __p2_565)); \
51004   __ret_565 = __builtin_shufflevector(__ret_565, __ret_565, 1, 0); \
51005   __ret_565; \
51006 })
51007 #endif
51008 
51009 #ifdef __LITTLE_ENDIAN__
51010 #define vmull_high_lane_u16(__p0_566, __p1_566, __p2_566) __extension__ ({ \
51011   uint32x4_t __ret_566; \
51012   uint16x8_t __s0_566 = __p0_566; \
51013   uint16x4_t __s1_566 = __p1_566; \
51014   __ret_566 = vmull_u16(vget_high_u16(__s0_566), splat_lane_u16(__s1_566, __p2_566)); \
51015   __ret_566; \
51016 })
51017 #else
51018 #define vmull_high_lane_u16(__p0_567, __p1_567, __p2_567) __extension__ ({ \
51019   uint32x4_t __ret_567; \
51020   uint16x8_t __s0_567 = __p0_567; \
51021   uint16x4_t __s1_567 = __p1_567; \
51022   uint16x8_t __rev0_567;  __rev0_567 = __builtin_shufflevector(__s0_567, __s0_567, 7, 6, 5, 4, 3, 2, 1, 0); \
51023   uint16x4_t __rev1_567;  __rev1_567 = __builtin_shufflevector(__s1_567, __s1_567, 3, 2, 1, 0); \
51024   __ret_567 = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0_567), __noswap_splat_lane_u16(__rev1_567, __p2_567)); \
51025   __ret_567 = __builtin_shufflevector(__ret_567, __ret_567, 3, 2, 1, 0); \
51026   __ret_567; \
51027 })
51028 #endif
51029 
51030 #ifdef __LITTLE_ENDIAN__
51031 #define vmull_high_lane_s32(__p0_568, __p1_568, __p2_568) __extension__ ({ \
51032   int64x2_t __ret_568; \
51033   int32x4_t __s0_568 = __p0_568; \
51034   int32x2_t __s1_568 = __p1_568; \
51035   __ret_568 = vmull_s32(vget_high_s32(__s0_568), splat_lane_s32(__s1_568, __p2_568)); \
51036   __ret_568; \
51037 })
51038 #else
51039 #define vmull_high_lane_s32(__p0_569, __p1_569, __p2_569) __extension__ ({ \
51040   int64x2_t __ret_569; \
51041   int32x4_t __s0_569 = __p0_569; \
51042   int32x2_t __s1_569 = __p1_569; \
51043   int32x4_t __rev0_569;  __rev0_569 = __builtin_shufflevector(__s0_569, __s0_569, 3, 2, 1, 0); \
51044   int32x2_t __rev1_569;  __rev1_569 = __builtin_shufflevector(__s1_569, __s1_569, 1, 0); \
51045   __ret_569 = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0_569), __noswap_splat_lane_s32(__rev1_569, __p2_569)); \
51046   __ret_569 = __builtin_shufflevector(__ret_569, __ret_569, 1, 0); \
51047   __ret_569; \
51048 })
51049 #endif
51050 
51051 #ifdef __LITTLE_ENDIAN__
51052 #define vmull_high_lane_s16(__p0_570, __p1_570, __p2_570) __extension__ ({ \
51053   int32x4_t __ret_570; \
51054   int16x8_t __s0_570 = __p0_570; \
51055   int16x4_t __s1_570 = __p1_570; \
51056   __ret_570 = vmull_s16(vget_high_s16(__s0_570), splat_lane_s16(__s1_570, __p2_570)); \
51057   __ret_570; \
51058 })
51059 #else
51060 #define vmull_high_lane_s16(__p0_571, __p1_571, __p2_571) __extension__ ({ \
51061   int32x4_t __ret_571; \
51062   int16x8_t __s0_571 = __p0_571; \
51063   int16x4_t __s1_571 = __p1_571; \
51064   int16x8_t __rev0_571;  __rev0_571 = __builtin_shufflevector(__s0_571, __s0_571, 7, 6, 5, 4, 3, 2, 1, 0); \
51065   int16x4_t __rev1_571;  __rev1_571 = __builtin_shufflevector(__s1_571, __s1_571, 3, 2, 1, 0); \
51066   __ret_571 = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0_571), __noswap_splat_lane_s16(__rev1_571, __p2_571)); \
51067   __ret_571 = __builtin_shufflevector(__ret_571, __ret_571, 3, 2, 1, 0); \
51068   __ret_571; \
51069 })
51070 #endif
51071 
51072 #ifdef __LITTLE_ENDIAN__
51073 #define vmull_high_laneq_u32(__p0_572, __p1_572, __p2_572) __extension__ ({ \
51074   uint64x2_t __ret_572; \
51075   uint32x4_t __s0_572 = __p0_572; \
51076   uint32x4_t __s1_572 = __p1_572; \
51077   __ret_572 = vmull_u32(vget_high_u32(__s0_572), splat_laneq_u32(__s1_572, __p2_572)); \
51078   __ret_572; \
51079 })
51080 #else
51081 #define vmull_high_laneq_u32(__p0_573, __p1_573, __p2_573) __extension__ ({ \
51082   uint64x2_t __ret_573; \
51083   uint32x4_t __s0_573 = __p0_573; \
51084   uint32x4_t __s1_573 = __p1_573; \
51085   uint32x4_t __rev0_573;  __rev0_573 = __builtin_shufflevector(__s0_573, __s0_573, 3, 2, 1, 0); \
51086   uint32x4_t __rev1_573;  __rev1_573 = __builtin_shufflevector(__s1_573, __s1_573, 3, 2, 1, 0); \
51087   __ret_573 = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0_573), __noswap_splat_laneq_u32(__rev1_573, __p2_573)); \
51088   __ret_573 = __builtin_shufflevector(__ret_573, __ret_573, 1, 0); \
51089   __ret_573; \
51090 })
51091 #endif
51092 
51093 #ifdef __LITTLE_ENDIAN__
51094 #define vmull_high_laneq_u16(__p0_574, __p1_574, __p2_574) __extension__ ({ \
51095   uint32x4_t __ret_574; \
51096   uint16x8_t __s0_574 = __p0_574; \
51097   uint16x8_t __s1_574 = __p1_574; \
51098   __ret_574 = vmull_u16(vget_high_u16(__s0_574), splat_laneq_u16(__s1_574, __p2_574)); \
51099   __ret_574; \
51100 })
51101 #else
51102 #define vmull_high_laneq_u16(__p0_575, __p1_575, __p2_575) __extension__ ({ \
51103   uint32x4_t __ret_575; \
51104   uint16x8_t __s0_575 = __p0_575; \
51105   uint16x8_t __s1_575 = __p1_575; \
51106   uint16x8_t __rev0_575;  __rev0_575 = __builtin_shufflevector(__s0_575, __s0_575, 7, 6, 5, 4, 3, 2, 1, 0); \
51107   uint16x8_t __rev1_575;  __rev1_575 = __builtin_shufflevector(__s1_575, __s1_575, 7, 6, 5, 4, 3, 2, 1, 0); \
51108   __ret_575 = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0_575), __noswap_splat_laneq_u16(__rev1_575, __p2_575)); \
51109   __ret_575 = __builtin_shufflevector(__ret_575, __ret_575, 3, 2, 1, 0); \
51110   __ret_575; \
51111 })
51112 #endif
51113 
51114 #ifdef __LITTLE_ENDIAN__
51115 #define vmull_high_laneq_s32(__p0_576, __p1_576, __p2_576) __extension__ ({ \
51116   int64x2_t __ret_576; \
51117   int32x4_t __s0_576 = __p0_576; \
51118   int32x4_t __s1_576 = __p1_576; \
51119   __ret_576 = vmull_s32(vget_high_s32(__s0_576), splat_laneq_s32(__s1_576, __p2_576)); \
51120   __ret_576; \
51121 })
51122 #else
51123 #define vmull_high_laneq_s32(__p0_577, __p1_577, __p2_577) __extension__ ({ \
51124   int64x2_t __ret_577; \
51125   int32x4_t __s0_577 = __p0_577; \
51126   int32x4_t __s1_577 = __p1_577; \
51127   int32x4_t __rev0_577;  __rev0_577 = __builtin_shufflevector(__s0_577, __s0_577, 3, 2, 1, 0); \
51128   int32x4_t __rev1_577;  __rev1_577 = __builtin_shufflevector(__s1_577, __s1_577, 3, 2, 1, 0); \
51129   __ret_577 = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0_577), __noswap_splat_laneq_s32(__rev1_577, __p2_577)); \
51130   __ret_577 = __builtin_shufflevector(__ret_577, __ret_577, 1, 0); \
51131   __ret_577; \
51132 })
51133 #endif
51134 
51135 #ifdef __LITTLE_ENDIAN__
51136 #define vmull_high_laneq_s16(__p0_578, __p1_578, __p2_578) __extension__ ({ \
51137   int32x4_t __ret_578; \
51138   int16x8_t __s0_578 = __p0_578; \
51139   int16x8_t __s1_578 = __p1_578; \
51140   __ret_578 = vmull_s16(vget_high_s16(__s0_578), splat_laneq_s16(__s1_578, __p2_578)); \
51141   __ret_578; \
51142 })
51143 #else
51144 #define vmull_high_laneq_s16(__p0_579, __p1_579, __p2_579) __extension__ ({ \
51145   int32x4_t __ret_579; \
51146   int16x8_t __s0_579 = __p0_579; \
51147   int16x8_t __s1_579 = __p1_579; \
51148   int16x8_t __rev0_579;  __rev0_579 = __builtin_shufflevector(__s0_579, __s0_579, 7, 6, 5, 4, 3, 2, 1, 0); \
51149   int16x8_t __rev1_579;  __rev1_579 = __builtin_shufflevector(__s1_579, __s1_579, 7, 6, 5, 4, 3, 2, 1, 0); \
51150   __ret_579 = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0_579), __noswap_splat_laneq_s16(__rev1_579, __p2_579)); \
51151   __ret_579 = __builtin_shufflevector(__ret_579, __ret_579, 3, 2, 1, 0); \
51152   __ret_579; \
51153 })
51154 #endif
51155 
51156 #ifdef __LITTLE_ENDIAN__
vmull_high_n_u32(uint32x4_t __p0,uint32_t __p1)51157 __ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
51158   uint64x2_t __ret;
51159   __ret = vmull_n_u32(vget_high_u32(__p0), __p1);
51160   return __ret;
51161 }
51162 #else
vmull_high_n_u32(uint32x4_t __p0,uint32_t __p1)51163 __ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
51164   uint64x2_t __ret;
51165   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51166   __ret = __noswap_vmull_n_u32(__noswap_vget_high_u32(__rev0), __p1);
51167   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51168   return __ret;
51169 }
51170 #endif
51171 
51172 #ifdef __LITTLE_ENDIAN__
vmull_high_n_u16(uint16x8_t __p0,uint16_t __p1)51173 __ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
51174   uint32x4_t __ret;
51175   __ret = vmull_n_u16(vget_high_u16(__p0), __p1);
51176   return __ret;
51177 }
51178 #else
vmull_high_n_u16(uint16x8_t __p0,uint16_t __p1)51179 __ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
51180   uint32x4_t __ret;
51181   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51182   __ret = __noswap_vmull_n_u16(__noswap_vget_high_u16(__rev0), __p1);
51183   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51184   return __ret;
51185 }
51186 #endif
51187 
51188 #ifdef __LITTLE_ENDIAN__
vmull_high_n_s32(int32x4_t __p0,int32_t __p1)51189 __ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
51190   int64x2_t __ret;
51191   __ret = vmull_n_s32(vget_high_s32(__p0), __p1);
51192   return __ret;
51193 }
51194 #else
vmull_high_n_s32(int32x4_t __p0,int32_t __p1)51195 __ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
51196   int64x2_t __ret;
51197   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51198   __ret = __noswap_vmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
51199   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51200   return __ret;
51201 }
51202 #endif
51203 
51204 #ifdef __LITTLE_ENDIAN__
vmull_high_n_s16(int16x8_t __p0,int16_t __p1)51205 __ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
51206   int32x4_t __ret;
51207   __ret = vmull_n_s16(vget_high_s16(__p0), __p1);
51208   return __ret;
51209 }
51210 #else
vmull_high_n_s16(int16x8_t __p0,int16_t __p1)51211 __ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
51212   int32x4_t __ret;
51213   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51214   __ret = __noswap_vmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
51215   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51216   return __ret;
51217 }
51218 #endif
51219 
51220 #ifdef __LITTLE_ENDIAN__
51221 #define vmull_laneq_u32(__p0_580, __p1_580, __p2_580) __extension__ ({ \
51222   uint64x2_t __ret_580; \
51223   uint32x2_t __s0_580 = __p0_580; \
51224   uint32x4_t __s1_580 = __p1_580; \
51225   __ret_580 = vmull_u32(__s0_580, splat_laneq_u32(__s1_580, __p2_580)); \
51226   __ret_580; \
51227 })
51228 #else
51229 #define vmull_laneq_u32(__p0_581, __p1_581, __p2_581) __extension__ ({ \
51230   uint64x2_t __ret_581; \
51231   uint32x2_t __s0_581 = __p0_581; \
51232   uint32x4_t __s1_581 = __p1_581; \
51233   uint32x2_t __rev0_581;  __rev0_581 = __builtin_shufflevector(__s0_581, __s0_581, 1, 0); \
51234   uint32x4_t __rev1_581;  __rev1_581 = __builtin_shufflevector(__s1_581, __s1_581, 3, 2, 1, 0); \
51235   __ret_581 = __noswap_vmull_u32(__rev0_581, __noswap_splat_laneq_u32(__rev1_581, __p2_581)); \
51236   __ret_581 = __builtin_shufflevector(__ret_581, __ret_581, 1, 0); \
51237   __ret_581; \
51238 })
51239 #endif
51240 
51241 #ifdef __LITTLE_ENDIAN__
51242 #define vmull_laneq_u16(__p0_582, __p1_582, __p2_582) __extension__ ({ \
51243   uint32x4_t __ret_582; \
51244   uint16x4_t __s0_582 = __p0_582; \
51245   uint16x8_t __s1_582 = __p1_582; \
51246   __ret_582 = vmull_u16(__s0_582, splat_laneq_u16(__s1_582, __p2_582)); \
51247   __ret_582; \
51248 })
51249 #else
51250 #define vmull_laneq_u16(__p0_583, __p1_583, __p2_583) __extension__ ({ \
51251   uint32x4_t __ret_583; \
51252   uint16x4_t __s0_583 = __p0_583; \
51253   uint16x8_t __s1_583 = __p1_583; \
51254   uint16x4_t __rev0_583;  __rev0_583 = __builtin_shufflevector(__s0_583, __s0_583, 3, 2, 1, 0); \
51255   uint16x8_t __rev1_583;  __rev1_583 = __builtin_shufflevector(__s1_583, __s1_583, 7, 6, 5, 4, 3, 2, 1, 0); \
51256   __ret_583 = __noswap_vmull_u16(__rev0_583, __noswap_splat_laneq_u16(__rev1_583, __p2_583)); \
51257   __ret_583 = __builtin_shufflevector(__ret_583, __ret_583, 3, 2, 1, 0); \
51258   __ret_583; \
51259 })
51260 #endif
51261 
51262 #ifdef __LITTLE_ENDIAN__
51263 #define vmull_laneq_s32(__p0_584, __p1_584, __p2_584) __extension__ ({ \
51264   int64x2_t __ret_584; \
51265   int32x2_t __s0_584 = __p0_584; \
51266   int32x4_t __s1_584 = __p1_584; \
51267   __ret_584 = vmull_s32(__s0_584, splat_laneq_s32(__s1_584, __p2_584)); \
51268   __ret_584; \
51269 })
51270 #else
51271 #define vmull_laneq_s32(__p0_585, __p1_585, __p2_585) __extension__ ({ \
51272   int64x2_t __ret_585; \
51273   int32x2_t __s0_585 = __p0_585; \
51274   int32x4_t __s1_585 = __p1_585; \
51275   int32x2_t __rev0_585;  __rev0_585 = __builtin_shufflevector(__s0_585, __s0_585, 1, 0); \
51276   int32x4_t __rev1_585;  __rev1_585 = __builtin_shufflevector(__s1_585, __s1_585, 3, 2, 1, 0); \
51277   __ret_585 = __noswap_vmull_s32(__rev0_585, __noswap_splat_laneq_s32(__rev1_585, __p2_585)); \
51278   __ret_585 = __builtin_shufflevector(__ret_585, __ret_585, 1, 0); \
51279   __ret_585; \
51280 })
51281 #endif
51282 
51283 #ifdef __LITTLE_ENDIAN__
51284 #define vmull_laneq_s16(__p0_586, __p1_586, __p2_586) __extension__ ({ \
51285   int32x4_t __ret_586; \
51286   int16x4_t __s0_586 = __p0_586; \
51287   int16x8_t __s1_586 = __p1_586; \
51288   __ret_586 = vmull_s16(__s0_586, splat_laneq_s16(__s1_586, __p2_586)); \
51289   __ret_586; \
51290 })
51291 #else
51292 #define vmull_laneq_s16(__p0_587, __p1_587, __p2_587) __extension__ ({ \
51293   int32x4_t __ret_587; \
51294   int16x4_t __s0_587 = __p0_587; \
51295   int16x8_t __s1_587 = __p1_587; \
51296   int16x4_t __rev0_587;  __rev0_587 = __builtin_shufflevector(__s0_587, __s0_587, 3, 2, 1, 0); \
51297   int16x8_t __rev1_587;  __rev1_587 = __builtin_shufflevector(__s1_587, __s1_587, 7, 6, 5, 4, 3, 2, 1, 0); \
51298   __ret_587 = __noswap_vmull_s16(__rev0_587, __noswap_splat_laneq_s16(__rev1_587, __p2_587)); \
51299   __ret_587 = __builtin_shufflevector(__ret_587, __ret_587, 3, 2, 1, 0); \
51300   __ret_587; \
51301 })
51302 #endif
51303 
51304 #ifdef __LITTLE_ENDIAN__
vmulxq_f64(float64x2_t __p0,float64x2_t __p1)51305 __ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
51306   float64x2_t __ret;
51307   __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
51308   return __ret;
51309 }
51310 #else
vmulxq_f64(float64x2_t __p0,float64x2_t __p1)51311 __ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
51312   float64x2_t __ret;
51313   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51314   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51315   __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
51316   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51317   return __ret;
51318 }
__noswap_vmulxq_f64(float64x2_t __p0,float64x2_t __p1)51319 __ai float64x2_t __noswap_vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
51320   float64x2_t __ret;
51321   __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
51322   return __ret;
51323 }
51324 #endif
51325 
51326 #ifdef __LITTLE_ENDIAN__
vmulxq_f32(float32x4_t __p0,float32x4_t __p1)51327 __ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
51328   float32x4_t __ret;
51329   __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
51330   return __ret;
51331 }
51332 #else
vmulxq_f32(float32x4_t __p0,float32x4_t __p1)51333 __ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
51334   float32x4_t __ret;
51335   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51336   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
51337   __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
51338   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51339   return __ret;
51340 }
__noswap_vmulxq_f32(float32x4_t __p0,float32x4_t __p1)51341 __ai float32x4_t __noswap_vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
51342   float32x4_t __ret;
51343   __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
51344   return __ret;
51345 }
51346 #endif
51347 
vmulx_f64(float64x1_t __p0,float64x1_t __p1)51348 __ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
51349   float64x1_t __ret;
51350   __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51351   return __ret;
51352 }
51353 #ifdef __LITTLE_ENDIAN__
vmulx_f32(float32x2_t __p0,float32x2_t __p1)51354 __ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
51355   float32x2_t __ret;
51356   __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
51357   return __ret;
51358 }
51359 #else
vmulx_f32(float32x2_t __p0,float32x2_t __p1)51360 __ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
51361   float32x2_t __ret;
51362   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51363   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51364   __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
51365   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51366   return __ret;
51367 }
__noswap_vmulx_f32(float32x2_t __p0,float32x2_t __p1)51368 __ai float32x2_t __noswap_vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
51369   float32x2_t __ret;
51370   __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
51371   return __ret;
51372 }
51373 #endif
51374 
vmulxd_f64(float64_t __p0,float64_t __p1)51375 __ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
51376   float64_t __ret;
51377   __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
51378   return __ret;
51379 }
vmulxs_f32(float32_t __p0,float32_t __p1)51380 __ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
51381   float32_t __ret;
51382   __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
51383   return __ret;
51384 }
51385 #define vmulxd_lane_f64(__p0_588, __p1_588, __p2_588) __extension__ ({ \
51386   float64_t __ret_588; \
51387   float64_t __s0_588 = __p0_588; \
51388   float64x1_t __s1_588 = __p1_588; \
51389   __ret_588 = vmulxd_f64(__s0_588, vget_lane_f64(__s1_588, __p2_588)); \
51390   __ret_588; \
51391 })
51392 #ifdef __LITTLE_ENDIAN__
51393 #define vmulxs_lane_f32(__p0_589, __p1_589, __p2_589) __extension__ ({ \
51394   float32_t __ret_589; \
51395   float32_t __s0_589 = __p0_589; \
51396   float32x2_t __s1_589 = __p1_589; \
51397   __ret_589 = vmulxs_f32(__s0_589, vget_lane_f32(__s1_589, __p2_589)); \
51398   __ret_589; \
51399 })
51400 #else
51401 #define vmulxs_lane_f32(__p0_590, __p1_590, __p2_590) __extension__ ({ \
51402   float32_t __ret_590; \
51403   float32_t __s0_590 = __p0_590; \
51404   float32x2_t __s1_590 = __p1_590; \
51405   float32x2_t __rev1_590;  __rev1_590 = __builtin_shufflevector(__s1_590, __s1_590, 1, 0); \
51406   __ret_590 = vmulxs_f32(__s0_590, __noswap_vget_lane_f32(__rev1_590, __p2_590)); \
51407   __ret_590; \
51408 })
51409 #endif
51410 
51411 #ifdef __LITTLE_ENDIAN__
51412 #define vmulxq_lane_f64(__p0_591, __p1_591, __p2_591) __extension__ ({ \
51413   float64x2_t __ret_591; \
51414   float64x2_t __s0_591 = __p0_591; \
51415   float64x1_t __s1_591 = __p1_591; \
51416   __ret_591 = vmulxq_f64(__s0_591, splatq_lane_f64(__s1_591, __p2_591)); \
51417   __ret_591; \
51418 })
51419 #else
51420 #define vmulxq_lane_f64(__p0_592, __p1_592, __p2_592) __extension__ ({ \
51421   float64x2_t __ret_592; \
51422   float64x2_t __s0_592 = __p0_592; \
51423   float64x1_t __s1_592 = __p1_592; \
51424   float64x2_t __rev0_592;  __rev0_592 = __builtin_shufflevector(__s0_592, __s0_592, 1, 0); \
51425   __ret_592 = __noswap_vmulxq_f64(__rev0_592, __noswap_splatq_lane_f64(__s1_592, __p2_592)); \
51426   __ret_592 = __builtin_shufflevector(__ret_592, __ret_592, 1, 0); \
51427   __ret_592; \
51428 })
51429 #endif
51430 
51431 #ifdef __LITTLE_ENDIAN__
51432 #define vmulxq_lane_f32(__p0_593, __p1_593, __p2_593) __extension__ ({ \
51433   float32x4_t __ret_593; \
51434   float32x4_t __s0_593 = __p0_593; \
51435   float32x2_t __s1_593 = __p1_593; \
51436   __ret_593 = vmulxq_f32(__s0_593, splatq_lane_f32(__s1_593, __p2_593)); \
51437   __ret_593; \
51438 })
51439 #else
51440 #define vmulxq_lane_f32(__p0_594, __p1_594, __p2_594) __extension__ ({ \
51441   float32x4_t __ret_594; \
51442   float32x4_t __s0_594 = __p0_594; \
51443   float32x2_t __s1_594 = __p1_594; \
51444   float32x4_t __rev0_594;  __rev0_594 = __builtin_shufflevector(__s0_594, __s0_594, 3, 2, 1, 0); \
51445   float32x2_t __rev1_594;  __rev1_594 = __builtin_shufflevector(__s1_594, __s1_594, 1, 0); \
51446   __ret_594 = __noswap_vmulxq_f32(__rev0_594, __noswap_splatq_lane_f32(__rev1_594, __p2_594)); \
51447   __ret_594 = __builtin_shufflevector(__ret_594, __ret_594, 3, 2, 1, 0); \
51448   __ret_594; \
51449 })
51450 #endif
51451 
51452 #ifdef __LITTLE_ENDIAN__
51453 #define vmulx_lane_f32(__p0_595, __p1_595, __p2_595) __extension__ ({ \
51454   float32x2_t __ret_595; \
51455   float32x2_t __s0_595 = __p0_595; \
51456   float32x2_t __s1_595 = __p1_595; \
51457   __ret_595 = vmulx_f32(__s0_595, splat_lane_f32(__s1_595, __p2_595)); \
51458   __ret_595; \
51459 })
51460 #else
51461 #define vmulx_lane_f32(__p0_596, __p1_596, __p2_596) __extension__ ({ \
51462   float32x2_t __ret_596; \
51463   float32x2_t __s0_596 = __p0_596; \
51464   float32x2_t __s1_596 = __p1_596; \
51465   float32x2_t __rev0_596;  __rev0_596 = __builtin_shufflevector(__s0_596, __s0_596, 1, 0); \
51466   float32x2_t __rev1_596;  __rev1_596 = __builtin_shufflevector(__s1_596, __s1_596, 1, 0); \
51467   __ret_596 = __noswap_vmulx_f32(__rev0_596, __noswap_splat_lane_f32(__rev1_596, __p2_596)); \
51468   __ret_596 = __builtin_shufflevector(__ret_596, __ret_596, 1, 0); \
51469   __ret_596; \
51470 })
51471 #endif
51472 
51473 #ifdef __LITTLE_ENDIAN__
51474 #define vmulxd_laneq_f64(__p0_597, __p1_597, __p2_597) __extension__ ({ \
51475   float64_t __ret_597; \
51476   float64_t __s0_597 = __p0_597; \
51477   float64x2_t __s1_597 = __p1_597; \
51478   __ret_597 = vmulxd_f64(__s0_597, vgetq_lane_f64(__s1_597, __p2_597)); \
51479   __ret_597; \
51480 })
51481 #else
51482 #define vmulxd_laneq_f64(__p0_598, __p1_598, __p2_598) __extension__ ({ \
51483   float64_t __ret_598; \
51484   float64_t __s0_598 = __p0_598; \
51485   float64x2_t __s1_598 = __p1_598; \
51486   float64x2_t __rev1_598;  __rev1_598 = __builtin_shufflevector(__s1_598, __s1_598, 1, 0); \
51487   __ret_598 = vmulxd_f64(__s0_598, __noswap_vgetq_lane_f64(__rev1_598, __p2_598)); \
51488   __ret_598; \
51489 })
51490 #endif
51491 
51492 #ifdef __LITTLE_ENDIAN__
51493 #define vmulxs_laneq_f32(__p0_599, __p1_599, __p2_599) __extension__ ({ \
51494   float32_t __ret_599; \
51495   float32_t __s0_599 = __p0_599; \
51496   float32x4_t __s1_599 = __p1_599; \
51497   __ret_599 = vmulxs_f32(__s0_599, vgetq_lane_f32(__s1_599, __p2_599)); \
51498   __ret_599; \
51499 })
51500 #else
51501 #define vmulxs_laneq_f32(__p0_600, __p1_600, __p2_600) __extension__ ({ \
51502   float32_t __ret_600; \
51503   float32_t __s0_600 = __p0_600; \
51504   float32x4_t __s1_600 = __p1_600; \
51505   float32x4_t __rev1_600;  __rev1_600 = __builtin_shufflevector(__s1_600, __s1_600, 3, 2, 1, 0); \
51506   __ret_600 = vmulxs_f32(__s0_600, __noswap_vgetq_lane_f32(__rev1_600, __p2_600)); \
51507   __ret_600; \
51508 })
51509 #endif
51510 
51511 #ifdef __LITTLE_ENDIAN__
51512 #define vmulxq_laneq_f64(__p0_601, __p1_601, __p2_601) __extension__ ({ \
51513   float64x2_t __ret_601; \
51514   float64x2_t __s0_601 = __p0_601; \
51515   float64x2_t __s1_601 = __p1_601; \
51516   __ret_601 = vmulxq_f64(__s0_601, splatq_laneq_f64(__s1_601, __p2_601)); \
51517   __ret_601; \
51518 })
51519 #else
51520 #define vmulxq_laneq_f64(__p0_602, __p1_602, __p2_602) __extension__ ({ \
51521   float64x2_t __ret_602; \
51522   float64x2_t __s0_602 = __p0_602; \
51523   float64x2_t __s1_602 = __p1_602; \
51524   float64x2_t __rev0_602;  __rev0_602 = __builtin_shufflevector(__s0_602, __s0_602, 1, 0); \
51525   float64x2_t __rev1_602;  __rev1_602 = __builtin_shufflevector(__s1_602, __s1_602, 1, 0); \
51526   __ret_602 = __noswap_vmulxq_f64(__rev0_602, __noswap_splatq_laneq_f64(__rev1_602, __p2_602)); \
51527   __ret_602 = __builtin_shufflevector(__ret_602, __ret_602, 1, 0); \
51528   __ret_602; \
51529 })
51530 #endif
51531 
51532 #ifdef __LITTLE_ENDIAN__
51533 #define vmulxq_laneq_f32(__p0_603, __p1_603, __p2_603) __extension__ ({ \
51534   float32x4_t __ret_603; \
51535   float32x4_t __s0_603 = __p0_603; \
51536   float32x4_t __s1_603 = __p1_603; \
51537   __ret_603 = vmulxq_f32(__s0_603, splatq_laneq_f32(__s1_603, __p2_603)); \
51538   __ret_603; \
51539 })
51540 #else
51541 #define vmulxq_laneq_f32(__p0_604, __p1_604, __p2_604) __extension__ ({ \
51542   float32x4_t __ret_604; \
51543   float32x4_t __s0_604 = __p0_604; \
51544   float32x4_t __s1_604 = __p1_604; \
51545   float32x4_t __rev0_604;  __rev0_604 = __builtin_shufflevector(__s0_604, __s0_604, 3, 2, 1, 0); \
51546   float32x4_t __rev1_604;  __rev1_604 = __builtin_shufflevector(__s1_604, __s1_604, 3, 2, 1, 0); \
51547   __ret_604 = __noswap_vmulxq_f32(__rev0_604, __noswap_splatq_laneq_f32(__rev1_604, __p2_604)); \
51548   __ret_604 = __builtin_shufflevector(__ret_604, __ret_604, 3, 2, 1, 0); \
51549   __ret_604; \
51550 })
51551 #endif
51552 
51553 #ifdef __LITTLE_ENDIAN__
51554 #define vmulx_laneq_f32(__p0_605, __p1_605, __p2_605) __extension__ ({ \
51555   float32x2_t __ret_605; \
51556   float32x2_t __s0_605 = __p0_605; \
51557   float32x4_t __s1_605 = __p1_605; \
51558   __ret_605 = vmulx_f32(__s0_605, splat_laneq_f32(__s1_605, __p2_605)); \
51559   __ret_605; \
51560 })
51561 #else
51562 #define vmulx_laneq_f32(__p0_606, __p1_606, __p2_606) __extension__ ({ \
51563   float32x2_t __ret_606; \
51564   float32x2_t __s0_606 = __p0_606; \
51565   float32x4_t __s1_606 = __p1_606; \
51566   float32x2_t __rev0_606;  __rev0_606 = __builtin_shufflevector(__s0_606, __s0_606, 1, 0); \
51567   float32x4_t __rev1_606;  __rev1_606 = __builtin_shufflevector(__s1_606, __s1_606, 3, 2, 1, 0); \
51568   __ret_606 = __noswap_vmulx_f32(__rev0_606, __noswap_splat_laneq_f32(__rev1_606, __p2_606)); \
51569   __ret_606 = __builtin_shufflevector(__ret_606, __ret_606, 1, 0); \
51570   __ret_606; \
51571 })
51572 #endif
51573 
51574 #ifdef __LITTLE_ENDIAN__
vnegq_f64(float64x2_t __p0)51575 __ai float64x2_t vnegq_f64(float64x2_t __p0) {
51576   float64x2_t __ret;
51577   __ret = -__p0;
51578   return __ret;
51579 }
51580 #else
vnegq_f64(float64x2_t __p0)51581 __ai float64x2_t vnegq_f64(float64x2_t __p0) {
51582   float64x2_t __ret;
51583   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51584   __ret = -__rev0;
51585   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51586   return __ret;
51587 }
51588 #endif
51589 
51590 #ifdef __LITTLE_ENDIAN__
vnegq_s64(int64x2_t __p0)51591 __ai int64x2_t vnegq_s64(int64x2_t __p0) {
51592   int64x2_t __ret;
51593   __ret = -__p0;
51594   return __ret;
51595 }
51596 #else
vnegq_s64(int64x2_t __p0)51597 __ai int64x2_t vnegq_s64(int64x2_t __p0) {
51598   int64x2_t __ret;
51599   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51600   __ret = -__rev0;
51601   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51602   return __ret;
51603 }
51604 #endif
51605 
vneg_f64(float64x1_t __p0)51606 __ai float64x1_t vneg_f64(float64x1_t __p0) {
51607   float64x1_t __ret;
51608   __ret = -__p0;
51609   return __ret;
51610 }
vneg_s64(int64x1_t __p0)51611 __ai int64x1_t vneg_s64(int64x1_t __p0) {
51612   int64x1_t __ret;
51613   __ret = -__p0;
51614   return __ret;
51615 }
vnegd_s64(int64_t __p0)51616 __ai int64_t vnegd_s64(int64_t __p0) {
51617   int64_t __ret;
51618   __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
51619   return __ret;
51620 }
51621 #ifdef __LITTLE_ENDIAN__
vpaddq_u8(uint8x16_t __p0,uint8x16_t __p1)51622 __ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
51623   uint8x16_t __ret;
51624   __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
51625   return __ret;
51626 }
51627 #else
vpaddq_u8(uint8x16_t __p0,uint8x16_t __p1)51628 __ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
51629   uint8x16_t __ret;
51630   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51631   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51632   __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
51633   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51634   return __ret;
51635 }
51636 #endif
51637 
51638 #ifdef __LITTLE_ENDIAN__
vpaddq_u32(uint32x4_t __p0,uint32x4_t __p1)51639 __ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
51640   uint32x4_t __ret;
51641   __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
51642   return __ret;
51643 }
51644 #else
vpaddq_u32(uint32x4_t __p0,uint32x4_t __p1)51645 __ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
51646   uint32x4_t __ret;
51647   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51648   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
51649   __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
51650   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51651   return __ret;
51652 }
51653 #endif
51654 
51655 #ifdef __LITTLE_ENDIAN__
vpaddq_u64(uint64x2_t __p0,uint64x2_t __p1)51656 __ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
51657   uint64x2_t __ret;
51658   __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
51659   return __ret;
51660 }
51661 #else
vpaddq_u64(uint64x2_t __p0,uint64x2_t __p1)51662 __ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
51663   uint64x2_t __ret;
51664   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51665   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51666   __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
51667   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51668   return __ret;
51669 }
51670 #endif
51671 
51672 #ifdef __LITTLE_ENDIAN__
vpaddq_u16(uint16x8_t __p0,uint16x8_t __p1)51673 __ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
51674   uint16x8_t __ret;
51675   __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
51676   return __ret;
51677 }
51678 #else
vpaddq_u16(uint16x8_t __p0,uint16x8_t __p1)51679 __ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
51680   uint16x8_t __ret;
51681   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51682   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
51683   __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
51684   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
51685   return __ret;
51686 }
51687 #endif
51688 
51689 #ifdef __LITTLE_ENDIAN__
vpaddq_s8(int8x16_t __p0,int8x16_t __p1)51690 __ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
51691   int8x16_t __ret;
51692   __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
51693   return __ret;
51694 }
51695 #else
vpaddq_s8(int8x16_t __p0,int8x16_t __p1)51696 __ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
51697   int8x16_t __ret;
51698   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51699   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51700   __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
51701   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51702   return __ret;
51703 }
51704 #endif
51705 
51706 #ifdef __LITTLE_ENDIAN__
vpaddq_f64(float64x2_t __p0,float64x2_t __p1)51707 __ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
51708   float64x2_t __ret;
51709   __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
51710   return __ret;
51711 }
51712 #else
vpaddq_f64(float64x2_t __p0,float64x2_t __p1)51713 __ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
51714   float64x2_t __ret;
51715   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51716   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51717   __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
51718   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51719   return __ret;
51720 }
51721 #endif
51722 
51723 #ifdef __LITTLE_ENDIAN__
vpaddq_f32(float32x4_t __p0,float32x4_t __p1)51724 __ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
51725   float32x4_t __ret;
51726   __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
51727   return __ret;
51728 }
51729 #else
vpaddq_f32(float32x4_t __p0,float32x4_t __p1)51730 __ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
51731   float32x4_t __ret;
51732   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51733   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
51734   __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
51735   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51736   return __ret;
51737 }
51738 #endif
51739 
51740 #ifdef __LITTLE_ENDIAN__
vpaddq_s32(int32x4_t __p0,int32x4_t __p1)51741 __ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
51742   int32x4_t __ret;
51743   __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
51744   return __ret;
51745 }
51746 #else
vpaddq_s32(int32x4_t __p0,int32x4_t __p1)51747 __ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
51748   int32x4_t __ret;
51749   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51750   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
51751   __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
51752   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51753   return __ret;
51754 }
51755 #endif
51756 
51757 #ifdef __LITTLE_ENDIAN__
vpaddq_s64(int64x2_t __p0,int64x2_t __p1)51758 __ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
51759   int64x2_t __ret;
51760   __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
51761   return __ret;
51762 }
51763 #else
vpaddq_s64(int64x2_t __p0,int64x2_t __p1)51764 __ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
51765   int64x2_t __ret;
51766   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51767   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51768   __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
51769   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51770   return __ret;
51771 }
51772 #endif
51773 
51774 #ifdef __LITTLE_ENDIAN__
vpaddq_s16(int16x8_t __p0,int16x8_t __p1)51775 __ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
51776   int16x8_t __ret;
51777   __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
51778   return __ret;
51779 }
51780 #else
vpaddq_s16(int16x8_t __p0,int16x8_t __p1)51781 __ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
51782   int16x8_t __ret;
51783   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51784   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
51785   __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
51786   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
51787   return __ret;
51788 }
51789 #endif
51790 
51791 #ifdef __LITTLE_ENDIAN__
vpaddd_u64(uint64x2_t __p0)51792 __ai uint64_t vpaddd_u64(uint64x2_t __p0) {
51793   uint64_t __ret;
51794   __ret = (uint64_t) __builtin_neon_vpaddd_u64(__p0);
51795   return __ret;
51796 }
51797 #else
vpaddd_u64(uint64x2_t __p0)51798 __ai uint64_t vpaddd_u64(uint64x2_t __p0) {
51799   uint64_t __ret;
51800   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51801   __ret = (uint64_t) __builtin_neon_vpaddd_u64(__rev0);
51802   return __ret;
51803 }
51804 #endif
51805 
51806 #ifdef __LITTLE_ENDIAN__
vpaddd_f64(float64x2_t __p0)51807 __ai float64_t vpaddd_f64(float64x2_t __p0) {
51808   float64_t __ret;
51809   __ret = (float64_t) __builtin_neon_vpaddd_f64(__p0);
51810   return __ret;
51811 }
51812 #else
vpaddd_f64(float64x2_t __p0)51813 __ai float64_t vpaddd_f64(float64x2_t __p0) {
51814   float64_t __ret;
51815   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51816   __ret = (float64_t) __builtin_neon_vpaddd_f64(__rev0);
51817   return __ret;
51818 }
51819 #endif
51820 
51821 #ifdef __LITTLE_ENDIAN__
vpaddd_s64(int64x2_t __p0)51822 __ai int64_t vpaddd_s64(int64x2_t __p0) {
51823   int64_t __ret;
51824   __ret = (int64_t) __builtin_neon_vpaddd_s64(__p0);
51825   return __ret;
51826 }
51827 #else
vpaddd_s64(int64x2_t __p0)51828 __ai int64_t vpaddd_s64(int64x2_t __p0) {
51829   int64_t __ret;
51830   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51831   __ret = (int64_t) __builtin_neon_vpaddd_s64(__rev0);
51832   return __ret;
51833 }
51834 #endif
51835 
51836 #ifdef __LITTLE_ENDIAN__
vpadds_f32(float32x2_t __p0)51837 __ai float32_t vpadds_f32(float32x2_t __p0) {
51838   float32_t __ret;
51839   __ret = (float32_t) __builtin_neon_vpadds_f32(__p0);
51840   return __ret;
51841 }
51842 #else
vpadds_f32(float32x2_t __p0)51843 __ai float32_t vpadds_f32(float32x2_t __p0) {
51844   float32_t __ret;
51845   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51846   __ret = (float32_t) __builtin_neon_vpadds_f32(__rev0);
51847   return __ret;
51848 }
51849 #endif
51850 
51851 #ifdef __LITTLE_ENDIAN__
vpmaxq_u8(uint8x16_t __p0,uint8x16_t __p1)51852 __ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
51853   uint8x16_t __ret;
51854   __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
51855   return __ret;
51856 }
51857 #else
vpmaxq_u8(uint8x16_t __p0,uint8x16_t __p1)51858 __ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
51859   uint8x16_t __ret;
51860   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51861   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51862   __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
51863   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51864   return __ret;
51865 }
51866 #endif
51867 
51868 #ifdef __LITTLE_ENDIAN__
vpmaxq_u32(uint32x4_t __p0,uint32x4_t __p1)51869 __ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
51870   uint32x4_t __ret;
51871   __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
51872   return __ret;
51873 }
51874 #else
vpmaxq_u32(uint32x4_t __p0,uint32x4_t __p1)51875 __ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
51876   uint32x4_t __ret;
51877   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51878   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
51879   __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
51880   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51881   return __ret;
51882 }
51883 #endif
51884 
51885 #ifdef __LITTLE_ENDIAN__
vpmaxq_u16(uint16x8_t __p0,uint16x8_t __p1)51886 __ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
51887   uint16x8_t __ret;
51888   __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
51889   return __ret;
51890 }
51891 #else
vpmaxq_u16(uint16x8_t __p0,uint16x8_t __p1)51892 __ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
51893   uint16x8_t __ret;
51894   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51895   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
51896   __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
51897   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
51898   return __ret;
51899 }
51900 #endif
51901 
51902 #ifdef __LITTLE_ENDIAN__
vpmaxq_s8(int8x16_t __p0,int8x16_t __p1)51903 __ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
51904   int8x16_t __ret;
51905   __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
51906   return __ret;
51907 }
51908 #else
vpmaxq_s8(int8x16_t __p0,int8x16_t __p1)51909 __ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
51910   int8x16_t __ret;
51911   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51912   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51913   __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
51914   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51915   return __ret;
51916 }
51917 #endif
51918 
51919 #ifdef __LITTLE_ENDIAN__
vpmaxq_f64(float64x2_t __p0,float64x2_t __p1)51920 __ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51921   float64x2_t __ret;
51922   __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
51923   return __ret;
51924 }
51925 #else
vpmaxq_f64(float64x2_t __p0,float64x2_t __p1)51926 __ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51927   float64x2_t __ret;
51928   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51929   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51930   __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
51931   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51932   return __ret;
51933 }
51934 #endif
51935 
51936 #ifdef __LITTLE_ENDIAN__
vpmaxq_f32(float32x4_t __p0,float32x4_t __p1)51937 __ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
51938   float32x4_t __ret;
51939   __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
51940   return __ret;
51941 }
51942 #else
vpmaxq_f32(float32x4_t __p0,float32x4_t __p1)51943 __ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
51944   float32x4_t __ret;
51945   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51946   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
51947   __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
51948   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51949   return __ret;
51950 }
51951 #endif
51952 
51953 #ifdef __LITTLE_ENDIAN__
vpmaxq_s32(int32x4_t __p0,int32x4_t __p1)51954 __ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
51955   int32x4_t __ret;
51956   __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
51957   return __ret;
51958 }
51959 #else
vpmaxq_s32(int32x4_t __p0,int32x4_t __p1)51960 __ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
51961   int32x4_t __ret;
51962   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51963   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
51964   __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
51965   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51966   return __ret;
51967 }
51968 #endif
51969 
51970 #ifdef __LITTLE_ENDIAN__
vpmaxq_s16(int16x8_t __p0,int16x8_t __p1)51971 __ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
51972   int16x8_t __ret;
51973   __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
51974   return __ret;
51975 }
51976 #else
vpmaxq_s16(int16x8_t __p0,int16x8_t __p1)51977 __ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
51978   int16x8_t __ret;
51979   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51980   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
51981   __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
51982   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
51983   return __ret;
51984 }
51985 #endif
51986 
51987 #ifdef __LITTLE_ENDIAN__
vpmaxqd_f64(float64x2_t __p0)51988 __ai float64_t vpmaxqd_f64(float64x2_t __p0) {
51989   float64_t __ret;
51990   __ret = (float64_t) __builtin_neon_vpmaxqd_f64(__p0);
51991   return __ret;
51992 }
51993 #else
vpmaxqd_f64(float64x2_t __p0)51994 __ai float64_t vpmaxqd_f64(float64x2_t __p0) {
51995   float64_t __ret;
51996   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51997   __ret = (float64_t) __builtin_neon_vpmaxqd_f64(__rev0);
51998   return __ret;
51999 }
52000 #endif
52001 
52002 #ifdef __LITTLE_ENDIAN__
vpmaxs_f32(float32x2_t __p0)52003 __ai float32_t vpmaxs_f32(float32x2_t __p0) {
52004   float32_t __ret;
52005   __ret = (float32_t) __builtin_neon_vpmaxs_f32(__p0);
52006   return __ret;
52007 }
52008 #else
vpmaxs_f32(float32x2_t __p0)52009 __ai float32_t vpmaxs_f32(float32x2_t __p0) {
52010   float32_t __ret;
52011   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52012   __ret = (float32_t) __builtin_neon_vpmaxs_f32(__rev0);
52013   return __ret;
52014 }
52015 #endif
52016 
52017 #ifdef __LITTLE_ENDIAN__
vpmaxnmq_f64(float64x2_t __p0,float64x2_t __p1)52018 __ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
52019   float64x2_t __ret;
52020   __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
52021   return __ret;
52022 }
52023 #else
vpmaxnmq_f64(float64x2_t __p0,float64x2_t __p1)52024 __ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
52025   float64x2_t __ret;
52026   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52027   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52028   __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
52029   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52030   return __ret;
52031 }
52032 #endif
52033 
52034 #ifdef __LITTLE_ENDIAN__
vpmaxnmq_f32(float32x4_t __p0,float32x4_t __p1)52035 __ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
52036   float32x4_t __ret;
52037   __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
52038   return __ret;
52039 }
52040 #else
vpmaxnmq_f32(float32x4_t __p0,float32x4_t __p1)52041 __ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
52042   float32x4_t __ret;
52043   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52044   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52045   __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
52046   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52047   return __ret;
52048 }
52049 #endif
52050 
52051 #ifdef __LITTLE_ENDIAN__
vpmaxnm_f32(float32x2_t __p0,float32x2_t __p1)52052 __ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
52053   float32x2_t __ret;
52054   __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
52055   return __ret;
52056 }
52057 #else
vpmaxnm_f32(float32x2_t __p0,float32x2_t __p1)52058 __ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
52059   float32x2_t __ret;
52060   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52061   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52062   __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
52063   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52064   return __ret;
52065 }
52066 #endif
52067 
52068 #ifdef __LITTLE_ENDIAN__
vpmaxnmqd_f64(float64x2_t __p0)52069 __ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
52070   float64_t __ret;
52071   __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64(__p0);
52072   return __ret;
52073 }
52074 #else
vpmaxnmqd_f64(float64x2_t __p0)52075 __ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
52076   float64_t __ret;
52077   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52078   __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64(__rev0);
52079   return __ret;
52080 }
52081 #endif
52082 
52083 #ifdef __LITTLE_ENDIAN__
vpmaxnms_f32(float32x2_t __p0)52084 __ai float32_t vpmaxnms_f32(float32x2_t __p0) {
52085   float32_t __ret;
52086   __ret = (float32_t) __builtin_neon_vpmaxnms_f32(__p0);
52087   return __ret;
52088 }
52089 #else
vpmaxnms_f32(float32x2_t __p0)52090 __ai float32_t vpmaxnms_f32(float32x2_t __p0) {
52091   float32_t __ret;
52092   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52093   __ret = (float32_t) __builtin_neon_vpmaxnms_f32(__rev0);
52094   return __ret;
52095 }
52096 #endif
52097 
52098 #ifdef __LITTLE_ENDIAN__
vpminq_u8(uint8x16_t __p0,uint8x16_t __p1)52099 __ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
52100   uint8x16_t __ret;
52101   __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
52102   return __ret;
52103 }
52104 #else
vpminq_u8(uint8x16_t __p0,uint8x16_t __p1)52105 __ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
52106   uint8x16_t __ret;
52107   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52108   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52109   __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
52110   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52111   return __ret;
52112 }
52113 #endif
52114 
52115 #ifdef __LITTLE_ENDIAN__
vpminq_u32(uint32x4_t __p0,uint32x4_t __p1)52116 __ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
52117   uint32x4_t __ret;
52118   __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
52119   return __ret;
52120 }
52121 #else
vpminq_u32(uint32x4_t __p0,uint32x4_t __p1)52122 __ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
52123   uint32x4_t __ret;
52124   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52125   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52126   __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
52127   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52128   return __ret;
52129 }
52130 #endif
52131 
52132 #ifdef __LITTLE_ENDIAN__
vpminq_u16(uint16x8_t __p0,uint16x8_t __p1)52133 __ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
52134   uint16x8_t __ret;
52135   __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
52136   return __ret;
52137 }
52138 #else
vpminq_u16(uint16x8_t __p0,uint16x8_t __p1)52139 __ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
52140   uint16x8_t __ret;
52141   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52142   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
52143   __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
52144   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
52145   return __ret;
52146 }
52147 #endif
52148 
52149 #ifdef __LITTLE_ENDIAN__
vpminq_s8(int8x16_t __p0,int8x16_t __p1)52150 __ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
52151   int8x16_t __ret;
52152   __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
52153   return __ret;
52154 }
52155 #else
vpminq_s8(int8x16_t __p0,int8x16_t __p1)52156 __ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
52157   int8x16_t __ret;
52158   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52159   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52160   __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
52161   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52162   return __ret;
52163 }
52164 #endif
52165 
52166 #ifdef __LITTLE_ENDIAN__
vpminq_f64(float64x2_t __p0,float64x2_t __p1)52167 __ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
52168   float64x2_t __ret;
52169   __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
52170   return __ret;
52171 }
52172 #else
vpminq_f64(float64x2_t __p0,float64x2_t __p1)52173 __ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
52174   float64x2_t __ret;
52175   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52176   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52177   __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
52178   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52179   return __ret;
52180 }
52181 #endif
52182 
52183 #ifdef __LITTLE_ENDIAN__
vpminq_f32(float32x4_t __p0,float32x4_t __p1)52184 __ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
52185   float32x4_t __ret;
52186   __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
52187   return __ret;
52188 }
52189 #else
vpminq_f32(float32x4_t __p0,float32x4_t __p1)52190 __ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
52191   float32x4_t __ret;
52192   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52193   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52194   __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
52195   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52196   return __ret;
52197 }
52198 #endif
52199 
52200 #ifdef __LITTLE_ENDIAN__
vpminq_s32(int32x4_t __p0,int32x4_t __p1)52201 __ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
52202   int32x4_t __ret;
52203   __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
52204   return __ret;
52205 }
52206 #else
vpminq_s32(int32x4_t __p0,int32x4_t __p1)52207 __ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
52208   int32x4_t __ret;
52209   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52210   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52211   __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
52212   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52213   return __ret;
52214 }
52215 #endif
52216 
52217 #ifdef __LITTLE_ENDIAN__
vpminq_s16(int16x8_t __p0,int16x8_t __p1)52218 __ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
52219   int16x8_t __ret;
52220   __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
52221   return __ret;
52222 }
52223 #else
vpminq_s16(int16x8_t __p0,int16x8_t __p1)52224 __ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
52225   int16x8_t __ret;
52226   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52227   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
52228   __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
52229   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
52230   return __ret;
52231 }
52232 #endif
52233 
52234 #ifdef __LITTLE_ENDIAN__
vpminqd_f64(float64x2_t __p0)52235 __ai float64_t vpminqd_f64(float64x2_t __p0) {
52236   float64_t __ret;
52237   __ret = (float64_t) __builtin_neon_vpminqd_f64(__p0);
52238   return __ret;
52239 }
52240 #else
vpminqd_f64(float64x2_t __p0)52241 __ai float64_t vpminqd_f64(float64x2_t __p0) {
52242   float64_t __ret;
52243   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52244   __ret = (float64_t) __builtin_neon_vpminqd_f64(__rev0);
52245   return __ret;
52246 }
52247 #endif
52248 
52249 #ifdef __LITTLE_ENDIAN__
vpmins_f32(float32x2_t __p0)52250 __ai float32_t vpmins_f32(float32x2_t __p0) {
52251   float32_t __ret;
52252   __ret = (float32_t) __builtin_neon_vpmins_f32(__p0);
52253   return __ret;
52254 }
52255 #else
vpmins_f32(float32x2_t __p0)52256 __ai float32_t vpmins_f32(float32x2_t __p0) {
52257   float32_t __ret;
52258   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52259   __ret = (float32_t) __builtin_neon_vpmins_f32(__rev0);
52260   return __ret;
52261 }
52262 #endif
52263 
52264 #ifdef __LITTLE_ENDIAN__
vpminnmq_f64(float64x2_t __p0,float64x2_t __p1)52265 __ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
52266   float64x2_t __ret;
52267   __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
52268   return __ret;
52269 }
52270 #else
vpminnmq_f64(float64x2_t __p0,float64x2_t __p1)52271 __ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
52272   float64x2_t __ret;
52273   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52274   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52275   __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
52276   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52277   return __ret;
52278 }
52279 #endif
52280 
52281 #ifdef __LITTLE_ENDIAN__
vpminnmq_f32(float32x4_t __p0,float32x4_t __p1)52282 __ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
52283   float32x4_t __ret;
52284   __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
52285   return __ret;
52286 }
52287 #else
vpminnmq_f32(float32x4_t __p0,float32x4_t __p1)52288 __ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
52289   float32x4_t __ret;
52290   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52291   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52292   __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
52293   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52294   return __ret;
52295 }
52296 #endif
52297 
52298 #ifdef __LITTLE_ENDIAN__
vpminnm_f32(float32x2_t __p0,float32x2_t __p1)52299 __ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
52300   float32x2_t __ret;
52301   __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
52302   return __ret;
52303 }
52304 #else
vpminnm_f32(float32x2_t __p0,float32x2_t __p1)52305 __ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
52306   float32x2_t __ret;
52307   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52308   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52309   __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
52310   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52311   return __ret;
52312 }
52313 #endif
52314 
52315 #ifdef __LITTLE_ENDIAN__
vpminnmqd_f64(float64x2_t __p0)52316 __ai float64_t vpminnmqd_f64(float64x2_t __p0) {
52317   float64_t __ret;
52318   __ret = (float64_t) __builtin_neon_vpminnmqd_f64(__p0);
52319   return __ret;
52320 }
52321 #else
vpminnmqd_f64(float64x2_t __p0)52322 __ai float64_t vpminnmqd_f64(float64x2_t __p0) {
52323   float64_t __ret;
52324   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52325   __ret = (float64_t) __builtin_neon_vpminnmqd_f64(__rev0);
52326   return __ret;
52327 }
52328 #endif
52329 
52330 #ifdef __LITTLE_ENDIAN__
vpminnms_f32(float32x2_t __p0)52331 __ai float32_t vpminnms_f32(float32x2_t __p0) {
52332   float32_t __ret;
52333   __ret = (float32_t) __builtin_neon_vpminnms_f32(__p0);
52334   return __ret;
52335 }
52336 #else
vpminnms_f32(float32x2_t __p0)52337 __ai float32_t vpminnms_f32(float32x2_t __p0) {
52338   float32_t __ret;
52339   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52340   __ret = (float32_t) __builtin_neon_vpminnms_f32(__rev0);
52341   return __ret;
52342 }
52343 #endif
52344 
52345 #ifdef __LITTLE_ENDIAN__
vqabsq_s64(int64x2_t __p0)52346 __ai int64x2_t vqabsq_s64(int64x2_t __p0) {
52347   int64x2_t __ret;
52348   __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 35);
52349   return __ret;
52350 }
52351 #else
vqabsq_s64(int64x2_t __p0)52352 __ai int64x2_t vqabsq_s64(int64x2_t __p0) {
52353   int64x2_t __ret;
52354   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52355   __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 35);
52356   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52357   return __ret;
52358 }
52359 #endif
52360 
vqabs_s64(int64x1_t __p0)52361 __ai int64x1_t vqabs_s64(int64x1_t __p0) {
52362   int64x1_t __ret;
52363   __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
52364   return __ret;
52365 }
vqabsb_s8(int8_t __p0)52366 __ai int8_t vqabsb_s8(int8_t __p0) {
52367   int8_t __ret;
52368   __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
52369   return __ret;
52370 }
vqabss_s32(int32_t __p0)52371 __ai int32_t vqabss_s32(int32_t __p0) {
52372   int32_t __ret;
52373   __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
52374   return __ret;
52375 }
vqabsd_s64(int64_t __p0)52376 __ai int64_t vqabsd_s64(int64_t __p0) {
52377   int64_t __ret;
52378   __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
52379   return __ret;
52380 }
vqabsh_s16(int16_t __p0)52381 __ai int16_t vqabsh_s16(int16_t __p0) {
52382   int16_t __ret;
52383   __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
52384   return __ret;
52385 }
vqaddb_u8(uint8_t __p0,uint8_t __p1)52386 __ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
52387   uint8_t __ret;
52388   __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
52389   return __ret;
52390 }
vqadds_u32(uint32_t __p0,uint32_t __p1)52391 __ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
52392   uint32_t __ret;
52393   __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
52394   return __ret;
52395 }
vqaddd_u64(uint64_t __p0,uint64_t __p1)52396 __ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
52397   uint64_t __ret;
52398   __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
52399   return __ret;
52400 }
vqaddh_u16(uint16_t __p0,uint16_t __p1)52401 __ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
52402   uint16_t __ret;
52403   __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
52404   return __ret;
52405 }
vqaddb_s8(int8_t __p0,int8_t __p1)52406 __ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
52407   int8_t __ret;
52408   __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
52409   return __ret;
52410 }
vqadds_s32(int32_t __p0,int32_t __p1)52411 __ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
52412   int32_t __ret;
52413   __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
52414   return __ret;
52415 }
vqaddd_s64(int64_t __p0,int64_t __p1)52416 __ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
52417   int64_t __ret;
52418   __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
52419   return __ret;
52420 }
vqaddh_s16(int16_t __p0,int16_t __p1)52421 __ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
52422   int16_t __ret;
52423   __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
52424   return __ret;
52425 }
vqdmlals_s32(int64_t __p0,int32_t __p1,int32_t __p2)52426 __ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
52427   int64_t __ret;
52428   __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
52429   return __ret;
52430 }
vqdmlalh_s16(int32_t __p0,int16_t __p1,int16_t __p2)52431 __ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
52432   int32_t __ret;
52433   __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
52434   return __ret;
52435 }
52436 #ifdef __LITTLE_ENDIAN__
vqdmlal_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)52437 __ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
52438   int64x2_t __ret;
52439   __ret = vqdmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
52440   return __ret;
52441 }
52442 #else
vqdmlal_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)52443 __ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
52444   int64x2_t __ret;
52445   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52446   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52447   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
52448   __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
52449   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52450   return __ret;
52451 }
52452 #endif
52453 
52454 #ifdef __LITTLE_ENDIAN__
vqdmlal_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)52455 __ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
52456   int32x4_t __ret;
52457   __ret = vqdmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
52458   return __ret;
52459 }
52460 #else
vqdmlal_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)52461 __ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
52462   int32x4_t __ret;
52463   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52464   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
52465   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
52466   __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
52467   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52468   return __ret;
52469 }
52470 #endif
52471 
52472 #ifdef __LITTLE_ENDIAN__
52473 #define vqdmlal_high_lane_s32(__p0_607, __p1_607, __p2_607, __p3_607) __extension__ ({ \
52474   int64x2_t __ret_607; \
52475   int64x2_t __s0_607 = __p0_607; \
52476   int32x4_t __s1_607 = __p1_607; \
52477   int32x2_t __s2_607 = __p2_607; \
52478   __ret_607 = vqdmlal_s32(__s0_607, vget_high_s32(__s1_607), splat_lane_s32(__s2_607, __p3_607)); \
52479   __ret_607; \
52480 })
52481 #else
52482 #define vqdmlal_high_lane_s32(__p0_608, __p1_608, __p2_608, __p3_608) __extension__ ({ \
52483   int64x2_t __ret_608; \
52484   int64x2_t __s0_608 = __p0_608; \
52485   int32x4_t __s1_608 = __p1_608; \
52486   int32x2_t __s2_608 = __p2_608; \
52487   int64x2_t __rev0_608;  __rev0_608 = __builtin_shufflevector(__s0_608, __s0_608, 1, 0); \
52488   int32x4_t __rev1_608;  __rev1_608 = __builtin_shufflevector(__s1_608, __s1_608, 3, 2, 1, 0); \
52489   int32x2_t __rev2_608;  __rev2_608 = __builtin_shufflevector(__s2_608, __s2_608, 1, 0); \
52490   __ret_608 = __noswap_vqdmlal_s32(__rev0_608, __noswap_vget_high_s32(__rev1_608), __noswap_splat_lane_s32(__rev2_608, __p3_608)); \
52491   __ret_608 = __builtin_shufflevector(__ret_608, __ret_608, 1, 0); \
52492   __ret_608; \
52493 })
52494 #endif
52495 
52496 #ifdef __LITTLE_ENDIAN__
52497 #define vqdmlal_high_lane_s16(__p0_609, __p1_609, __p2_609, __p3_609) __extension__ ({ \
52498   int32x4_t __ret_609; \
52499   int32x4_t __s0_609 = __p0_609; \
52500   int16x8_t __s1_609 = __p1_609; \
52501   int16x4_t __s2_609 = __p2_609; \
52502   __ret_609 = vqdmlal_s16(__s0_609, vget_high_s16(__s1_609), splat_lane_s16(__s2_609, __p3_609)); \
52503   __ret_609; \
52504 })
52505 #else
52506 #define vqdmlal_high_lane_s16(__p0_610, __p1_610, __p2_610, __p3_610) __extension__ ({ \
52507   int32x4_t __ret_610; \
52508   int32x4_t __s0_610 = __p0_610; \
52509   int16x8_t __s1_610 = __p1_610; \
52510   int16x4_t __s2_610 = __p2_610; \
52511   int32x4_t __rev0_610;  __rev0_610 = __builtin_shufflevector(__s0_610, __s0_610, 3, 2, 1, 0); \
52512   int16x8_t __rev1_610;  __rev1_610 = __builtin_shufflevector(__s1_610, __s1_610, 7, 6, 5, 4, 3, 2, 1, 0); \
52513   int16x4_t __rev2_610;  __rev2_610 = __builtin_shufflevector(__s2_610, __s2_610, 3, 2, 1, 0); \
52514   __ret_610 = __noswap_vqdmlal_s16(__rev0_610, __noswap_vget_high_s16(__rev1_610), __noswap_splat_lane_s16(__rev2_610, __p3_610)); \
52515   __ret_610 = __builtin_shufflevector(__ret_610, __ret_610, 3, 2, 1, 0); \
52516   __ret_610; \
52517 })
52518 #endif
52519 
52520 #ifdef __LITTLE_ENDIAN__
52521 #define vqdmlal_high_laneq_s32(__p0_611, __p1_611, __p2_611, __p3_611) __extension__ ({ \
52522   int64x2_t __ret_611; \
52523   int64x2_t __s0_611 = __p0_611; \
52524   int32x4_t __s1_611 = __p1_611; \
52525   int32x4_t __s2_611 = __p2_611; \
52526   __ret_611 = vqdmlal_s32(__s0_611, vget_high_s32(__s1_611), splat_laneq_s32(__s2_611, __p3_611)); \
52527   __ret_611; \
52528 })
52529 #else
52530 #define vqdmlal_high_laneq_s32(__p0_612, __p1_612, __p2_612, __p3_612) __extension__ ({ \
52531   int64x2_t __ret_612; \
52532   int64x2_t __s0_612 = __p0_612; \
52533   int32x4_t __s1_612 = __p1_612; \
52534   int32x4_t __s2_612 = __p2_612; \
52535   int64x2_t __rev0_612;  __rev0_612 = __builtin_shufflevector(__s0_612, __s0_612, 1, 0); \
52536   int32x4_t __rev1_612;  __rev1_612 = __builtin_shufflevector(__s1_612, __s1_612, 3, 2, 1, 0); \
52537   int32x4_t __rev2_612;  __rev2_612 = __builtin_shufflevector(__s2_612, __s2_612, 3, 2, 1, 0); \
52538   __ret_612 = __noswap_vqdmlal_s32(__rev0_612, __noswap_vget_high_s32(__rev1_612), __noswap_splat_laneq_s32(__rev2_612, __p3_612)); \
52539   __ret_612 = __builtin_shufflevector(__ret_612, __ret_612, 1, 0); \
52540   __ret_612; \
52541 })
52542 #endif
52543 
52544 #ifdef __LITTLE_ENDIAN__
52545 #define vqdmlal_high_laneq_s16(__p0_613, __p1_613, __p2_613, __p3_613) __extension__ ({ \
52546   int32x4_t __ret_613; \
52547   int32x4_t __s0_613 = __p0_613; \
52548   int16x8_t __s1_613 = __p1_613; \
52549   int16x8_t __s2_613 = __p2_613; \
52550   __ret_613 = vqdmlal_s16(__s0_613, vget_high_s16(__s1_613), splat_laneq_s16(__s2_613, __p3_613)); \
52551   __ret_613; \
52552 })
52553 #else
52554 #define vqdmlal_high_laneq_s16(__p0_614, __p1_614, __p2_614, __p3_614) __extension__ ({ \
52555   int32x4_t __ret_614; \
52556   int32x4_t __s0_614 = __p0_614; \
52557   int16x8_t __s1_614 = __p1_614; \
52558   int16x8_t __s2_614 = __p2_614; \
52559   int32x4_t __rev0_614;  __rev0_614 = __builtin_shufflevector(__s0_614, __s0_614, 3, 2, 1, 0); \
52560   int16x8_t __rev1_614;  __rev1_614 = __builtin_shufflevector(__s1_614, __s1_614, 7, 6, 5, 4, 3, 2, 1, 0); \
52561   int16x8_t __rev2_614;  __rev2_614 = __builtin_shufflevector(__s2_614, __s2_614, 7, 6, 5, 4, 3, 2, 1, 0); \
52562   __ret_614 = __noswap_vqdmlal_s16(__rev0_614, __noswap_vget_high_s16(__rev1_614), __noswap_splat_laneq_s16(__rev2_614, __p3_614)); \
52563   __ret_614 = __builtin_shufflevector(__ret_614, __ret_614, 3, 2, 1, 0); \
52564   __ret_614; \
52565 })
52566 #endif
52567 
52568 #ifdef __LITTLE_ENDIAN__
vqdmlal_high_n_s32(int64x2_t __p0,int32x4_t __p1,int32_t __p2)52569 __ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
52570   int64x2_t __ret;
52571   __ret = vqdmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
52572   return __ret;
52573 }
52574 #else
vqdmlal_high_n_s32(int64x2_t __p0,int32x4_t __p1,int32_t __p2)52575 __ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
52576   int64x2_t __ret;
52577   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52578   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52579   __ret = __noswap_vqdmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
52580   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52581   return __ret;
52582 }
52583 #endif
52584 
52585 #ifdef __LITTLE_ENDIAN__
vqdmlal_high_n_s16(int32x4_t __p0,int16x8_t __p1,int16_t __p2)52586 __ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
52587   int32x4_t __ret;
52588   __ret = vqdmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
52589   return __ret;
52590 }
52591 #else
vqdmlal_high_n_s16(int32x4_t __p0,int16x8_t __p1,int16_t __p2)52592 __ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
52593   int32x4_t __ret;
52594   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52595   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
52596   __ret = __noswap_vqdmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
52597   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52598   return __ret;
52599 }
52600 #endif
52601 
52602 #ifdef __LITTLE_ENDIAN__
52603 #define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52604   int64_t __ret; \
52605   int64_t __s0 = __p0; \
52606   int32_t __s1 = __p1; \
52607   int32x2_t __s2 = __p2; \
52608   __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, __s2, __p3); \
52609   __ret; \
52610 })
52611 #else
52612 #define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52613   int64_t __ret; \
52614   int64_t __s0 = __p0; \
52615   int32_t __s1 = __p1; \
52616   int32x2_t __s2 = __p2; \
52617   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52618   __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, __rev2, __p3); \
52619   __ret; \
52620 })
52621 #endif
52622 
52623 #ifdef __LITTLE_ENDIAN__
52624 #define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52625   int32_t __ret; \
52626   int32_t __s0 = __p0; \
52627   int16_t __s1 = __p1; \
52628   int16x4_t __s2 = __p2; \
52629   __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, __s2, __p3); \
52630   __ret; \
52631 })
52632 #else
52633 #define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52634   int32_t __ret; \
52635   int32_t __s0 = __p0; \
52636   int16_t __s1 = __p1; \
52637   int16x4_t __s2 = __p2; \
52638   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52639   __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, __rev2, __p3); \
52640   __ret; \
52641 })
52642 #endif
52643 
52644 #ifdef __LITTLE_ENDIAN__
52645 #define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52646   int64_t __ret; \
52647   int64_t __s0 = __p0; \
52648   int32_t __s1 = __p1; \
52649   int32x4_t __s2 = __p2; \
52650   __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, __s2, __p3); \
52651   __ret; \
52652 })
52653 #else
52654 #define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52655   int64_t __ret; \
52656   int64_t __s0 = __p0; \
52657   int32_t __s1 = __p1; \
52658   int32x4_t __s2 = __p2; \
52659   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52660   __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, __rev2, __p3); \
52661   __ret; \
52662 })
52663 #endif
52664 
52665 #ifdef __LITTLE_ENDIAN__
52666 #define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52667   int32_t __ret; \
52668   int32_t __s0 = __p0; \
52669   int16_t __s1 = __p1; \
52670   int16x8_t __s2 = __p2; \
52671   __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, __s2, __p3); \
52672   __ret; \
52673 })
52674 #else
52675 #define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52676   int32_t __ret; \
52677   int32_t __s0 = __p0; \
52678   int16_t __s1 = __p1; \
52679   int16x8_t __s2 = __p2; \
52680   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52681   __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, __rev2, __p3); \
52682   __ret; \
52683 })
52684 #endif
52685 
52686 #ifdef __LITTLE_ENDIAN__
52687 #define vqdmlal_laneq_s32(__p0_615, __p1_615, __p2_615, __p3_615) __extension__ ({ \
52688   int64x2_t __ret_615; \
52689   int64x2_t __s0_615 = __p0_615; \
52690   int32x2_t __s1_615 = __p1_615; \
52691   int32x4_t __s2_615 = __p2_615; \
52692   __ret_615 = vqdmlal_s32(__s0_615, __s1_615, splat_laneq_s32(__s2_615, __p3_615)); \
52693   __ret_615; \
52694 })
52695 #else
52696 #define vqdmlal_laneq_s32(__p0_616, __p1_616, __p2_616, __p3_616) __extension__ ({ \
52697   int64x2_t __ret_616; \
52698   int64x2_t __s0_616 = __p0_616; \
52699   int32x2_t __s1_616 = __p1_616; \
52700   int32x4_t __s2_616 = __p2_616; \
52701   int64x2_t __rev0_616;  __rev0_616 = __builtin_shufflevector(__s0_616, __s0_616, 1, 0); \
52702   int32x2_t __rev1_616;  __rev1_616 = __builtin_shufflevector(__s1_616, __s1_616, 1, 0); \
52703   int32x4_t __rev2_616;  __rev2_616 = __builtin_shufflevector(__s2_616, __s2_616, 3, 2, 1, 0); \
52704   __ret_616 = __noswap_vqdmlal_s32(__rev0_616, __rev1_616, __noswap_splat_laneq_s32(__rev2_616, __p3_616)); \
52705   __ret_616 = __builtin_shufflevector(__ret_616, __ret_616, 1, 0); \
52706   __ret_616; \
52707 })
52708 #endif
52709 
52710 #ifdef __LITTLE_ENDIAN__
52711 #define vqdmlal_laneq_s16(__p0_617, __p1_617, __p2_617, __p3_617) __extension__ ({ \
52712   int32x4_t __ret_617; \
52713   int32x4_t __s0_617 = __p0_617; \
52714   int16x4_t __s1_617 = __p1_617; \
52715   int16x8_t __s2_617 = __p2_617; \
52716   __ret_617 = vqdmlal_s16(__s0_617, __s1_617, splat_laneq_s16(__s2_617, __p3_617)); \
52717   __ret_617; \
52718 })
52719 #else
52720 #define vqdmlal_laneq_s16(__p0_618, __p1_618, __p2_618, __p3_618) __extension__ ({ \
52721   int32x4_t __ret_618; \
52722   int32x4_t __s0_618 = __p0_618; \
52723   int16x4_t __s1_618 = __p1_618; \
52724   int16x8_t __s2_618 = __p2_618; \
52725   int32x4_t __rev0_618;  __rev0_618 = __builtin_shufflevector(__s0_618, __s0_618, 3, 2, 1, 0); \
52726   int16x4_t __rev1_618;  __rev1_618 = __builtin_shufflevector(__s1_618, __s1_618, 3, 2, 1, 0); \
52727   int16x8_t __rev2_618;  __rev2_618 = __builtin_shufflevector(__s2_618, __s2_618, 7, 6, 5, 4, 3, 2, 1, 0); \
52728   __ret_618 = __noswap_vqdmlal_s16(__rev0_618, __rev1_618, __noswap_splat_laneq_s16(__rev2_618, __p3_618)); \
52729   __ret_618 = __builtin_shufflevector(__ret_618, __ret_618, 3, 2, 1, 0); \
52730   __ret_618; \
52731 })
52732 #endif
52733 
vqdmlsls_s32(int64_t __p0,int32_t __p1,int32_t __p2)52734 __ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
52735   int64_t __ret;
52736   __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
52737   return __ret;
52738 }
vqdmlslh_s16(int32_t __p0,int16_t __p1,int16_t __p2)52739 __ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
52740   int32_t __ret;
52741   __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
52742   return __ret;
52743 }
52744 #ifdef __LITTLE_ENDIAN__
vqdmlsl_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)52745 __ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
52746   int64x2_t __ret;
52747   __ret = vqdmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
52748   return __ret;
52749 }
52750 #else
vqdmlsl_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)52751 __ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
52752   int64x2_t __ret;
52753   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52754   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52755   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
52756   __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
52757   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52758   return __ret;
52759 }
52760 #endif
52761 
52762 #ifdef __LITTLE_ENDIAN__
vqdmlsl_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)52763 __ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
52764   int32x4_t __ret;
52765   __ret = vqdmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
52766   return __ret;
52767 }
52768 #else
vqdmlsl_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)52769 __ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
52770   int32x4_t __ret;
52771   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52772   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
52773   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
52774   __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
52775   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52776   return __ret;
52777 }
52778 #endif
52779 
52780 #ifdef __LITTLE_ENDIAN__
52781 #define vqdmlsl_high_lane_s32(__p0_619, __p1_619, __p2_619, __p3_619) __extension__ ({ \
52782   int64x2_t __ret_619; \
52783   int64x2_t __s0_619 = __p0_619; \
52784   int32x4_t __s1_619 = __p1_619; \
52785   int32x2_t __s2_619 = __p2_619; \
52786   __ret_619 = vqdmlsl_s32(__s0_619, vget_high_s32(__s1_619), splat_lane_s32(__s2_619, __p3_619)); \
52787   __ret_619; \
52788 })
52789 #else
52790 #define vqdmlsl_high_lane_s32(__p0_620, __p1_620, __p2_620, __p3_620) __extension__ ({ \
52791   int64x2_t __ret_620; \
52792   int64x2_t __s0_620 = __p0_620; \
52793   int32x4_t __s1_620 = __p1_620; \
52794   int32x2_t __s2_620 = __p2_620; \
52795   int64x2_t __rev0_620;  __rev0_620 = __builtin_shufflevector(__s0_620, __s0_620, 1, 0); \
52796   int32x4_t __rev1_620;  __rev1_620 = __builtin_shufflevector(__s1_620, __s1_620, 3, 2, 1, 0); \
52797   int32x2_t __rev2_620;  __rev2_620 = __builtin_shufflevector(__s2_620, __s2_620, 1, 0); \
52798   __ret_620 = __noswap_vqdmlsl_s32(__rev0_620, __noswap_vget_high_s32(__rev1_620), __noswap_splat_lane_s32(__rev2_620, __p3_620)); \
52799   __ret_620 = __builtin_shufflevector(__ret_620, __ret_620, 1, 0); \
52800   __ret_620; \
52801 })
52802 #endif
52803 
52804 #ifdef __LITTLE_ENDIAN__
52805 #define vqdmlsl_high_lane_s16(__p0_621, __p1_621, __p2_621, __p3_621) __extension__ ({ \
52806   int32x4_t __ret_621; \
52807   int32x4_t __s0_621 = __p0_621; \
52808   int16x8_t __s1_621 = __p1_621; \
52809   int16x4_t __s2_621 = __p2_621; \
52810   __ret_621 = vqdmlsl_s16(__s0_621, vget_high_s16(__s1_621), splat_lane_s16(__s2_621, __p3_621)); \
52811   __ret_621; \
52812 })
52813 #else
52814 #define vqdmlsl_high_lane_s16(__p0_622, __p1_622, __p2_622, __p3_622) __extension__ ({ \
52815   int32x4_t __ret_622; \
52816   int32x4_t __s0_622 = __p0_622; \
52817   int16x8_t __s1_622 = __p1_622; \
52818   int16x4_t __s2_622 = __p2_622; \
52819   int32x4_t __rev0_622;  __rev0_622 = __builtin_shufflevector(__s0_622, __s0_622, 3, 2, 1, 0); \
52820   int16x8_t __rev1_622;  __rev1_622 = __builtin_shufflevector(__s1_622, __s1_622, 7, 6, 5, 4, 3, 2, 1, 0); \
52821   int16x4_t __rev2_622;  __rev2_622 = __builtin_shufflevector(__s2_622, __s2_622, 3, 2, 1, 0); \
52822   __ret_622 = __noswap_vqdmlsl_s16(__rev0_622, __noswap_vget_high_s16(__rev1_622), __noswap_splat_lane_s16(__rev2_622, __p3_622)); \
52823   __ret_622 = __builtin_shufflevector(__ret_622, __ret_622, 3, 2, 1, 0); \
52824   __ret_622; \
52825 })
52826 #endif
52827 
52828 #ifdef __LITTLE_ENDIAN__
52829 #define vqdmlsl_high_laneq_s32(__p0_623, __p1_623, __p2_623, __p3_623) __extension__ ({ \
52830   int64x2_t __ret_623; \
52831   int64x2_t __s0_623 = __p0_623; \
52832   int32x4_t __s1_623 = __p1_623; \
52833   int32x4_t __s2_623 = __p2_623; \
52834   __ret_623 = vqdmlsl_s32(__s0_623, vget_high_s32(__s1_623), splat_laneq_s32(__s2_623, __p3_623)); \
52835   __ret_623; \
52836 })
52837 #else
52838 #define vqdmlsl_high_laneq_s32(__p0_624, __p1_624, __p2_624, __p3_624) __extension__ ({ \
52839   int64x2_t __ret_624; \
52840   int64x2_t __s0_624 = __p0_624; \
52841   int32x4_t __s1_624 = __p1_624; \
52842   int32x4_t __s2_624 = __p2_624; \
52843   int64x2_t __rev0_624;  __rev0_624 = __builtin_shufflevector(__s0_624, __s0_624, 1, 0); \
52844   int32x4_t __rev1_624;  __rev1_624 = __builtin_shufflevector(__s1_624, __s1_624, 3, 2, 1, 0); \
52845   int32x4_t __rev2_624;  __rev2_624 = __builtin_shufflevector(__s2_624, __s2_624, 3, 2, 1, 0); \
52846   __ret_624 = __noswap_vqdmlsl_s32(__rev0_624, __noswap_vget_high_s32(__rev1_624), __noswap_splat_laneq_s32(__rev2_624, __p3_624)); \
52847   __ret_624 = __builtin_shufflevector(__ret_624, __ret_624, 1, 0); \
52848   __ret_624; \
52849 })
52850 #endif
52851 
52852 #ifdef __LITTLE_ENDIAN__
52853 #define vqdmlsl_high_laneq_s16(__p0_625, __p1_625, __p2_625, __p3_625) __extension__ ({ \
52854   int32x4_t __ret_625; \
52855   int32x4_t __s0_625 = __p0_625; \
52856   int16x8_t __s1_625 = __p1_625; \
52857   int16x8_t __s2_625 = __p2_625; \
52858   __ret_625 = vqdmlsl_s16(__s0_625, vget_high_s16(__s1_625), splat_laneq_s16(__s2_625, __p3_625)); \
52859   __ret_625; \
52860 })
52861 #else
52862 #define vqdmlsl_high_laneq_s16(__p0_626, __p1_626, __p2_626, __p3_626) __extension__ ({ \
52863   int32x4_t __ret_626; \
52864   int32x4_t __s0_626 = __p0_626; \
52865   int16x8_t __s1_626 = __p1_626; \
52866   int16x8_t __s2_626 = __p2_626; \
52867   int32x4_t __rev0_626;  __rev0_626 = __builtin_shufflevector(__s0_626, __s0_626, 3, 2, 1, 0); \
52868   int16x8_t __rev1_626;  __rev1_626 = __builtin_shufflevector(__s1_626, __s1_626, 7, 6, 5, 4, 3, 2, 1, 0); \
52869   int16x8_t __rev2_626;  __rev2_626 = __builtin_shufflevector(__s2_626, __s2_626, 7, 6, 5, 4, 3, 2, 1, 0); \
52870   __ret_626 = __noswap_vqdmlsl_s16(__rev0_626, __noswap_vget_high_s16(__rev1_626), __noswap_splat_laneq_s16(__rev2_626, __p3_626)); \
52871   __ret_626 = __builtin_shufflevector(__ret_626, __ret_626, 3, 2, 1, 0); \
52872   __ret_626; \
52873 })
52874 #endif
52875 
52876 #ifdef __LITTLE_ENDIAN__
vqdmlsl_high_n_s32(int64x2_t __p0,int32x4_t __p1,int32_t __p2)52877 __ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
52878   int64x2_t __ret;
52879   __ret = vqdmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
52880   return __ret;
52881 }
52882 #else
vqdmlsl_high_n_s32(int64x2_t __p0,int32x4_t __p1,int32_t __p2)52883 __ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
52884   int64x2_t __ret;
52885   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52886   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52887   __ret = __noswap_vqdmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
52888   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52889   return __ret;
52890 }
52891 #endif
52892 
52893 #ifdef __LITTLE_ENDIAN__
vqdmlsl_high_n_s16(int32x4_t __p0,int16x8_t __p1,int16_t __p2)52894 __ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
52895   int32x4_t __ret;
52896   __ret = vqdmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
52897   return __ret;
52898 }
52899 #else
vqdmlsl_high_n_s16(int32x4_t __p0,int16x8_t __p1,int16_t __p2)52900 __ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
52901   int32x4_t __ret;
52902   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52903   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
52904   __ret = __noswap_vqdmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
52905   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52906   return __ret;
52907 }
52908 #endif
52909 
52910 #ifdef __LITTLE_ENDIAN__
52911 #define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52912   int64_t __ret; \
52913   int64_t __s0 = __p0; \
52914   int32_t __s1 = __p1; \
52915   int32x2_t __s2 = __p2; \
52916   __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, __s2, __p3); \
52917   __ret; \
52918 })
52919 #else
52920 #define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52921   int64_t __ret; \
52922   int64_t __s0 = __p0; \
52923   int32_t __s1 = __p1; \
52924   int32x2_t __s2 = __p2; \
52925   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52926   __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, __rev2, __p3); \
52927   __ret; \
52928 })
52929 #endif
52930 
52931 #ifdef __LITTLE_ENDIAN__
52932 #define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52933   int32_t __ret; \
52934   int32_t __s0 = __p0; \
52935   int16_t __s1 = __p1; \
52936   int16x4_t __s2 = __p2; \
52937   __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, __s2, __p3); \
52938   __ret; \
52939 })
52940 #else
52941 #define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52942   int32_t __ret; \
52943   int32_t __s0 = __p0; \
52944   int16_t __s1 = __p1; \
52945   int16x4_t __s2 = __p2; \
52946   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52947   __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, __rev2, __p3); \
52948   __ret; \
52949 })
52950 #endif
52951 
52952 #ifdef __LITTLE_ENDIAN__
52953 #define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52954   int64_t __ret; \
52955   int64_t __s0 = __p0; \
52956   int32_t __s1 = __p1; \
52957   int32x4_t __s2 = __p2; \
52958   __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, __s2, __p3); \
52959   __ret; \
52960 })
52961 #else
52962 #define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52963   int64_t __ret; \
52964   int64_t __s0 = __p0; \
52965   int32_t __s1 = __p1; \
52966   int32x4_t __s2 = __p2; \
52967   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52968   __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, __rev2, __p3); \
52969   __ret; \
52970 })
52971 #endif
52972 
52973 #ifdef __LITTLE_ENDIAN__
52974 #define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52975   int32_t __ret; \
52976   int32_t __s0 = __p0; \
52977   int16_t __s1 = __p1; \
52978   int16x8_t __s2 = __p2; \
52979   __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, __s2, __p3); \
52980   __ret; \
52981 })
52982 #else
52983 #define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52984   int32_t __ret; \
52985   int32_t __s0 = __p0; \
52986   int16_t __s1 = __p1; \
52987   int16x8_t __s2 = __p2; \
52988   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52989   __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, __rev2, __p3); \
52990   __ret; \
52991 })
52992 #endif
52993 
52994 #ifdef __LITTLE_ENDIAN__
52995 #define vqdmlsl_laneq_s32(__p0_627, __p1_627, __p2_627, __p3_627) __extension__ ({ \
52996   int64x2_t __ret_627; \
52997   int64x2_t __s0_627 = __p0_627; \
52998   int32x2_t __s1_627 = __p1_627; \
52999   int32x4_t __s2_627 = __p2_627; \
53000   __ret_627 = vqdmlsl_s32(__s0_627, __s1_627, splat_laneq_s32(__s2_627, __p3_627)); \
53001   __ret_627; \
53002 })
53003 #else
53004 #define vqdmlsl_laneq_s32(__p0_628, __p1_628, __p2_628, __p3_628) __extension__ ({ \
53005   int64x2_t __ret_628; \
53006   int64x2_t __s0_628 = __p0_628; \
53007   int32x2_t __s1_628 = __p1_628; \
53008   int32x4_t __s2_628 = __p2_628; \
53009   int64x2_t __rev0_628;  __rev0_628 = __builtin_shufflevector(__s0_628, __s0_628, 1, 0); \
53010   int32x2_t __rev1_628;  __rev1_628 = __builtin_shufflevector(__s1_628, __s1_628, 1, 0); \
53011   int32x4_t __rev2_628;  __rev2_628 = __builtin_shufflevector(__s2_628, __s2_628, 3, 2, 1, 0); \
53012   __ret_628 = __noswap_vqdmlsl_s32(__rev0_628, __rev1_628, __noswap_splat_laneq_s32(__rev2_628, __p3_628)); \
53013   __ret_628 = __builtin_shufflevector(__ret_628, __ret_628, 1, 0); \
53014   __ret_628; \
53015 })
53016 #endif
53017 
53018 #ifdef __LITTLE_ENDIAN__
53019 #define vqdmlsl_laneq_s16(__p0_629, __p1_629, __p2_629, __p3_629) __extension__ ({ \
53020   int32x4_t __ret_629; \
53021   int32x4_t __s0_629 = __p0_629; \
53022   int16x4_t __s1_629 = __p1_629; \
53023   int16x8_t __s2_629 = __p2_629; \
53024   __ret_629 = vqdmlsl_s16(__s0_629, __s1_629, splat_laneq_s16(__s2_629, __p3_629)); \
53025   __ret_629; \
53026 })
53027 #else
53028 #define vqdmlsl_laneq_s16(__p0_630, __p1_630, __p2_630, __p3_630) __extension__ ({ \
53029   int32x4_t __ret_630; \
53030   int32x4_t __s0_630 = __p0_630; \
53031   int16x4_t __s1_630 = __p1_630; \
53032   int16x8_t __s2_630 = __p2_630; \
53033   int32x4_t __rev0_630;  __rev0_630 = __builtin_shufflevector(__s0_630, __s0_630, 3, 2, 1, 0); \
53034   int16x4_t __rev1_630;  __rev1_630 = __builtin_shufflevector(__s1_630, __s1_630, 3, 2, 1, 0); \
53035   int16x8_t __rev2_630;  __rev2_630 = __builtin_shufflevector(__s2_630, __s2_630, 7, 6, 5, 4, 3, 2, 1, 0); \
53036   __ret_630 = __noswap_vqdmlsl_s16(__rev0_630, __rev1_630, __noswap_splat_laneq_s16(__rev2_630, __p3_630)); \
53037   __ret_630 = __builtin_shufflevector(__ret_630, __ret_630, 3, 2, 1, 0); \
53038   __ret_630; \
53039 })
53040 #endif
53041 
vqdmulhs_s32(int32_t __p0,int32_t __p1)53042 __ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
53043   int32_t __ret;
53044   __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
53045   return __ret;
53046 }
vqdmulhh_s16(int16_t __p0,int16_t __p1)53047 __ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
53048   int16_t __ret;
53049   __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
53050   return __ret;
53051 }
53052 #ifdef __LITTLE_ENDIAN__
53053 #define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53054   int32x4_t __ret; \
53055   int32x4_t __s0 = __p0; \
53056   int32x2_t __s1 = __p1; \
53057   __ret = (int32x4_t) __builtin_neon_vqdmulhq_lane_v((int8x16_t)__s0, (int8x8_t)__s1, __p2, 2); \
53058   __ret; \
53059 })
53060 #else
53061 #define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53062   int32x4_t __ret; \
53063   int32x4_t __s0 = __p0; \
53064   int32x2_t __s1 = __p1; \
53065   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53066   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53067   __ret = (int32x4_t) __builtin_neon_vqdmulhq_lane_v((int8x16_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
53068   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53069   __ret; \
53070 })
53071 #endif
53072 
53073 #ifdef __LITTLE_ENDIAN__
53074 #define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53075   int16x8_t __ret; \
53076   int16x8_t __s0 = __p0; \
53077   int16x4_t __s1 = __p1; \
53078   __ret = (int16x8_t) __builtin_neon_vqdmulhq_lane_v((int8x16_t)__s0, (int8x8_t)__s1, __p2, 1); \
53079   __ret; \
53080 })
53081 #else
53082 #define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53083   int16x8_t __ret; \
53084   int16x8_t __s0 = __p0; \
53085   int16x4_t __s1 = __p1; \
53086   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53087   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53088   __ret = (int16x8_t) __builtin_neon_vqdmulhq_lane_v((int8x16_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
53089   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53090   __ret; \
53091 })
53092 #endif
53093 
53094 #ifdef __LITTLE_ENDIAN__
53095 #define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53096   int32x2_t __ret; \
53097   int32x2_t __s0 = __p0; \
53098   int32x2_t __s1 = __p1; \
53099   __ret = (int32x2_t) __builtin_neon_vqdmulh_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
53100   __ret; \
53101 })
53102 #else
53103 #define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53104   int32x2_t __ret; \
53105   int32x2_t __s0 = __p0; \
53106   int32x2_t __s1 = __p1; \
53107   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53108   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53109   __ret = (int32x2_t) __builtin_neon_vqdmulh_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
53110   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53111   __ret; \
53112 })
53113 #endif
53114 
53115 #ifdef __LITTLE_ENDIAN__
53116 #define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53117   int16x4_t __ret; \
53118   int16x4_t __s0 = __p0; \
53119   int16x4_t __s1 = __p1; \
53120   __ret = (int16x4_t) __builtin_neon_vqdmulh_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
53121   __ret; \
53122 })
53123 #else
53124 #define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53125   int16x4_t __ret; \
53126   int16x4_t __s0 = __p0; \
53127   int16x4_t __s1 = __p1; \
53128   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53129   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53130   __ret = (int16x4_t) __builtin_neon_vqdmulh_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
53131   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53132   __ret; \
53133 })
53134 #endif
53135 
53136 #ifdef __LITTLE_ENDIAN__
53137 #define vqdmulhs_lane_s32(__p0_631, __p1_631, __p2_631) __extension__ ({ \
53138   int32_t __ret_631; \
53139   int32_t __s0_631 = __p0_631; \
53140   int32x2_t __s1_631 = __p1_631; \
53141   __ret_631 = vqdmulhs_s32(__s0_631, vget_lane_s32(__s1_631, __p2_631)); \
53142   __ret_631; \
53143 })
53144 #else
53145 #define vqdmulhs_lane_s32(__p0_632, __p1_632, __p2_632) __extension__ ({ \
53146   int32_t __ret_632; \
53147   int32_t __s0_632 = __p0_632; \
53148   int32x2_t __s1_632 = __p1_632; \
53149   int32x2_t __rev1_632;  __rev1_632 = __builtin_shufflevector(__s1_632, __s1_632, 1, 0); \
53150   __ret_632 = vqdmulhs_s32(__s0_632, __noswap_vget_lane_s32(__rev1_632, __p2_632)); \
53151   __ret_632; \
53152 })
53153 #endif
53154 
53155 #ifdef __LITTLE_ENDIAN__
53156 #define vqdmulhh_lane_s16(__p0_633, __p1_633, __p2_633) __extension__ ({ \
53157   int16_t __ret_633; \
53158   int16_t __s0_633 = __p0_633; \
53159   int16x4_t __s1_633 = __p1_633; \
53160   __ret_633 = vqdmulhh_s16(__s0_633, vget_lane_s16(__s1_633, __p2_633)); \
53161   __ret_633; \
53162 })
53163 #else
53164 #define vqdmulhh_lane_s16(__p0_634, __p1_634, __p2_634) __extension__ ({ \
53165   int16_t __ret_634; \
53166   int16_t __s0_634 = __p0_634; \
53167   int16x4_t __s1_634 = __p1_634; \
53168   int16x4_t __rev1_634;  __rev1_634 = __builtin_shufflevector(__s1_634, __s1_634, 3, 2, 1, 0); \
53169   __ret_634 = vqdmulhh_s16(__s0_634, __noswap_vget_lane_s16(__rev1_634, __p2_634)); \
53170   __ret_634; \
53171 })
53172 #endif
53173 
53174 #ifdef __LITTLE_ENDIAN__
53175 #define vqdmulhs_laneq_s32(__p0_635, __p1_635, __p2_635) __extension__ ({ \
53176   int32_t __ret_635; \
53177   int32_t __s0_635 = __p0_635; \
53178   int32x4_t __s1_635 = __p1_635; \
53179   __ret_635 = vqdmulhs_s32(__s0_635, vgetq_lane_s32(__s1_635, __p2_635)); \
53180   __ret_635; \
53181 })
53182 #else
53183 #define vqdmulhs_laneq_s32(__p0_636, __p1_636, __p2_636) __extension__ ({ \
53184   int32_t __ret_636; \
53185   int32_t __s0_636 = __p0_636; \
53186   int32x4_t __s1_636 = __p1_636; \
53187   int32x4_t __rev1_636;  __rev1_636 = __builtin_shufflevector(__s1_636, __s1_636, 3, 2, 1, 0); \
53188   __ret_636 = vqdmulhs_s32(__s0_636, __noswap_vgetq_lane_s32(__rev1_636, __p2_636)); \
53189   __ret_636; \
53190 })
53191 #endif
53192 
53193 #ifdef __LITTLE_ENDIAN__
53194 #define vqdmulhh_laneq_s16(__p0_637, __p1_637, __p2_637) __extension__ ({ \
53195   int16_t __ret_637; \
53196   int16_t __s0_637 = __p0_637; \
53197   int16x8_t __s1_637 = __p1_637; \
53198   __ret_637 = vqdmulhh_s16(__s0_637, vgetq_lane_s16(__s1_637, __p2_637)); \
53199   __ret_637; \
53200 })
53201 #else
53202 #define vqdmulhh_laneq_s16(__p0_638, __p1_638, __p2_638) __extension__ ({ \
53203   int16_t __ret_638; \
53204   int16_t __s0_638 = __p0_638; \
53205   int16x8_t __s1_638 = __p1_638; \
53206   int16x8_t __rev1_638;  __rev1_638 = __builtin_shufflevector(__s1_638, __s1_638, 7, 6, 5, 4, 3, 2, 1, 0); \
53207   __ret_638 = vqdmulhh_s16(__s0_638, __noswap_vgetq_lane_s16(__rev1_638, __p2_638)); \
53208   __ret_638; \
53209 })
53210 #endif
53211 
53212 #ifdef __LITTLE_ENDIAN__
53213 #define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53214   int32x4_t __ret; \
53215   int32x4_t __s0 = __p0; \
53216   int32x4_t __s1 = __p1; \
53217   __ret = (int32x4_t) __builtin_neon_vqdmulhq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
53218   __ret; \
53219 })
53220 #else
53221 #define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53222   int32x4_t __ret; \
53223   int32x4_t __s0 = __p0; \
53224   int32x4_t __s1 = __p1; \
53225   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53226   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53227   __ret = (int32x4_t) __builtin_neon_vqdmulhq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
53228   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53229   __ret; \
53230 })
53231 #endif
53232 
53233 #ifdef __LITTLE_ENDIAN__
53234 #define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53235   int16x8_t __ret; \
53236   int16x8_t __s0 = __p0; \
53237   int16x8_t __s1 = __p1; \
53238   __ret = (int16x8_t) __builtin_neon_vqdmulhq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
53239   __ret; \
53240 })
53241 #else
53242 #define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53243   int16x8_t __ret; \
53244   int16x8_t __s0 = __p0; \
53245   int16x8_t __s1 = __p1; \
53246   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53247   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53248   __ret = (int16x8_t) __builtin_neon_vqdmulhq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
53249   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53250   __ret; \
53251 })
53252 #endif
53253 
53254 #ifdef __LITTLE_ENDIAN__
53255 #define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53256   int32x2_t __ret; \
53257   int32x2_t __s0 = __p0; \
53258   int32x4_t __s1 = __p1; \
53259   __ret = (int32x2_t) __builtin_neon_vqdmulh_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 2); \
53260   __ret; \
53261 })
53262 #else
53263 #define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53264   int32x2_t __ret; \
53265   int32x2_t __s0 = __p0; \
53266   int32x4_t __s1 = __p1; \
53267   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53268   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53269   __ret = (int32x2_t) __builtin_neon_vqdmulh_laneq_v((int8x8_t)__rev0, (int8x16_t)__rev1, __p2, 2); \
53270   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53271   __ret; \
53272 })
53273 #endif
53274 
53275 #ifdef __LITTLE_ENDIAN__
53276 #define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53277   int16x4_t __ret; \
53278   int16x4_t __s0 = __p0; \
53279   int16x8_t __s1 = __p1; \
53280   __ret = (int16x4_t) __builtin_neon_vqdmulh_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 1); \
53281   __ret; \
53282 })
53283 #else
53284 #define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53285   int16x4_t __ret; \
53286   int16x4_t __s0 = __p0; \
53287   int16x8_t __s1 = __p1; \
53288   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53289   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53290   __ret = (int16x4_t) __builtin_neon_vqdmulh_laneq_v((int8x8_t)__rev0, (int8x16_t)__rev1, __p2, 1); \
53291   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53292   __ret; \
53293 })
53294 #endif
53295 
vqdmulls_s32(int32_t __p0,int32_t __p1)53296 __ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
53297   int64_t __ret;
53298   __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
53299   return __ret;
53300 }
vqdmullh_s16(int16_t __p0,int16_t __p1)53301 __ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
53302   int32_t __ret;
53303   __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
53304   return __ret;
53305 }
53306 #ifdef __LITTLE_ENDIAN__
vqdmull_high_s32(int32x4_t __p0,int32x4_t __p1)53307 __ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
53308   int64x2_t __ret;
53309   __ret = vqdmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
53310   return __ret;
53311 }
53312 #else
vqdmull_high_s32(int32x4_t __p0,int32x4_t __p1)53313 __ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
53314   int64x2_t __ret;
53315   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53316   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53317   __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
53318   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53319   return __ret;
53320 }
53321 #endif
53322 
53323 #ifdef __LITTLE_ENDIAN__
vqdmull_high_s16(int16x8_t __p0,int16x8_t __p1)53324 __ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
53325   int32x4_t __ret;
53326   __ret = vqdmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
53327   return __ret;
53328 }
53329 #else
vqdmull_high_s16(int16x8_t __p0,int16x8_t __p1)53330 __ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
53331   int32x4_t __ret;
53332   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53333   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53334   __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
53335   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53336   return __ret;
53337 }
53338 #endif
53339 
53340 #ifdef __LITTLE_ENDIAN__
53341 #define vqdmull_high_lane_s32(__p0_639, __p1_639, __p2_639) __extension__ ({ \
53342   int64x2_t __ret_639; \
53343   int32x4_t __s0_639 = __p0_639; \
53344   int32x2_t __s1_639 = __p1_639; \
53345   __ret_639 = vqdmull_s32(vget_high_s32(__s0_639), splat_lane_s32(__s1_639, __p2_639)); \
53346   __ret_639; \
53347 })
53348 #else
53349 #define vqdmull_high_lane_s32(__p0_640, __p1_640, __p2_640) __extension__ ({ \
53350   int64x2_t __ret_640; \
53351   int32x4_t __s0_640 = __p0_640; \
53352   int32x2_t __s1_640 = __p1_640; \
53353   int32x4_t __rev0_640;  __rev0_640 = __builtin_shufflevector(__s0_640, __s0_640, 3, 2, 1, 0); \
53354   int32x2_t __rev1_640;  __rev1_640 = __builtin_shufflevector(__s1_640, __s1_640, 1, 0); \
53355   __ret_640 = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0_640), __noswap_splat_lane_s32(__rev1_640, __p2_640)); \
53356   __ret_640 = __builtin_shufflevector(__ret_640, __ret_640, 1, 0); \
53357   __ret_640; \
53358 })
53359 #endif
53360 
53361 #ifdef __LITTLE_ENDIAN__
53362 #define vqdmull_high_lane_s16(__p0_641, __p1_641, __p2_641) __extension__ ({ \
53363   int32x4_t __ret_641; \
53364   int16x8_t __s0_641 = __p0_641; \
53365   int16x4_t __s1_641 = __p1_641; \
53366   __ret_641 = vqdmull_s16(vget_high_s16(__s0_641), splat_lane_s16(__s1_641, __p2_641)); \
53367   __ret_641; \
53368 })
53369 #else
53370 #define vqdmull_high_lane_s16(__p0_642, __p1_642, __p2_642) __extension__ ({ \
53371   int32x4_t __ret_642; \
53372   int16x8_t __s0_642 = __p0_642; \
53373   int16x4_t __s1_642 = __p1_642; \
53374   int16x8_t __rev0_642;  __rev0_642 = __builtin_shufflevector(__s0_642, __s0_642, 7, 6, 5, 4, 3, 2, 1, 0); \
53375   int16x4_t __rev1_642;  __rev1_642 = __builtin_shufflevector(__s1_642, __s1_642, 3, 2, 1, 0); \
53376   __ret_642 = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0_642), __noswap_splat_lane_s16(__rev1_642, __p2_642)); \
53377   __ret_642 = __builtin_shufflevector(__ret_642, __ret_642, 3, 2, 1, 0); \
53378   __ret_642; \
53379 })
53380 #endif
53381 
53382 #ifdef __LITTLE_ENDIAN__
53383 #define vqdmull_high_laneq_s32(__p0_643, __p1_643, __p2_643) __extension__ ({ \
53384   int64x2_t __ret_643; \
53385   int32x4_t __s0_643 = __p0_643; \
53386   int32x4_t __s1_643 = __p1_643; \
53387   __ret_643 = vqdmull_s32(vget_high_s32(__s0_643), splat_laneq_s32(__s1_643, __p2_643)); \
53388   __ret_643; \
53389 })
53390 #else
53391 #define vqdmull_high_laneq_s32(__p0_644, __p1_644, __p2_644) __extension__ ({ \
53392   int64x2_t __ret_644; \
53393   int32x4_t __s0_644 = __p0_644; \
53394   int32x4_t __s1_644 = __p1_644; \
53395   int32x4_t __rev0_644;  __rev0_644 = __builtin_shufflevector(__s0_644, __s0_644, 3, 2, 1, 0); \
53396   int32x4_t __rev1_644;  __rev1_644 = __builtin_shufflevector(__s1_644, __s1_644, 3, 2, 1, 0); \
53397   __ret_644 = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0_644), __noswap_splat_laneq_s32(__rev1_644, __p2_644)); \
53398   __ret_644 = __builtin_shufflevector(__ret_644, __ret_644, 1, 0); \
53399   __ret_644; \
53400 })
53401 #endif
53402 
53403 #ifdef __LITTLE_ENDIAN__
53404 #define vqdmull_high_laneq_s16(__p0_645, __p1_645, __p2_645) __extension__ ({ \
53405   int32x4_t __ret_645; \
53406   int16x8_t __s0_645 = __p0_645; \
53407   int16x8_t __s1_645 = __p1_645; \
53408   __ret_645 = vqdmull_s16(vget_high_s16(__s0_645), splat_laneq_s16(__s1_645, __p2_645)); \
53409   __ret_645; \
53410 })
53411 #else
53412 #define vqdmull_high_laneq_s16(__p0_646, __p1_646, __p2_646) __extension__ ({ \
53413   int32x4_t __ret_646; \
53414   int16x8_t __s0_646 = __p0_646; \
53415   int16x8_t __s1_646 = __p1_646; \
53416   int16x8_t __rev0_646;  __rev0_646 = __builtin_shufflevector(__s0_646, __s0_646, 7, 6, 5, 4, 3, 2, 1, 0); \
53417   int16x8_t __rev1_646;  __rev1_646 = __builtin_shufflevector(__s1_646, __s1_646, 7, 6, 5, 4, 3, 2, 1, 0); \
53418   __ret_646 = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0_646), __noswap_splat_laneq_s16(__rev1_646, __p2_646)); \
53419   __ret_646 = __builtin_shufflevector(__ret_646, __ret_646, 3, 2, 1, 0); \
53420   __ret_646; \
53421 })
53422 #endif
53423 
53424 #ifdef __LITTLE_ENDIAN__
vqdmull_high_n_s32(int32x4_t __p0,int32_t __p1)53425 __ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
53426   int64x2_t __ret;
53427   __ret = vqdmull_n_s32(vget_high_s32(__p0), __p1);
53428   return __ret;
53429 }
53430 #else
vqdmull_high_n_s32(int32x4_t __p0,int32_t __p1)53431 __ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
53432   int64x2_t __ret;
53433   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53434   __ret = __noswap_vqdmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
53435   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53436   return __ret;
53437 }
53438 #endif
53439 
53440 #ifdef __LITTLE_ENDIAN__
vqdmull_high_n_s16(int16x8_t __p0,int16_t __p1)53441 __ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
53442   int32x4_t __ret;
53443   __ret = vqdmull_n_s16(vget_high_s16(__p0), __p1);
53444   return __ret;
53445 }
53446 #else
vqdmull_high_n_s16(int16x8_t __p0,int16_t __p1)53447 __ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
53448   int32x4_t __ret;
53449   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53450   __ret = __noswap_vqdmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
53451   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53452   return __ret;
53453 }
53454 #endif
53455 
53456 #ifdef __LITTLE_ENDIAN__
53457 #define vqdmulls_lane_s32(__p0_647, __p1_647, __p2_647) __extension__ ({ \
53458   int64_t __ret_647; \
53459   int32_t __s0_647 = __p0_647; \
53460   int32x2_t __s1_647 = __p1_647; \
53461   __ret_647 = vqdmulls_s32(__s0_647, vget_lane_s32(__s1_647, __p2_647)); \
53462   __ret_647; \
53463 })
53464 #else
53465 #define vqdmulls_lane_s32(__p0_648, __p1_648, __p2_648) __extension__ ({ \
53466   int64_t __ret_648; \
53467   int32_t __s0_648 = __p0_648; \
53468   int32x2_t __s1_648 = __p1_648; \
53469   int32x2_t __rev1_648;  __rev1_648 = __builtin_shufflevector(__s1_648, __s1_648, 1, 0); \
53470   __ret_648 = vqdmulls_s32(__s0_648, __noswap_vget_lane_s32(__rev1_648, __p2_648)); \
53471   __ret_648; \
53472 })
53473 #endif
53474 
53475 #ifdef __LITTLE_ENDIAN__
53476 #define vqdmullh_lane_s16(__p0_649, __p1_649, __p2_649) __extension__ ({ \
53477   int32_t __ret_649; \
53478   int16_t __s0_649 = __p0_649; \
53479   int16x4_t __s1_649 = __p1_649; \
53480   __ret_649 = vqdmullh_s16(__s0_649, vget_lane_s16(__s1_649, __p2_649)); \
53481   __ret_649; \
53482 })
53483 #else
53484 #define vqdmullh_lane_s16(__p0_650, __p1_650, __p2_650) __extension__ ({ \
53485   int32_t __ret_650; \
53486   int16_t __s0_650 = __p0_650; \
53487   int16x4_t __s1_650 = __p1_650; \
53488   int16x4_t __rev1_650;  __rev1_650 = __builtin_shufflevector(__s1_650, __s1_650, 3, 2, 1, 0); \
53489   __ret_650 = vqdmullh_s16(__s0_650, __noswap_vget_lane_s16(__rev1_650, __p2_650)); \
53490   __ret_650; \
53491 })
53492 #endif
53493 
53494 #ifdef __LITTLE_ENDIAN__
53495 #define vqdmulls_laneq_s32(__p0_651, __p1_651, __p2_651) __extension__ ({ \
53496   int64_t __ret_651; \
53497   int32_t __s0_651 = __p0_651; \
53498   int32x4_t __s1_651 = __p1_651; \
53499   __ret_651 = vqdmulls_s32(__s0_651, vgetq_lane_s32(__s1_651, __p2_651)); \
53500   __ret_651; \
53501 })
53502 #else
53503 #define vqdmulls_laneq_s32(__p0_652, __p1_652, __p2_652) __extension__ ({ \
53504   int64_t __ret_652; \
53505   int32_t __s0_652 = __p0_652; \
53506   int32x4_t __s1_652 = __p1_652; \
53507   int32x4_t __rev1_652;  __rev1_652 = __builtin_shufflevector(__s1_652, __s1_652, 3, 2, 1, 0); \
53508   __ret_652 = vqdmulls_s32(__s0_652, __noswap_vgetq_lane_s32(__rev1_652, __p2_652)); \
53509   __ret_652; \
53510 })
53511 #endif
53512 
53513 #ifdef __LITTLE_ENDIAN__
53514 #define vqdmullh_laneq_s16(__p0_653, __p1_653, __p2_653) __extension__ ({ \
53515   int32_t __ret_653; \
53516   int16_t __s0_653 = __p0_653; \
53517   int16x8_t __s1_653 = __p1_653; \
53518   __ret_653 = vqdmullh_s16(__s0_653, vgetq_lane_s16(__s1_653, __p2_653)); \
53519   __ret_653; \
53520 })
53521 #else
53522 #define vqdmullh_laneq_s16(__p0_654, __p1_654, __p2_654) __extension__ ({ \
53523   int32_t __ret_654; \
53524   int16_t __s0_654 = __p0_654; \
53525   int16x8_t __s1_654 = __p1_654; \
53526   int16x8_t __rev1_654;  __rev1_654 = __builtin_shufflevector(__s1_654, __s1_654, 7, 6, 5, 4, 3, 2, 1, 0); \
53527   __ret_654 = vqdmullh_s16(__s0_654, __noswap_vgetq_lane_s16(__rev1_654, __p2_654)); \
53528   __ret_654; \
53529 })
53530 #endif
53531 
53532 #ifdef __LITTLE_ENDIAN__
53533 #define vqdmull_laneq_s32(__p0_655, __p1_655, __p2_655) __extension__ ({ \
53534   int64x2_t __ret_655; \
53535   int32x2_t __s0_655 = __p0_655; \
53536   int32x4_t __s1_655 = __p1_655; \
53537   __ret_655 = vqdmull_s32(__s0_655, splat_laneq_s32(__s1_655, __p2_655)); \
53538   __ret_655; \
53539 })
53540 #else
53541 #define vqdmull_laneq_s32(__p0_656, __p1_656, __p2_656) __extension__ ({ \
53542   int64x2_t __ret_656; \
53543   int32x2_t __s0_656 = __p0_656; \
53544   int32x4_t __s1_656 = __p1_656; \
53545   int32x2_t __rev0_656;  __rev0_656 = __builtin_shufflevector(__s0_656, __s0_656, 1, 0); \
53546   int32x4_t __rev1_656;  __rev1_656 = __builtin_shufflevector(__s1_656, __s1_656, 3, 2, 1, 0); \
53547   __ret_656 = __noswap_vqdmull_s32(__rev0_656, __noswap_splat_laneq_s32(__rev1_656, __p2_656)); \
53548   __ret_656 = __builtin_shufflevector(__ret_656, __ret_656, 1, 0); \
53549   __ret_656; \
53550 })
53551 #endif
53552 
53553 #ifdef __LITTLE_ENDIAN__
53554 #define vqdmull_laneq_s16(__p0_657, __p1_657, __p2_657) __extension__ ({ \
53555   int32x4_t __ret_657; \
53556   int16x4_t __s0_657 = __p0_657; \
53557   int16x8_t __s1_657 = __p1_657; \
53558   __ret_657 = vqdmull_s16(__s0_657, splat_laneq_s16(__s1_657, __p2_657)); \
53559   __ret_657; \
53560 })
53561 #else
53562 #define vqdmull_laneq_s16(__p0_658, __p1_658, __p2_658) __extension__ ({ \
53563   int32x4_t __ret_658; \
53564   int16x4_t __s0_658 = __p0_658; \
53565   int16x8_t __s1_658 = __p1_658; \
53566   int16x4_t __rev0_658;  __rev0_658 = __builtin_shufflevector(__s0_658, __s0_658, 3, 2, 1, 0); \
53567   int16x8_t __rev1_658;  __rev1_658 = __builtin_shufflevector(__s1_658, __s1_658, 7, 6, 5, 4, 3, 2, 1, 0); \
53568   __ret_658 = __noswap_vqdmull_s16(__rev0_658, __noswap_splat_laneq_s16(__rev1_658, __p2_658)); \
53569   __ret_658 = __builtin_shufflevector(__ret_658, __ret_658, 3, 2, 1, 0); \
53570   __ret_658; \
53571 })
53572 #endif
53573 
vqmovns_s32(int32_t __p0)53574 __ai int16_t vqmovns_s32(int32_t __p0) {
53575   int16_t __ret;
53576   __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
53577   return __ret;
53578 }
vqmovnd_s64(int64_t __p0)53579 __ai int32_t vqmovnd_s64(int64_t __p0) {
53580   int32_t __ret;
53581   __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
53582   return __ret;
53583 }
vqmovnh_s16(int16_t __p0)53584 __ai int8_t vqmovnh_s16(int16_t __p0) {
53585   int8_t __ret;
53586   __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
53587   return __ret;
53588 }
vqmovns_u32(uint32_t __p0)53589 __ai uint16_t vqmovns_u32(uint32_t __p0) {
53590   uint16_t __ret;
53591   __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
53592   return __ret;
53593 }
vqmovnd_u64(uint64_t __p0)53594 __ai uint32_t vqmovnd_u64(uint64_t __p0) {
53595   uint32_t __ret;
53596   __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
53597   return __ret;
53598 }
vqmovnh_u16(uint16_t __p0)53599 __ai uint8_t vqmovnh_u16(uint16_t __p0) {
53600   uint8_t __ret;
53601   __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
53602   return __ret;
53603 }
53604 #ifdef __LITTLE_ENDIAN__
vqmovn_high_u32(uint16x4_t __p0,uint32x4_t __p1)53605 __ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53606   uint16x8_t __ret;
53607   __ret = vcombine_u16(__p0, vqmovn_u32(__p1));
53608   return __ret;
53609 }
53610 #else
vqmovn_high_u32(uint16x4_t __p0,uint32x4_t __p1)53611 __ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53612   uint16x8_t __ret;
53613   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53614   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53615   __ret = __noswap_vcombine_u16(__rev0, __noswap_vqmovn_u32(__rev1));
53616   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53617   return __ret;
53618 }
53619 #endif
53620 
53621 #ifdef __LITTLE_ENDIAN__
vqmovn_high_u64(uint32x2_t __p0,uint64x2_t __p1)53622 __ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53623   uint32x4_t __ret;
53624   __ret = vcombine_u32(__p0, vqmovn_u64(__p1));
53625   return __ret;
53626 }
53627 #else
vqmovn_high_u64(uint32x2_t __p0,uint64x2_t __p1)53628 __ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53629   uint32x4_t __ret;
53630   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53631   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53632   __ret = __noswap_vcombine_u32(__rev0, __noswap_vqmovn_u64(__rev1));
53633   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53634   return __ret;
53635 }
53636 #endif
53637 
53638 #ifdef __LITTLE_ENDIAN__
vqmovn_high_u16(uint8x8_t __p0,uint16x8_t __p1)53639 __ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53640   uint8x16_t __ret;
53641   __ret = vcombine_u8(__p0, vqmovn_u16(__p1));
53642   return __ret;
53643 }
53644 #else
vqmovn_high_u16(uint8x8_t __p0,uint16x8_t __p1)53645 __ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53646   uint8x16_t __ret;
53647   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53648   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53649   __ret = __noswap_vcombine_u8(__rev0, __noswap_vqmovn_u16(__rev1));
53650   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53651   return __ret;
53652 }
53653 #endif
53654 
53655 #ifdef __LITTLE_ENDIAN__
vqmovn_high_s32(int16x4_t __p0,int32x4_t __p1)53656 __ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53657   int16x8_t __ret;
53658   __ret = vcombine_s16(__p0, vqmovn_s32(__p1));
53659   return __ret;
53660 }
53661 #else
vqmovn_high_s32(int16x4_t __p0,int32x4_t __p1)53662 __ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53663   int16x8_t __ret;
53664   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53665   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53666   __ret = __noswap_vcombine_s16(__rev0, __noswap_vqmovn_s32(__rev1));
53667   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53668   return __ret;
53669 }
53670 #endif
53671 
53672 #ifdef __LITTLE_ENDIAN__
vqmovn_high_s64(int32x2_t __p0,int64x2_t __p1)53673 __ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53674   int32x4_t __ret;
53675   __ret = vcombine_s32(__p0, vqmovn_s64(__p1));
53676   return __ret;
53677 }
53678 #else
vqmovn_high_s64(int32x2_t __p0,int64x2_t __p1)53679 __ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53680   int32x4_t __ret;
53681   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53682   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53683   __ret = __noswap_vcombine_s32(__rev0, __noswap_vqmovn_s64(__rev1));
53684   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53685   return __ret;
53686 }
53687 #endif
53688 
53689 #ifdef __LITTLE_ENDIAN__
vqmovn_high_s16(int8x8_t __p0,int16x8_t __p1)53690 __ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53691   int8x16_t __ret;
53692   __ret = vcombine_s8(__p0, vqmovn_s16(__p1));
53693   return __ret;
53694 }
53695 #else
vqmovn_high_s16(int8x8_t __p0,int16x8_t __p1)53696 __ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53697   int8x16_t __ret;
53698   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53699   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53700   __ret = __noswap_vcombine_s8(__rev0, __noswap_vqmovn_s16(__rev1));
53701   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53702   return __ret;
53703 }
53704 #endif
53705 
vqmovuns_s32(int32_t __p0)53706 __ai uint16_t vqmovuns_s32(int32_t __p0) {
53707   uint16_t __ret;
53708   __ret = (uint16_t) __builtin_neon_vqmovuns_s32(__p0);
53709   return __ret;
53710 }
vqmovund_s64(int64_t __p0)53711 __ai uint32_t vqmovund_s64(int64_t __p0) {
53712   uint32_t __ret;
53713   __ret = (uint32_t) __builtin_neon_vqmovund_s64(__p0);
53714   return __ret;
53715 }
vqmovunh_s16(int16_t __p0)53716 __ai uint8_t vqmovunh_s16(int16_t __p0) {
53717   uint8_t __ret;
53718   __ret = (uint8_t) __builtin_neon_vqmovunh_s16(__p0);
53719   return __ret;
53720 }
53721 #ifdef __LITTLE_ENDIAN__
vqmovun_high_s32(uint16x4_t __p0,int32x4_t __p1)53722 __ai uint16x8_t vqmovun_high_s32(uint16x4_t __p0, int32x4_t __p1) {
53723   uint16x8_t __ret;
53724   __ret = vcombine_u16((uint16x4_t)(__p0), vqmovun_s32(__p1));
53725   return __ret;
53726 }
53727 #else
vqmovun_high_s32(uint16x4_t __p0,int32x4_t __p1)53728 __ai uint16x8_t vqmovun_high_s32(uint16x4_t __p0, int32x4_t __p1) {
53729   uint16x8_t __ret;
53730   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53731   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53732   __ret = __noswap_vcombine_u16((uint16x4_t)(__rev0), __noswap_vqmovun_s32(__rev1));
53733   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53734   return __ret;
53735 }
53736 #endif
53737 
53738 #ifdef __LITTLE_ENDIAN__
vqmovun_high_s64(uint32x2_t __p0,int64x2_t __p1)53739 __ai uint32x4_t vqmovun_high_s64(uint32x2_t __p0, int64x2_t __p1) {
53740   uint32x4_t __ret;
53741   __ret = vcombine_u32((uint32x2_t)(__p0), vqmovun_s64(__p1));
53742   return __ret;
53743 }
53744 #else
vqmovun_high_s64(uint32x2_t __p0,int64x2_t __p1)53745 __ai uint32x4_t vqmovun_high_s64(uint32x2_t __p0, int64x2_t __p1) {
53746   uint32x4_t __ret;
53747   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53748   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53749   __ret = __noswap_vcombine_u32((uint32x2_t)(__rev0), __noswap_vqmovun_s64(__rev1));
53750   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53751   return __ret;
53752 }
53753 #endif
53754 
53755 #ifdef __LITTLE_ENDIAN__
vqmovun_high_s16(uint8x8_t __p0,int16x8_t __p1)53756 __ai uint8x16_t vqmovun_high_s16(uint8x8_t __p0, int16x8_t __p1) {
53757   uint8x16_t __ret;
53758   __ret = vcombine_u8((uint8x8_t)(__p0), vqmovun_s16(__p1));
53759   return __ret;
53760 }
53761 #else
vqmovun_high_s16(uint8x8_t __p0,int16x8_t __p1)53762 __ai uint8x16_t vqmovun_high_s16(uint8x8_t __p0, int16x8_t __p1) {
53763   uint8x16_t __ret;
53764   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53765   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53766   __ret = __noswap_vcombine_u8((uint8x8_t)(__rev0), __noswap_vqmovun_s16(__rev1));
53767   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53768   return __ret;
53769 }
53770 #endif
53771 
53772 #ifdef __LITTLE_ENDIAN__
vqnegq_s64(int64x2_t __p0)53773 __ai int64x2_t vqnegq_s64(int64x2_t __p0) {
53774   int64x2_t __ret;
53775   __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 35);
53776   return __ret;
53777 }
53778 #else
vqnegq_s64(int64x2_t __p0)53779 __ai int64x2_t vqnegq_s64(int64x2_t __p0) {
53780   int64x2_t __ret;
53781   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53782   __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 35);
53783   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53784   return __ret;
53785 }
53786 #endif
53787 
vqneg_s64(int64x1_t __p0)53788 __ai int64x1_t vqneg_s64(int64x1_t __p0) {
53789   int64x1_t __ret;
53790   __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
53791   return __ret;
53792 }
vqnegb_s8(int8_t __p0)53793 __ai int8_t vqnegb_s8(int8_t __p0) {
53794   int8_t __ret;
53795   __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
53796   return __ret;
53797 }
vqnegs_s32(int32_t __p0)53798 __ai int32_t vqnegs_s32(int32_t __p0) {
53799   int32_t __ret;
53800   __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
53801   return __ret;
53802 }
vqnegd_s64(int64_t __p0)53803 __ai int64_t vqnegd_s64(int64_t __p0) {
53804   int64_t __ret;
53805   __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
53806   return __ret;
53807 }
vqnegh_s16(int16_t __p0)53808 __ai int16_t vqnegh_s16(int16_t __p0) {
53809   int16_t __ret;
53810   __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
53811   return __ret;
53812 }
vqrdmulhs_s32(int32_t __p0,int32_t __p1)53813 __ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
53814   int32_t __ret;
53815   __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
53816   return __ret;
53817 }
vqrdmulhh_s16(int16_t __p0,int16_t __p1)53818 __ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
53819   int16_t __ret;
53820   __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
53821   return __ret;
53822 }
53823 #ifdef __LITTLE_ENDIAN__
53824 #define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53825   int32x4_t __ret; \
53826   int32x4_t __s0 = __p0; \
53827   int32x2_t __s1 = __p1; \
53828   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_lane_v((int8x16_t)__s0, (int8x8_t)__s1, __p2, 2); \
53829   __ret; \
53830 })
53831 #else
53832 #define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53833   int32x4_t __ret; \
53834   int32x4_t __s0 = __p0; \
53835   int32x2_t __s1 = __p1; \
53836   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53837   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53838   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_lane_v((int8x16_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
53839   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53840   __ret; \
53841 })
53842 #endif
53843 
53844 #ifdef __LITTLE_ENDIAN__
53845 #define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53846   int16x8_t __ret; \
53847   int16x8_t __s0 = __p0; \
53848   int16x4_t __s1 = __p1; \
53849   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_lane_v((int8x16_t)__s0, (int8x8_t)__s1, __p2, 1); \
53850   __ret; \
53851 })
53852 #else
53853 #define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53854   int16x8_t __ret; \
53855   int16x8_t __s0 = __p0; \
53856   int16x4_t __s1 = __p1; \
53857   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53858   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53859   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_lane_v((int8x16_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
53860   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53861   __ret; \
53862 })
53863 #endif
53864 
53865 #ifdef __LITTLE_ENDIAN__
53866 #define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53867   int32x2_t __ret; \
53868   int32x2_t __s0 = __p0; \
53869   int32x2_t __s1 = __p1; \
53870   __ret = (int32x2_t) __builtin_neon_vqrdmulh_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
53871   __ret; \
53872 })
53873 #else
53874 #define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53875   int32x2_t __ret; \
53876   int32x2_t __s0 = __p0; \
53877   int32x2_t __s1 = __p1; \
53878   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53879   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53880   __ret = (int32x2_t) __builtin_neon_vqrdmulh_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
53881   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53882   __ret; \
53883 })
53884 #endif
53885 
53886 #ifdef __LITTLE_ENDIAN__
53887 #define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53888   int16x4_t __ret; \
53889   int16x4_t __s0 = __p0; \
53890   int16x4_t __s1 = __p1; \
53891   __ret = (int16x4_t) __builtin_neon_vqrdmulh_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
53892   __ret; \
53893 })
53894 #else
53895 #define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53896   int16x4_t __ret; \
53897   int16x4_t __s0 = __p0; \
53898   int16x4_t __s1 = __p1; \
53899   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53900   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53901   __ret = (int16x4_t) __builtin_neon_vqrdmulh_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
53902   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53903   __ret; \
53904 })
53905 #endif
53906 
53907 #ifdef __LITTLE_ENDIAN__
53908 #define vqrdmulhs_lane_s32(__p0_659, __p1_659, __p2_659) __extension__ ({ \
53909   int32_t __ret_659; \
53910   int32_t __s0_659 = __p0_659; \
53911   int32x2_t __s1_659 = __p1_659; \
53912   __ret_659 = vqrdmulhs_s32(__s0_659, vget_lane_s32(__s1_659, __p2_659)); \
53913   __ret_659; \
53914 })
53915 #else
53916 #define vqrdmulhs_lane_s32(__p0_660, __p1_660, __p2_660) __extension__ ({ \
53917   int32_t __ret_660; \
53918   int32_t __s0_660 = __p0_660; \
53919   int32x2_t __s1_660 = __p1_660; \
53920   int32x2_t __rev1_660;  __rev1_660 = __builtin_shufflevector(__s1_660, __s1_660, 1, 0); \
53921   __ret_660 = vqrdmulhs_s32(__s0_660, __noswap_vget_lane_s32(__rev1_660, __p2_660)); \
53922   __ret_660; \
53923 })
53924 #endif
53925 
53926 #ifdef __LITTLE_ENDIAN__
53927 #define vqrdmulhh_lane_s16(__p0_661, __p1_661, __p2_661) __extension__ ({ \
53928   int16_t __ret_661; \
53929   int16_t __s0_661 = __p0_661; \
53930   int16x4_t __s1_661 = __p1_661; \
53931   __ret_661 = vqrdmulhh_s16(__s0_661, vget_lane_s16(__s1_661, __p2_661)); \
53932   __ret_661; \
53933 })
53934 #else
53935 #define vqrdmulhh_lane_s16(__p0_662, __p1_662, __p2_662) __extension__ ({ \
53936   int16_t __ret_662; \
53937   int16_t __s0_662 = __p0_662; \
53938   int16x4_t __s1_662 = __p1_662; \
53939   int16x4_t __rev1_662;  __rev1_662 = __builtin_shufflevector(__s1_662, __s1_662, 3, 2, 1, 0); \
53940   __ret_662 = vqrdmulhh_s16(__s0_662, __noswap_vget_lane_s16(__rev1_662, __p2_662)); \
53941   __ret_662; \
53942 })
53943 #endif
53944 
53945 #ifdef __LITTLE_ENDIAN__
53946 #define vqrdmulhs_laneq_s32(__p0_663, __p1_663, __p2_663) __extension__ ({ \
53947   int32_t __ret_663; \
53948   int32_t __s0_663 = __p0_663; \
53949   int32x4_t __s1_663 = __p1_663; \
53950   __ret_663 = vqrdmulhs_s32(__s0_663, vgetq_lane_s32(__s1_663, __p2_663)); \
53951   __ret_663; \
53952 })
53953 #else
53954 #define vqrdmulhs_laneq_s32(__p0_664, __p1_664, __p2_664) __extension__ ({ \
53955   int32_t __ret_664; \
53956   int32_t __s0_664 = __p0_664; \
53957   int32x4_t __s1_664 = __p1_664; \
53958   int32x4_t __rev1_664;  __rev1_664 = __builtin_shufflevector(__s1_664, __s1_664, 3, 2, 1, 0); \
53959   __ret_664 = vqrdmulhs_s32(__s0_664, __noswap_vgetq_lane_s32(__rev1_664, __p2_664)); \
53960   __ret_664; \
53961 })
53962 #endif
53963 
53964 #ifdef __LITTLE_ENDIAN__
53965 #define vqrdmulhh_laneq_s16(__p0_665, __p1_665, __p2_665) __extension__ ({ \
53966   int16_t __ret_665; \
53967   int16_t __s0_665 = __p0_665; \
53968   int16x8_t __s1_665 = __p1_665; \
53969   __ret_665 = vqrdmulhh_s16(__s0_665, vgetq_lane_s16(__s1_665, __p2_665)); \
53970   __ret_665; \
53971 })
53972 #else
53973 #define vqrdmulhh_laneq_s16(__p0_666, __p1_666, __p2_666) __extension__ ({ \
53974   int16_t __ret_666; \
53975   int16_t __s0_666 = __p0_666; \
53976   int16x8_t __s1_666 = __p1_666; \
53977   int16x8_t __rev1_666;  __rev1_666 = __builtin_shufflevector(__s1_666, __s1_666, 7, 6, 5, 4, 3, 2, 1, 0); \
53978   __ret_666 = vqrdmulhh_s16(__s0_666, __noswap_vgetq_lane_s16(__rev1_666, __p2_666)); \
53979   __ret_666; \
53980 })
53981 #endif
53982 
53983 #ifdef __LITTLE_ENDIAN__
53984 #define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53985   int32x4_t __ret; \
53986   int32x4_t __s0 = __p0; \
53987   int32x4_t __s1 = __p1; \
53988   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
53989   __ret; \
53990 })
53991 #else
53992 #define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53993   int32x4_t __ret; \
53994   int32x4_t __s0 = __p0; \
53995   int32x4_t __s1 = __p1; \
53996   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53997   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53998   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
53999   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54000   __ret; \
54001 })
54002 #endif
54003 
54004 #ifdef __LITTLE_ENDIAN__
54005 #define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54006   int16x8_t __ret; \
54007   int16x8_t __s0 = __p0; \
54008   int16x8_t __s1 = __p1; \
54009   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
54010   __ret; \
54011 })
54012 #else
54013 #define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54014   int16x8_t __ret; \
54015   int16x8_t __s0 = __p0; \
54016   int16x8_t __s1 = __p1; \
54017   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54018   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54019   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
54020   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
54021   __ret; \
54022 })
54023 #endif
54024 
54025 #ifdef __LITTLE_ENDIAN__
54026 #define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54027   int32x2_t __ret; \
54028   int32x2_t __s0 = __p0; \
54029   int32x4_t __s1 = __p1; \
54030   __ret = (int32x2_t) __builtin_neon_vqrdmulh_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 2); \
54031   __ret; \
54032 })
54033 #else
54034 #define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54035   int32x2_t __ret; \
54036   int32x2_t __s0 = __p0; \
54037   int32x4_t __s1 = __p1; \
54038   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54039   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54040   __ret = (int32x2_t) __builtin_neon_vqrdmulh_laneq_v((int8x8_t)__rev0, (int8x16_t)__rev1, __p2, 2); \
54041   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54042   __ret; \
54043 })
54044 #endif
54045 
54046 #ifdef __LITTLE_ENDIAN__
54047 #define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54048   int16x4_t __ret; \
54049   int16x4_t __s0 = __p0; \
54050   int16x8_t __s1 = __p1; \
54051   __ret = (int16x4_t) __builtin_neon_vqrdmulh_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 1); \
54052   __ret; \
54053 })
54054 #else
54055 #define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54056   int16x4_t __ret; \
54057   int16x4_t __s0 = __p0; \
54058   int16x8_t __s1 = __p1; \
54059   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54060   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54061   __ret = (int16x4_t) __builtin_neon_vqrdmulh_laneq_v((int8x8_t)__rev0, (int8x16_t)__rev1, __p2, 1); \
54062   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54063   __ret; \
54064 })
54065 #endif
54066 
vqrshlb_u8(uint8_t __p0,int8_t __p1)54067 __ai uint8_t vqrshlb_u8(uint8_t __p0, int8_t __p1) {
54068   uint8_t __ret;
54069   __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
54070   return __ret;
54071 }
vqrshls_u32(uint32_t __p0,int32_t __p1)54072 __ai uint32_t vqrshls_u32(uint32_t __p0, int32_t __p1) {
54073   uint32_t __ret;
54074   __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
54075   return __ret;
54076 }
vqrshld_u64(uint64_t __p0,int64_t __p1)54077 __ai uint64_t vqrshld_u64(uint64_t __p0, int64_t __p1) {
54078   uint64_t __ret;
54079   __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
54080   return __ret;
54081 }
vqrshlh_u16(uint16_t __p0,int16_t __p1)54082 __ai uint16_t vqrshlh_u16(uint16_t __p0, int16_t __p1) {
54083   uint16_t __ret;
54084   __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
54085   return __ret;
54086 }
vqrshlb_s8(int8_t __p0,int8_t __p1)54087 __ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
54088   int8_t __ret;
54089   __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
54090   return __ret;
54091 }
vqrshls_s32(int32_t __p0,int32_t __p1)54092 __ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
54093   int32_t __ret;
54094   __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
54095   return __ret;
54096 }
vqrshld_s64(int64_t __p0,int64_t __p1)54097 __ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
54098   int64_t __ret;
54099   __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
54100   return __ret;
54101 }
vqrshlh_s16(int16_t __p0,int16_t __p1)54102 __ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
54103   int16_t __ret;
54104   __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
54105   return __ret;
54106 }
54107 #ifdef __LITTLE_ENDIAN__
54108 #define vqrshrn_high_n_u32(__p0_667, __p1_667, __p2_667) __extension__ ({ \
54109   uint16x8_t __ret_667; \
54110   uint16x4_t __s0_667 = __p0_667; \
54111   uint32x4_t __s1_667 = __p1_667; \
54112   __ret_667 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_667), (uint16x4_t)(vqrshrn_n_u32(__s1_667, __p2_667)))); \
54113   __ret_667; \
54114 })
54115 #else
54116 #define vqrshrn_high_n_u32(__p0_668, __p1_668, __p2_668) __extension__ ({ \
54117   uint16x8_t __ret_668; \
54118   uint16x4_t __s0_668 = __p0_668; \
54119   uint32x4_t __s1_668 = __p1_668; \
54120   uint16x4_t __rev0_668;  __rev0_668 = __builtin_shufflevector(__s0_668, __s0_668, 3, 2, 1, 0); \
54121   uint32x4_t __rev1_668;  __rev1_668 = __builtin_shufflevector(__s1_668, __s1_668, 3, 2, 1, 0); \
54122   __ret_668 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_668), (uint16x4_t)(__noswap_vqrshrn_n_u32(__rev1_668, __p2_668)))); \
54123   __ret_668 = __builtin_shufflevector(__ret_668, __ret_668, 7, 6, 5, 4, 3, 2, 1, 0); \
54124   __ret_668; \
54125 })
54126 #endif
54127 
54128 #ifdef __LITTLE_ENDIAN__
54129 #define vqrshrn_high_n_u64(__p0_669, __p1_669, __p2_669) __extension__ ({ \
54130   uint32x4_t __ret_669; \
54131   uint32x2_t __s0_669 = __p0_669; \
54132   uint64x2_t __s1_669 = __p1_669; \
54133   __ret_669 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_669), (uint32x2_t)(vqrshrn_n_u64(__s1_669, __p2_669)))); \
54134   __ret_669; \
54135 })
54136 #else
54137 #define vqrshrn_high_n_u64(__p0_670, __p1_670, __p2_670) __extension__ ({ \
54138   uint32x4_t __ret_670; \
54139   uint32x2_t __s0_670 = __p0_670; \
54140   uint64x2_t __s1_670 = __p1_670; \
54141   uint32x2_t __rev0_670;  __rev0_670 = __builtin_shufflevector(__s0_670, __s0_670, 1, 0); \
54142   uint64x2_t __rev1_670;  __rev1_670 = __builtin_shufflevector(__s1_670, __s1_670, 1, 0); \
54143   __ret_670 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_670), (uint32x2_t)(__noswap_vqrshrn_n_u64(__rev1_670, __p2_670)))); \
54144   __ret_670 = __builtin_shufflevector(__ret_670, __ret_670, 3, 2, 1, 0); \
54145   __ret_670; \
54146 })
54147 #endif
54148 
54149 #ifdef __LITTLE_ENDIAN__
54150 #define vqrshrn_high_n_u16(__p0_671, __p1_671, __p2_671) __extension__ ({ \
54151   uint8x16_t __ret_671; \
54152   uint8x8_t __s0_671 = __p0_671; \
54153   uint16x8_t __s1_671 = __p1_671; \
54154   __ret_671 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_671), (uint8x8_t)(vqrshrn_n_u16(__s1_671, __p2_671)))); \
54155   __ret_671; \
54156 })
54157 #else
54158 #define vqrshrn_high_n_u16(__p0_672, __p1_672, __p2_672) __extension__ ({ \
54159   uint8x16_t __ret_672; \
54160   uint8x8_t __s0_672 = __p0_672; \
54161   uint16x8_t __s1_672 = __p1_672; \
54162   uint8x8_t __rev0_672;  __rev0_672 = __builtin_shufflevector(__s0_672, __s0_672, 7, 6, 5, 4, 3, 2, 1, 0); \
54163   uint16x8_t __rev1_672;  __rev1_672 = __builtin_shufflevector(__s1_672, __s1_672, 7, 6, 5, 4, 3, 2, 1, 0); \
54164   __ret_672 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_672), (uint8x8_t)(__noswap_vqrshrn_n_u16(__rev1_672, __p2_672)))); \
54165   __ret_672 = __builtin_shufflevector(__ret_672, __ret_672, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
54166   __ret_672; \
54167 })
54168 #endif
54169 
54170 #ifdef __LITTLE_ENDIAN__
54171 #define vqrshrn_high_n_s32(__p0_673, __p1_673, __p2_673) __extension__ ({ \
54172   int16x8_t __ret_673; \
54173   int16x4_t __s0_673 = __p0_673; \
54174   int32x4_t __s1_673 = __p1_673; \
54175   __ret_673 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_673), (int16x4_t)(vqrshrn_n_s32(__s1_673, __p2_673)))); \
54176   __ret_673; \
54177 })
54178 #else
54179 #define vqrshrn_high_n_s32(__p0_674, __p1_674, __p2_674) __extension__ ({ \
54180   int16x8_t __ret_674; \
54181   int16x4_t __s0_674 = __p0_674; \
54182   int32x4_t __s1_674 = __p1_674; \
54183   int16x4_t __rev0_674;  __rev0_674 = __builtin_shufflevector(__s0_674, __s0_674, 3, 2, 1, 0); \
54184   int32x4_t __rev1_674;  __rev1_674 = __builtin_shufflevector(__s1_674, __s1_674, 3, 2, 1, 0); \
54185   __ret_674 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_674), (int16x4_t)(__noswap_vqrshrn_n_s32(__rev1_674, __p2_674)))); \
54186   __ret_674 = __builtin_shufflevector(__ret_674, __ret_674, 7, 6, 5, 4, 3, 2, 1, 0); \
54187   __ret_674; \
54188 })
54189 #endif
54190 
54191 #ifdef __LITTLE_ENDIAN__
54192 #define vqrshrn_high_n_s64(__p0_675, __p1_675, __p2_675) __extension__ ({ \
54193   int32x4_t __ret_675; \
54194   int32x2_t __s0_675 = __p0_675; \
54195   int64x2_t __s1_675 = __p1_675; \
54196   __ret_675 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_675), (int32x2_t)(vqrshrn_n_s64(__s1_675, __p2_675)))); \
54197   __ret_675; \
54198 })
54199 #else
54200 #define vqrshrn_high_n_s64(__p0_676, __p1_676, __p2_676) __extension__ ({ \
54201   int32x4_t __ret_676; \
54202   int32x2_t __s0_676 = __p0_676; \
54203   int64x2_t __s1_676 = __p1_676; \
54204   int32x2_t __rev0_676;  __rev0_676 = __builtin_shufflevector(__s0_676, __s0_676, 1, 0); \
54205   int64x2_t __rev1_676;  __rev1_676 = __builtin_shufflevector(__s1_676, __s1_676, 1, 0); \
54206   __ret_676 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_676), (int32x2_t)(__noswap_vqrshrn_n_s64(__rev1_676, __p2_676)))); \
54207   __ret_676 = __builtin_shufflevector(__ret_676, __ret_676, 3, 2, 1, 0); \
54208   __ret_676; \
54209 })
54210 #endif
54211 
54212 #ifdef __LITTLE_ENDIAN__
54213 #define vqrshrn_high_n_s16(__p0_677, __p1_677, __p2_677) __extension__ ({ \
54214   int8x16_t __ret_677; \
54215   int8x8_t __s0_677 = __p0_677; \
54216   int16x8_t __s1_677 = __p1_677; \
54217   __ret_677 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_677), (int8x8_t)(vqrshrn_n_s16(__s1_677, __p2_677)))); \
54218   __ret_677; \
54219 })
54220 #else
54221 #define vqrshrn_high_n_s16(__p0_678, __p1_678, __p2_678) __extension__ ({ \
54222   int8x16_t __ret_678; \
54223   int8x8_t __s0_678 = __p0_678; \
54224   int16x8_t __s1_678 = __p1_678; \
54225   int8x8_t __rev0_678;  __rev0_678 = __builtin_shufflevector(__s0_678, __s0_678, 7, 6, 5, 4, 3, 2, 1, 0); \
54226   int16x8_t __rev1_678;  __rev1_678 = __builtin_shufflevector(__s1_678, __s1_678, 7, 6, 5, 4, 3, 2, 1, 0); \
54227   __ret_678 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_678), (int8x8_t)(__noswap_vqrshrn_n_s16(__rev1_678, __p2_678)))); \
54228   __ret_678 = __builtin_shufflevector(__ret_678, __ret_678, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
54229   __ret_678; \
54230 })
54231 #endif
54232 
54233 #define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
54234   uint16_t __ret; \
54235   uint32_t __s0 = __p0; \
54236   __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
54237   __ret; \
54238 })
54239 #define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
54240   uint32_t __ret; \
54241   uint64_t __s0 = __p0; \
54242   __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
54243   __ret; \
54244 })
54245 #define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
54246   uint8_t __ret; \
54247   uint16_t __s0 = __p0; \
54248   __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
54249   __ret; \
54250 })
54251 #define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
54252   int16_t __ret; \
54253   int32_t __s0 = __p0; \
54254   __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
54255   __ret; \
54256 })
54257 #define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
54258   int32_t __ret; \
54259   int64_t __s0 = __p0; \
54260   __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
54261   __ret; \
54262 })
54263 #define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
54264   int8_t __ret; \
54265   int16_t __s0 = __p0; \
54266   __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
54267   __ret; \
54268 })
54269 #ifdef __LITTLE_ENDIAN__
54270 #define vqrshrun_high_n_s32(__p0_679, __p1_679, __p2_679) __extension__ ({ \
54271   int16x8_t __ret_679; \
54272   int16x4_t __s0_679 = __p0_679; \
54273   int32x4_t __s1_679 = __p1_679; \
54274   __ret_679 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_679), (int16x4_t)(vqrshrun_n_s32(__s1_679, __p2_679)))); \
54275   __ret_679; \
54276 })
54277 #else
54278 #define vqrshrun_high_n_s32(__p0_680, __p1_680, __p2_680) __extension__ ({ \
54279   int16x8_t __ret_680; \
54280   int16x4_t __s0_680 = __p0_680; \
54281   int32x4_t __s1_680 = __p1_680; \
54282   int16x4_t __rev0_680;  __rev0_680 = __builtin_shufflevector(__s0_680, __s0_680, 3, 2, 1, 0); \
54283   int32x4_t __rev1_680;  __rev1_680 = __builtin_shufflevector(__s1_680, __s1_680, 3, 2, 1, 0); \
54284   __ret_680 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_680), (int16x4_t)(__noswap_vqrshrun_n_s32(__rev1_680, __p2_680)))); \
54285   __ret_680 = __builtin_shufflevector(__ret_680, __ret_680, 7, 6, 5, 4, 3, 2, 1, 0); \
54286   __ret_680; \
54287 })
54288 #endif
54289 
54290 #ifdef __LITTLE_ENDIAN__
54291 #define vqrshrun_high_n_s64(__p0_681, __p1_681, __p2_681) __extension__ ({ \
54292   int32x4_t __ret_681; \
54293   int32x2_t __s0_681 = __p0_681; \
54294   int64x2_t __s1_681 = __p1_681; \
54295   __ret_681 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_681), (int32x2_t)(vqrshrun_n_s64(__s1_681, __p2_681)))); \
54296   __ret_681; \
54297 })
54298 #else
54299 #define vqrshrun_high_n_s64(__p0_682, __p1_682, __p2_682) __extension__ ({ \
54300   int32x4_t __ret_682; \
54301   int32x2_t __s0_682 = __p0_682; \
54302   int64x2_t __s1_682 = __p1_682; \
54303   int32x2_t __rev0_682;  __rev0_682 = __builtin_shufflevector(__s0_682, __s0_682, 1, 0); \
54304   int64x2_t __rev1_682;  __rev1_682 = __builtin_shufflevector(__s1_682, __s1_682, 1, 0); \
54305   __ret_682 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_682), (int32x2_t)(__noswap_vqrshrun_n_s64(__rev1_682, __p2_682)))); \
54306   __ret_682 = __builtin_shufflevector(__ret_682, __ret_682, 3, 2, 1, 0); \
54307   __ret_682; \
54308 })
54309 #endif
54310 
54311 #ifdef __LITTLE_ENDIAN__
54312 #define vqrshrun_high_n_s16(__p0_683, __p1_683, __p2_683) __extension__ ({ \
54313   int8x16_t __ret_683; \
54314   int8x8_t __s0_683 = __p0_683; \
54315   int16x8_t __s1_683 = __p1_683; \
54316   __ret_683 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_683), (int8x8_t)(vqrshrun_n_s16(__s1_683, __p2_683)))); \
54317   __ret_683; \
54318 })
54319 #else
54320 #define vqrshrun_high_n_s16(__p0_684, __p1_684, __p2_684) __extension__ ({ \
54321   int8x16_t __ret_684; \
54322   int8x8_t __s0_684 = __p0_684; \
54323   int16x8_t __s1_684 = __p1_684; \
54324   int8x8_t __rev0_684;  __rev0_684 = __builtin_shufflevector(__s0_684, __s0_684, 7, 6, 5, 4, 3, 2, 1, 0); \
54325   int16x8_t __rev1_684;  __rev1_684 = __builtin_shufflevector(__s1_684, __s1_684, 7, 6, 5, 4, 3, 2, 1, 0); \
54326   __ret_684 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_684), (int8x8_t)(__noswap_vqrshrun_n_s16(__rev1_684, __p2_684)))); \
54327   __ret_684 = __builtin_shufflevector(__ret_684, __ret_684, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
54328   __ret_684; \
54329 })
54330 #endif
54331 
54332 #define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
54333   uint16_t __ret; \
54334   int32_t __s0 = __p0; \
54335   __ret = (uint16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
54336   __ret; \
54337 })
54338 #define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
54339   uint32_t __ret; \
54340   int64_t __s0 = __p0; \
54341   __ret = (uint32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
54342   __ret; \
54343 })
54344 #define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
54345   uint8_t __ret; \
54346   int16_t __s0 = __p0; \
54347   __ret = (uint8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
54348   __ret; \
54349 })
vqshlb_u8(uint8_t __p0,int8_t __p1)54350 __ai uint8_t vqshlb_u8(uint8_t __p0, int8_t __p1) {
54351   uint8_t __ret;
54352   __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
54353   return __ret;
54354 }
vqshls_u32(uint32_t __p0,int32_t __p1)54355 __ai uint32_t vqshls_u32(uint32_t __p0, int32_t __p1) {
54356   uint32_t __ret;
54357   __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
54358   return __ret;
54359 }
vqshld_u64(uint64_t __p0,int64_t __p1)54360 __ai uint64_t vqshld_u64(uint64_t __p0, int64_t __p1) {
54361   uint64_t __ret;
54362   __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
54363   return __ret;
54364 }
vqshlh_u16(uint16_t __p0,int16_t __p1)54365 __ai uint16_t vqshlh_u16(uint16_t __p0, int16_t __p1) {
54366   uint16_t __ret;
54367   __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
54368   return __ret;
54369 }
vqshlb_s8(int8_t __p0,int8_t __p1)54370 __ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
54371   int8_t __ret;
54372   __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
54373   return __ret;
54374 }
vqshls_s32(int32_t __p0,int32_t __p1)54375 __ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
54376   int32_t __ret;
54377   __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
54378   return __ret;
54379 }
vqshld_s64(int64_t __p0,int64_t __p1)54380 __ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
54381   int64_t __ret;
54382   __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
54383   return __ret;
54384 }
vqshlh_s16(int16_t __p0,int16_t __p1)54385 __ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
54386   int16_t __ret;
54387   __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
54388   return __ret;
54389 }
54390 #define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
54391   uint8_t __ret; \
54392   uint8_t __s0 = __p0; \
54393   __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
54394   __ret; \
54395 })
54396 #define vqshls_n_u32(__p0, __p1) __extension__ ({ \
54397   uint32_t __ret; \
54398   uint32_t __s0 = __p0; \
54399   __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
54400   __ret; \
54401 })
54402 #define vqshld_n_u64(__p0, __p1) __extension__ ({ \
54403   uint64_t __ret; \
54404   uint64_t __s0 = __p0; \
54405   __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
54406   __ret; \
54407 })
54408 #define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
54409   uint16_t __ret; \
54410   uint16_t __s0 = __p0; \
54411   __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
54412   __ret; \
54413 })
54414 #define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
54415   int8_t __ret; \
54416   int8_t __s0 = __p0; \
54417   __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
54418   __ret; \
54419 })
54420 #define vqshls_n_s32(__p0, __p1) __extension__ ({ \
54421   int32_t __ret; \
54422   int32_t __s0 = __p0; \
54423   __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
54424   __ret; \
54425 })
54426 #define vqshld_n_s64(__p0, __p1) __extension__ ({ \
54427   int64_t __ret; \
54428   int64_t __s0 = __p0; \
54429   __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
54430   __ret; \
54431 })
54432 #define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
54433   int16_t __ret; \
54434   int16_t __s0 = __p0; \
54435   __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
54436   __ret; \
54437 })
54438 #define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
54439   int8_t __ret; \
54440   int8_t __s0 = __p0; \
54441   __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
54442   __ret; \
54443 })
54444 #define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
54445   int32_t __ret; \
54446   int32_t __s0 = __p0; \
54447   __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
54448   __ret; \
54449 })
54450 #define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
54451   int64_t __ret; \
54452   int64_t __s0 = __p0; \
54453   __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
54454   __ret; \
54455 })
54456 #define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
54457   int16_t __ret; \
54458   int16_t __s0 = __p0; \
54459   __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
54460   __ret; \
54461 })
54462 #ifdef __LITTLE_ENDIAN__
54463 #define vqshrn_high_n_u32(__p0_685, __p1_685, __p2_685) __extension__ ({ \
54464   uint16x8_t __ret_685; \
54465   uint16x4_t __s0_685 = __p0_685; \
54466   uint32x4_t __s1_685 = __p1_685; \
54467   __ret_685 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_685), (uint16x4_t)(vqshrn_n_u32(__s1_685, __p2_685)))); \
54468   __ret_685; \
54469 })
54470 #else
54471 #define vqshrn_high_n_u32(__p0_686, __p1_686, __p2_686) __extension__ ({ \
54472   uint16x8_t __ret_686; \
54473   uint16x4_t __s0_686 = __p0_686; \
54474   uint32x4_t __s1_686 = __p1_686; \
54475   uint16x4_t __rev0_686;  __rev0_686 = __builtin_shufflevector(__s0_686, __s0_686, 3, 2, 1, 0); \
54476   uint32x4_t __rev1_686;  __rev1_686 = __builtin_shufflevector(__s1_686, __s1_686, 3, 2, 1, 0); \
54477   __ret_686 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_686), (uint16x4_t)(__noswap_vqshrn_n_u32(__rev1_686, __p2_686)))); \
54478   __ret_686 = __builtin_shufflevector(__ret_686, __ret_686, 7, 6, 5, 4, 3, 2, 1, 0); \
54479   __ret_686; \
54480 })
54481 #endif
54482 
54483 #ifdef __LITTLE_ENDIAN__
54484 #define vqshrn_high_n_u64(__p0_687, __p1_687, __p2_687) __extension__ ({ \
54485   uint32x4_t __ret_687; \
54486   uint32x2_t __s0_687 = __p0_687; \
54487   uint64x2_t __s1_687 = __p1_687; \
54488   __ret_687 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_687), (uint32x2_t)(vqshrn_n_u64(__s1_687, __p2_687)))); \
54489   __ret_687; \
54490 })
54491 #else
54492 #define vqshrn_high_n_u64(__p0_688, __p1_688, __p2_688) __extension__ ({ \
54493   uint32x4_t __ret_688; \
54494   uint32x2_t __s0_688 = __p0_688; \
54495   uint64x2_t __s1_688 = __p1_688; \
54496   uint32x2_t __rev0_688;  __rev0_688 = __builtin_shufflevector(__s0_688, __s0_688, 1, 0); \
54497   uint64x2_t __rev1_688;  __rev1_688 = __builtin_shufflevector(__s1_688, __s1_688, 1, 0); \
54498   __ret_688 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_688), (uint32x2_t)(__noswap_vqshrn_n_u64(__rev1_688, __p2_688)))); \
54499   __ret_688 = __builtin_shufflevector(__ret_688, __ret_688, 3, 2, 1, 0); \
54500   __ret_688; \
54501 })
54502 #endif
54503 
54504 #ifdef __LITTLE_ENDIAN__
54505 #define vqshrn_high_n_u16(__p0_689, __p1_689, __p2_689) __extension__ ({ \
54506   uint8x16_t __ret_689; \
54507   uint8x8_t __s0_689 = __p0_689; \
54508   uint16x8_t __s1_689 = __p1_689; \
54509   __ret_689 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_689), (uint8x8_t)(vqshrn_n_u16(__s1_689, __p2_689)))); \
54510   __ret_689; \
54511 })
54512 #else
54513 #define vqshrn_high_n_u16(__p0_690, __p1_690, __p2_690) __extension__ ({ \
54514   uint8x16_t __ret_690; \
54515   uint8x8_t __s0_690 = __p0_690; \
54516   uint16x8_t __s1_690 = __p1_690; \
54517   uint8x8_t __rev0_690;  __rev0_690 = __builtin_shufflevector(__s0_690, __s0_690, 7, 6, 5, 4, 3, 2, 1, 0); \
54518   uint16x8_t __rev1_690;  __rev1_690 = __builtin_shufflevector(__s1_690, __s1_690, 7, 6, 5, 4, 3, 2, 1, 0); \
54519   __ret_690 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_690), (uint8x8_t)(__noswap_vqshrn_n_u16(__rev1_690, __p2_690)))); \
54520   __ret_690 = __builtin_shufflevector(__ret_690, __ret_690, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
54521   __ret_690; \
54522 })
54523 #endif
54524 
54525 #ifdef __LITTLE_ENDIAN__
54526 #define vqshrn_high_n_s32(__p0_691, __p1_691, __p2_691) __extension__ ({ \
54527   int16x8_t __ret_691; \
54528   int16x4_t __s0_691 = __p0_691; \
54529   int32x4_t __s1_691 = __p1_691; \
54530   __ret_691 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_691), (int16x4_t)(vqshrn_n_s32(__s1_691, __p2_691)))); \
54531   __ret_691; \
54532 })
54533 #else
54534 #define vqshrn_high_n_s32(__p0_692, __p1_692, __p2_692) __extension__ ({ \
54535   int16x8_t __ret_692; \
54536   int16x4_t __s0_692 = __p0_692; \
54537   int32x4_t __s1_692 = __p1_692; \
54538   int16x4_t __rev0_692;  __rev0_692 = __builtin_shufflevector(__s0_692, __s0_692, 3, 2, 1, 0); \
54539   int32x4_t __rev1_692;  __rev1_692 = __builtin_shufflevector(__s1_692, __s1_692, 3, 2, 1, 0); \
54540   __ret_692 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_692), (int16x4_t)(__noswap_vqshrn_n_s32(__rev1_692, __p2_692)))); \
54541   __ret_692 = __builtin_shufflevector(__ret_692, __ret_692, 7, 6, 5, 4, 3, 2, 1, 0); \
54542   __ret_692; \
54543 })
54544 #endif
54545 
54546 #ifdef __LITTLE_ENDIAN__
54547 #define vqshrn_high_n_s64(__p0_693, __p1_693, __p2_693) __extension__ ({ \
54548   int32x4_t __ret_693; \
54549   int32x2_t __s0_693 = __p0_693; \
54550   int64x2_t __s1_693 = __p1_693; \
54551   __ret_693 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_693), (int32x2_t)(vqshrn_n_s64(__s1_693, __p2_693)))); \
54552   __ret_693; \
54553 })
54554 #else
54555 #define vqshrn_high_n_s64(__p0_694, __p1_694, __p2_694) __extension__ ({ \
54556   int32x4_t __ret_694; \
54557   int32x2_t __s0_694 = __p0_694; \
54558   int64x2_t __s1_694 = __p1_694; \
54559   int32x2_t __rev0_694;  __rev0_694 = __builtin_shufflevector(__s0_694, __s0_694, 1, 0); \
54560   int64x2_t __rev1_694;  __rev1_694 = __builtin_shufflevector(__s1_694, __s1_694, 1, 0); \
54561   __ret_694 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_694), (int32x2_t)(__noswap_vqshrn_n_s64(__rev1_694, __p2_694)))); \
54562   __ret_694 = __builtin_shufflevector(__ret_694, __ret_694, 3, 2, 1, 0); \
54563   __ret_694; \
54564 })
54565 #endif
54566 
54567 #ifdef __LITTLE_ENDIAN__
54568 #define vqshrn_high_n_s16(__p0_695, __p1_695, __p2_695) __extension__ ({ \
54569   int8x16_t __ret_695; \
54570   int8x8_t __s0_695 = __p0_695; \
54571   int16x8_t __s1_695 = __p1_695; \
54572   __ret_695 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_695), (int8x8_t)(vqshrn_n_s16(__s1_695, __p2_695)))); \
54573   __ret_695; \
54574 })
54575 #else
54576 #define vqshrn_high_n_s16(__p0_696, __p1_696, __p2_696) __extension__ ({ \
54577   int8x16_t __ret_696; \
54578   int8x8_t __s0_696 = __p0_696; \
54579   int16x8_t __s1_696 = __p1_696; \
54580   int8x8_t __rev0_696;  __rev0_696 = __builtin_shufflevector(__s0_696, __s0_696, 7, 6, 5, 4, 3, 2, 1, 0); \
54581   int16x8_t __rev1_696;  __rev1_696 = __builtin_shufflevector(__s1_696, __s1_696, 7, 6, 5, 4, 3, 2, 1, 0); \
54582   __ret_696 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_696), (int8x8_t)(__noswap_vqshrn_n_s16(__rev1_696, __p2_696)))); \
54583   __ret_696 = __builtin_shufflevector(__ret_696, __ret_696, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
54584   __ret_696; \
54585 })
54586 #endif
54587 
54588 #define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
54589   uint16_t __ret; \
54590   uint32_t __s0 = __p0; \
54591   __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
54592   __ret; \
54593 })
54594 #define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
54595   uint32_t __ret; \
54596   uint64_t __s0 = __p0; \
54597   __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
54598   __ret; \
54599 })
54600 #define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
54601   uint8_t __ret; \
54602   uint16_t __s0 = __p0; \
54603   __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
54604   __ret; \
54605 })
54606 #define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
54607   int16_t __ret; \
54608   int32_t __s0 = __p0; \
54609   __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
54610   __ret; \
54611 })
54612 #define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
54613   int32_t __ret; \
54614   int64_t __s0 = __p0; \
54615   __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
54616   __ret; \
54617 })
54618 #define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
54619   int8_t __ret; \
54620   int16_t __s0 = __p0; \
54621   __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
54622   __ret; \
54623 })
54624 #ifdef __LITTLE_ENDIAN__
54625 #define vqshrun_high_n_s32(__p0_697, __p1_697, __p2_697) __extension__ ({ \
54626   int16x8_t __ret_697; \
54627   int16x4_t __s0_697 = __p0_697; \
54628   int32x4_t __s1_697 = __p1_697; \
54629   __ret_697 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_697), (int16x4_t)(vqshrun_n_s32(__s1_697, __p2_697)))); \
54630   __ret_697; \
54631 })
54632 #else
54633 #define vqshrun_high_n_s32(__p0_698, __p1_698, __p2_698) __extension__ ({ \
54634   int16x8_t __ret_698; \
54635   int16x4_t __s0_698 = __p0_698; \
54636   int32x4_t __s1_698 = __p1_698; \
54637   int16x4_t __rev0_698;  __rev0_698 = __builtin_shufflevector(__s0_698, __s0_698, 3, 2, 1, 0); \
54638   int32x4_t __rev1_698;  __rev1_698 = __builtin_shufflevector(__s1_698, __s1_698, 3, 2, 1, 0); \
54639   __ret_698 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_698), (int16x4_t)(__noswap_vqshrun_n_s32(__rev1_698, __p2_698)))); \
54640   __ret_698 = __builtin_shufflevector(__ret_698, __ret_698, 7, 6, 5, 4, 3, 2, 1, 0); \
54641   __ret_698; \
54642 })
54643 #endif
54644 
54645 #ifdef __LITTLE_ENDIAN__
54646 #define vqshrun_high_n_s64(__p0_699, __p1_699, __p2_699) __extension__ ({ \
54647   int32x4_t __ret_699; \
54648   int32x2_t __s0_699 = __p0_699; \
54649   int64x2_t __s1_699 = __p1_699; \
54650   __ret_699 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_699), (int32x2_t)(vqshrun_n_s64(__s1_699, __p2_699)))); \
54651   __ret_699; \
54652 })
54653 #else
54654 #define vqshrun_high_n_s64(__p0_700, __p1_700, __p2_700) __extension__ ({ \
54655   int32x4_t __ret_700; \
54656   int32x2_t __s0_700 = __p0_700; \
54657   int64x2_t __s1_700 = __p1_700; \
54658   int32x2_t __rev0_700;  __rev0_700 = __builtin_shufflevector(__s0_700, __s0_700, 1, 0); \
54659   int64x2_t __rev1_700;  __rev1_700 = __builtin_shufflevector(__s1_700, __s1_700, 1, 0); \
54660   __ret_700 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_700), (int32x2_t)(__noswap_vqshrun_n_s64(__rev1_700, __p2_700)))); \
54661   __ret_700 = __builtin_shufflevector(__ret_700, __ret_700, 3, 2, 1, 0); \
54662   __ret_700; \
54663 })
54664 #endif
54665 
54666 #ifdef __LITTLE_ENDIAN__
54667 #define vqshrun_high_n_s16(__p0_701, __p1_701, __p2_701) __extension__ ({ \
54668   int8x16_t __ret_701; \
54669   int8x8_t __s0_701 = __p0_701; \
54670   int16x8_t __s1_701 = __p1_701; \
54671   __ret_701 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_701), (int8x8_t)(vqshrun_n_s16(__s1_701, __p2_701)))); \
54672   __ret_701; \
54673 })
54674 #else
54675 #define vqshrun_high_n_s16(__p0_702, __p1_702, __p2_702) __extension__ ({ \
54676   int8x16_t __ret_702; \
54677   int8x8_t __s0_702 = __p0_702; \
54678   int16x8_t __s1_702 = __p1_702; \
54679   int8x8_t __rev0_702;  __rev0_702 = __builtin_shufflevector(__s0_702, __s0_702, 7, 6, 5, 4, 3, 2, 1, 0); \
54680   int16x8_t __rev1_702;  __rev1_702 = __builtin_shufflevector(__s1_702, __s1_702, 7, 6, 5, 4, 3, 2, 1, 0); \
54681   __ret_702 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_702), (int8x8_t)(__noswap_vqshrun_n_s16(__rev1_702, __p2_702)))); \
54682   __ret_702 = __builtin_shufflevector(__ret_702, __ret_702, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
54683   __ret_702; \
54684 })
54685 #endif
54686 
54687 #define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
54688   uint16_t __ret; \
54689   int32_t __s0 = __p0; \
54690   __ret = (uint16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
54691   __ret; \
54692 })
54693 #define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
54694   uint32_t __ret; \
54695   int64_t __s0 = __p0; \
54696   __ret = (uint32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
54697   __ret; \
54698 })
54699 #define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
54700   uint8_t __ret; \
54701   int16_t __s0 = __p0; \
54702   __ret = (uint8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
54703   __ret; \
54704 })
vqsubb_u8(uint8_t __p0,uint8_t __p1)54705 __ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
54706   uint8_t __ret;
54707   __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
54708   return __ret;
54709 }
vqsubs_u32(uint32_t __p0,uint32_t __p1)54710 __ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
54711   uint32_t __ret;
54712   __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
54713   return __ret;
54714 }
vqsubd_u64(uint64_t __p0,uint64_t __p1)54715 __ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
54716   uint64_t __ret;
54717   __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
54718   return __ret;
54719 }
vqsubh_u16(uint16_t __p0,uint16_t __p1)54720 __ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
54721   uint16_t __ret;
54722   __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
54723   return __ret;
54724 }
vqsubb_s8(int8_t __p0,int8_t __p1)54725 __ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
54726   int8_t __ret;
54727   __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
54728   return __ret;
54729 }
vqsubs_s32(int32_t __p0,int32_t __p1)54730 __ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
54731   int32_t __ret;
54732   __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
54733   return __ret;
54734 }
vqsubd_s64(int64_t __p0,int64_t __p1)54735 __ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
54736   int64_t __ret;
54737   __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
54738   return __ret;
54739 }
vqsubh_s16(int16_t __p0,int16_t __p1)54740 __ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
54741   int16_t __ret;
54742   __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
54743   return __ret;
54744 }
54745 #ifdef __LITTLE_ENDIAN__
vqtbl1_p8(poly8x16_t __p0,uint8x8_t __p1)54746 __ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
54747   poly8x8_t __ret;
54748   __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 4);
54749   return __ret;
54750 }
54751 #else
vqtbl1_p8(poly8x16_t __p0,uint8x8_t __p1)54752 __ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
54753   poly8x8_t __ret;
54754   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54755   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54756   __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 4);
54757   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54758   return __ret;
54759 }
54760 #endif
54761 
54762 #ifdef __LITTLE_ENDIAN__
vqtbl1q_p8(poly8x16_t __p0,uint8x16_t __p1)54763 __ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
54764   poly8x16_t __ret;
54765   __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
54766   return __ret;
54767 }
54768 #else
vqtbl1q_p8(poly8x16_t __p0,uint8x16_t __p1)54769 __ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
54770   poly8x16_t __ret;
54771   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54772   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54773   __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
54774   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54775   return __ret;
54776 }
54777 #endif
54778 
54779 #ifdef __LITTLE_ENDIAN__
vqtbl1q_u8(uint8x16_t __p0,uint8x16_t __p1)54780 __ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
54781   uint8x16_t __ret;
54782   __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
54783   return __ret;
54784 }
54785 #else
vqtbl1q_u8(uint8x16_t __p0,uint8x16_t __p1)54786 __ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
54787   uint8x16_t __ret;
54788   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54789   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54790   __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
54791   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54792   return __ret;
54793 }
54794 #endif
54795 
54796 #ifdef __LITTLE_ENDIAN__
vqtbl1q_s8(int8x16_t __p0,uint8x16_t __p1)54797 __ai int8x16_t vqtbl1q_s8(int8x16_t __p0, uint8x16_t __p1) {
54798   int8x16_t __ret;
54799   __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
54800   return __ret;
54801 }
54802 #else
vqtbl1q_s8(int8x16_t __p0,uint8x16_t __p1)54803 __ai int8x16_t vqtbl1q_s8(int8x16_t __p0, uint8x16_t __p1) {
54804   int8x16_t __ret;
54805   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54806   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54807   __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
54808   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54809   return __ret;
54810 }
54811 #endif
54812 
54813 #ifdef __LITTLE_ENDIAN__
vqtbl1_u8(uint8x16_t __p0,uint8x8_t __p1)54814 __ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
54815   uint8x8_t __ret;
54816   __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 16);
54817   return __ret;
54818 }
54819 #else
vqtbl1_u8(uint8x16_t __p0,uint8x8_t __p1)54820 __ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
54821   uint8x8_t __ret;
54822   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54823   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54824   __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 16);
54825   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54826   return __ret;
54827 }
54828 #endif
54829 
54830 #ifdef __LITTLE_ENDIAN__
vqtbl1_s8(int8x16_t __p0,uint8x8_t __p1)54831 __ai int8x8_t vqtbl1_s8(int8x16_t __p0, uint8x8_t __p1) {
54832   int8x8_t __ret;
54833   __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 0);
54834   return __ret;
54835 }
54836 #else
vqtbl1_s8(int8x16_t __p0,uint8x8_t __p1)54837 __ai int8x8_t vqtbl1_s8(int8x16_t __p0, uint8x8_t __p1) {
54838   int8x8_t __ret;
54839   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54840   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54841   __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 0);
54842   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54843   return __ret;
54844 }
54845 #endif
54846 
54847 #ifdef __LITTLE_ENDIAN__
vqtbl2_p8(poly8x16x2_t __p0,uint8x8_t __p1)54848 __ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
54849   poly8x8_t __ret;
54850   __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 4);
54851   return __ret;
54852 }
54853 #else
vqtbl2_p8(poly8x16x2_t __p0,uint8x8_t __p1)54854 __ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
54855   poly8x8_t __ret;
54856   poly8x16x2_t __rev0;
54857   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54858   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54859   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54860   __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 4);
54861   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54862   return __ret;
54863 }
54864 #endif
54865 
54866 #ifdef __LITTLE_ENDIAN__
vqtbl2q_p8(poly8x16x2_t __p0,uint8x16_t __p1)54867 __ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
54868   poly8x16_t __ret;
54869   __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 36);
54870   return __ret;
54871 }
54872 #else
vqtbl2q_p8(poly8x16x2_t __p0,uint8x16_t __p1)54873 __ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
54874   poly8x16_t __ret;
54875   poly8x16x2_t __rev0;
54876   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54877   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54878   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54879   __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 36);
54880   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54881   return __ret;
54882 }
54883 #endif
54884 
54885 #ifdef __LITTLE_ENDIAN__
vqtbl2q_u8(uint8x16x2_t __p0,uint8x16_t __p1)54886 __ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
54887   uint8x16_t __ret;
54888   __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 48);
54889   return __ret;
54890 }
54891 #else
vqtbl2q_u8(uint8x16x2_t __p0,uint8x16_t __p1)54892 __ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
54893   uint8x16_t __ret;
54894   uint8x16x2_t __rev0;
54895   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54896   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54897   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54898   __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 48);
54899   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54900   return __ret;
54901 }
54902 #endif
54903 
54904 #ifdef __LITTLE_ENDIAN__
vqtbl2q_s8(int8x16x2_t __p0,uint8x16_t __p1)54905 __ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, uint8x16_t __p1) {
54906   int8x16_t __ret;
54907   __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 32);
54908   return __ret;
54909 }
54910 #else
vqtbl2q_s8(int8x16x2_t __p0,uint8x16_t __p1)54911 __ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, uint8x16_t __p1) {
54912   int8x16_t __ret;
54913   int8x16x2_t __rev0;
54914   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54915   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54916   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54917   __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 32);
54918   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54919   return __ret;
54920 }
54921 #endif
54922 
54923 #ifdef __LITTLE_ENDIAN__
vqtbl2_u8(uint8x16x2_t __p0,uint8x8_t __p1)54924 __ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
54925   uint8x8_t __ret;
54926   __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 16);
54927   return __ret;
54928 }
54929 #else
vqtbl2_u8(uint8x16x2_t __p0,uint8x8_t __p1)54930 __ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
54931   uint8x8_t __ret;
54932   uint8x16x2_t __rev0;
54933   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54934   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54935   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54936   __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 16);
54937   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54938   return __ret;
54939 }
54940 #endif
54941 
54942 #ifdef __LITTLE_ENDIAN__
vqtbl2_s8(int8x16x2_t __p0,uint8x8_t __p1)54943 __ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, uint8x8_t __p1) {
54944   int8x8_t __ret;
54945   __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 0);
54946   return __ret;
54947 }
54948 #else
vqtbl2_s8(int8x16x2_t __p0,uint8x8_t __p1)54949 __ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, uint8x8_t __p1) {
54950   int8x8_t __ret;
54951   int8x16x2_t __rev0;
54952   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54953   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54954   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54955   __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 0);
54956   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54957   return __ret;
54958 }
54959 #endif
54960 
54961 #ifdef __LITTLE_ENDIAN__
vqtbl3_p8(poly8x16x3_t __p0,uint8x8_t __p1)54962 __ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
54963   poly8x8_t __ret;
54964   __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 4);
54965   return __ret;
54966 }
54967 #else
vqtbl3_p8(poly8x16x3_t __p0,uint8x8_t __p1)54968 __ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
54969   poly8x8_t __ret;
54970   poly8x16x3_t __rev0;
54971   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54972   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54973   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54974   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54975   __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 4);
54976   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54977   return __ret;
54978 }
54979 #endif
54980 
54981 #ifdef __LITTLE_ENDIAN__
vqtbl3q_p8(poly8x16x3_t __p0,uint8x16_t __p1)54982 __ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
54983   poly8x16_t __ret;
54984   __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 36);
54985   return __ret;
54986 }
54987 #else
vqtbl3q_p8(poly8x16x3_t __p0,uint8x16_t __p1)54988 __ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
54989   poly8x16_t __ret;
54990   poly8x16x3_t __rev0;
54991   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54992   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54993   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54994   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54995   __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 36);
54996   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54997   return __ret;
54998 }
54999 #endif
55000 
55001 #ifdef __LITTLE_ENDIAN__
vqtbl3q_u8(uint8x16x3_t __p0,uint8x16_t __p1)55002 __ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
55003   uint8x16_t __ret;
55004   __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 48);
55005   return __ret;
55006 }
55007 #else
vqtbl3q_u8(uint8x16x3_t __p0,uint8x16_t __p1)55008 __ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
55009   uint8x16_t __ret;
55010   uint8x16x3_t __rev0;
55011   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55012   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55013   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55014   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55015   __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 48);
55016   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55017   return __ret;
55018 }
55019 #endif
55020 
55021 #ifdef __LITTLE_ENDIAN__
vqtbl3q_s8(int8x16x3_t __p0,uint8x16_t __p1)55022 __ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, uint8x16_t __p1) {
55023   int8x16_t __ret;
55024   __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 32);
55025   return __ret;
55026 }
55027 #else
vqtbl3q_s8(int8x16x3_t __p0,uint8x16_t __p1)55028 __ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, uint8x16_t __p1) {
55029   int8x16_t __ret;
55030   int8x16x3_t __rev0;
55031   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55032   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55033   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55034   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55035   __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 32);
55036   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55037   return __ret;
55038 }
55039 #endif
55040 
55041 #ifdef __LITTLE_ENDIAN__
vqtbl3_u8(uint8x16x3_t __p0,uint8x8_t __p1)55042 __ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
55043   uint8x8_t __ret;
55044   __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 16);
55045   return __ret;
55046 }
55047 #else
vqtbl3_u8(uint8x16x3_t __p0,uint8x8_t __p1)55048 __ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
55049   uint8x8_t __ret;
55050   uint8x16x3_t __rev0;
55051   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55052   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55053   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55054   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55055   __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 16);
55056   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55057   return __ret;
55058 }
55059 #endif
55060 
55061 #ifdef __LITTLE_ENDIAN__
vqtbl3_s8(int8x16x3_t __p0,uint8x8_t __p1)55062 __ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, uint8x8_t __p1) {
55063   int8x8_t __ret;
55064   __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 0);
55065   return __ret;
55066 }
55067 #else
vqtbl3_s8(int8x16x3_t __p0,uint8x8_t __p1)55068 __ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, uint8x8_t __p1) {
55069   int8x8_t __ret;
55070   int8x16x3_t __rev0;
55071   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55072   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55073   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55074   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55075   __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 0);
55076   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55077   return __ret;
55078 }
55079 #endif
55080 
55081 #ifdef __LITTLE_ENDIAN__
vqtbl4_p8(poly8x16x4_t __p0,uint8x8_t __p1)55082 __ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
55083   poly8x8_t __ret;
55084   __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 4);
55085   return __ret;
55086 }
55087 #else
vqtbl4_p8(poly8x16x4_t __p0,uint8x8_t __p1)55088 __ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
55089   poly8x8_t __ret;
55090   poly8x16x4_t __rev0;
55091   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55092   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55093   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55094   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55095   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55096   __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 4);
55097   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55098   return __ret;
55099 }
55100 #endif
55101 
55102 #ifdef __LITTLE_ENDIAN__
vqtbl4q_p8(poly8x16x4_t __p0,uint8x16_t __p1)55103 __ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
55104   poly8x16_t __ret;
55105   __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 36);
55106   return __ret;
55107 }
55108 #else
vqtbl4q_p8(poly8x16x4_t __p0,uint8x16_t __p1)55109 __ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
55110   poly8x16_t __ret;
55111   poly8x16x4_t __rev0;
55112   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55113   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55114   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55115   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55116   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55117   __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 36);
55118   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55119   return __ret;
55120 }
55121 #endif
55122 
55123 #ifdef __LITTLE_ENDIAN__
vqtbl4q_u8(uint8x16x4_t __p0,uint8x16_t __p1)55124 __ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
55125   uint8x16_t __ret;
55126   __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 48);
55127   return __ret;
55128 }
55129 #else
vqtbl4q_u8(uint8x16x4_t __p0,uint8x16_t __p1)55130 __ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
55131   uint8x16_t __ret;
55132   uint8x16x4_t __rev0;
55133   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55134   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55135   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55136   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55137   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55138   __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 48);
55139   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55140   return __ret;
55141 }
55142 #endif
55143 
55144 #ifdef __LITTLE_ENDIAN__
vqtbl4q_s8(int8x16x4_t __p0,uint8x16_t __p1)55145 __ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, uint8x16_t __p1) {
55146   int8x16_t __ret;
55147   __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 32);
55148   return __ret;
55149 }
55150 #else
vqtbl4q_s8(int8x16x4_t __p0,uint8x16_t __p1)55151 __ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, uint8x16_t __p1) {
55152   int8x16_t __ret;
55153   int8x16x4_t __rev0;
55154   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55155   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55156   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55157   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55158   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55159   __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 32);
55160   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55161   return __ret;
55162 }
55163 #endif
55164 
55165 #ifdef __LITTLE_ENDIAN__
vqtbl4_u8(uint8x16x4_t __p0,uint8x8_t __p1)55166 __ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
55167   uint8x8_t __ret;
55168   __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 16);
55169   return __ret;
55170 }
55171 #else
vqtbl4_u8(uint8x16x4_t __p0,uint8x8_t __p1)55172 __ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
55173   uint8x8_t __ret;
55174   uint8x16x4_t __rev0;
55175   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55176   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55177   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55178   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55179   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55180   __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 16);
55181   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55182   return __ret;
55183 }
55184 #endif
55185 
55186 #ifdef __LITTLE_ENDIAN__
vqtbl4_s8(int8x16x4_t __p0,uint8x8_t __p1)55187 __ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, uint8x8_t __p1) {
55188   int8x8_t __ret;
55189   __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 0);
55190   return __ret;
55191 }
55192 #else
vqtbl4_s8(int8x16x4_t __p0,uint8x8_t __p1)55193 __ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, uint8x8_t __p1) {
55194   int8x8_t __ret;
55195   int8x16x4_t __rev0;
55196   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55197   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55198   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55199   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55200   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55201   __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 0);
55202   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55203   return __ret;
55204 }
55205 #endif
55206 
55207 #ifdef __LITTLE_ENDIAN__
vqtbx1_p8(poly8x8_t __p0,poly8x16_t __p1,uint8x8_t __p2)55208 __ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
55209   poly8x8_t __ret;
55210   __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 4);
55211   return __ret;
55212 }
55213 #else
vqtbx1_p8(poly8x8_t __p0,poly8x16_t __p1,uint8x8_t __p2)55214 __ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
55215   poly8x8_t __ret;
55216   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55217   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55218   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55219   __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 4);
55220   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55221   return __ret;
55222 }
55223 #endif
55224 
55225 #ifdef __LITTLE_ENDIAN__
vqtbx1q_p8(poly8x16_t __p0,poly8x16_t __p1,uint8x16_t __p2)55226 __ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
55227   poly8x16_t __ret;
55228   __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
55229   return __ret;
55230 }
55231 #else
vqtbx1q_p8(poly8x16_t __p0,poly8x16_t __p1,uint8x16_t __p2)55232 __ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
55233   poly8x16_t __ret;
55234   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55235   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55236   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55237   __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
55238   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55239   return __ret;
55240 }
55241 #endif
55242 
55243 #ifdef __LITTLE_ENDIAN__
vqtbx1q_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)55244 __ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
55245   uint8x16_t __ret;
55246   __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
55247   return __ret;
55248 }
55249 #else
vqtbx1q_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)55250 __ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
55251   uint8x16_t __ret;
55252   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55253   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55254   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55255   __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
55256   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55257   return __ret;
55258 }
55259 #endif
55260 
55261 #ifdef __LITTLE_ENDIAN__
vqtbx1q_s8(int8x16_t __p0,int8x16_t __p1,uint8x16_t __p2)55262 __ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, uint8x16_t __p2) {
55263   int8x16_t __ret;
55264   __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
55265   return __ret;
55266 }
55267 #else
vqtbx1q_s8(int8x16_t __p0,int8x16_t __p1,uint8x16_t __p2)55268 __ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, uint8x16_t __p2) {
55269   int8x16_t __ret;
55270   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55271   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55272   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55273   __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
55274   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55275   return __ret;
55276 }
55277 #endif
55278 
55279 #ifdef __LITTLE_ENDIAN__
vqtbx1_u8(uint8x8_t __p0,uint8x16_t __p1,uint8x8_t __p2)55280 __ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
55281   uint8x8_t __ret;
55282   __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 16);
55283   return __ret;
55284 }
55285 #else
vqtbx1_u8(uint8x8_t __p0,uint8x16_t __p1,uint8x8_t __p2)55286 __ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
55287   uint8x8_t __ret;
55288   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55289   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55290   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55291   __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 16);
55292   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55293   return __ret;
55294 }
55295 #endif
55296 
55297 #ifdef __LITTLE_ENDIAN__
vqtbx1_s8(int8x8_t __p0,int8x16_t __p1,uint8x8_t __p2)55298 __ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, uint8x8_t __p2) {
55299   int8x8_t __ret;
55300   __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 0);
55301   return __ret;
55302 }
55303 #else
vqtbx1_s8(int8x8_t __p0,int8x16_t __p1,uint8x8_t __p2)55304 __ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, uint8x8_t __p2) {
55305   int8x8_t __ret;
55306   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55307   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55308   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55309   __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 0);
55310   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55311   return __ret;
55312 }
55313 #endif
55314 
55315 #ifdef __LITTLE_ENDIAN__
vqtbx2_p8(poly8x8_t __p0,poly8x16x2_t __p1,uint8x8_t __p2)55316 __ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
55317   poly8x8_t __ret;
55318   __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 4);
55319   return __ret;
55320 }
55321 #else
vqtbx2_p8(poly8x8_t __p0,poly8x16x2_t __p1,uint8x8_t __p2)55322 __ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
55323   poly8x8_t __ret;
55324   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55325   poly8x16x2_t __rev1;
55326   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55327   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55328   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55329   __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 4);
55330   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55331   return __ret;
55332 }
55333 #endif
55334 
55335 #ifdef __LITTLE_ENDIAN__
vqtbx2q_p8(poly8x16_t __p0,poly8x16x2_t __p1,uint8x16_t __p2)55336 __ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
55337   poly8x16_t __ret;
55338   __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 36);
55339   return __ret;
55340 }
55341 #else
vqtbx2q_p8(poly8x16_t __p0,poly8x16x2_t __p1,uint8x16_t __p2)55342 __ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
55343   poly8x16_t __ret;
55344   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55345   poly8x16x2_t __rev1;
55346   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55347   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55348   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55349   __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 36);
55350   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55351   return __ret;
55352 }
55353 #endif
55354 
55355 #ifdef __LITTLE_ENDIAN__
vqtbx2q_u8(uint8x16_t __p0,uint8x16x2_t __p1,uint8x16_t __p2)55356 __ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
55357   uint8x16_t __ret;
55358   __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 48);
55359   return __ret;
55360 }
55361 #else
vqtbx2q_u8(uint8x16_t __p0,uint8x16x2_t __p1,uint8x16_t __p2)55362 __ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
55363   uint8x16_t __ret;
55364   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55365   uint8x16x2_t __rev1;
55366   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55367   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55368   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55369   __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 48);
55370   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55371   return __ret;
55372 }
55373 #endif
55374 
55375 #ifdef __LITTLE_ENDIAN__
vqtbx2q_s8(int8x16_t __p0,int8x16x2_t __p1,uint8x16_t __p2)55376 __ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, uint8x16_t __p2) {
55377   int8x16_t __ret;
55378   __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 32);
55379   return __ret;
55380 }
55381 #else
vqtbx2q_s8(int8x16_t __p0,int8x16x2_t __p1,uint8x16_t __p2)55382 __ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, uint8x16_t __p2) {
55383   int8x16_t __ret;
55384   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55385   int8x16x2_t __rev1;
55386   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55387   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55388   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55389   __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 32);
55390   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55391   return __ret;
55392 }
55393 #endif
55394 
55395 #ifdef __LITTLE_ENDIAN__
vqtbx2_u8(uint8x8_t __p0,uint8x16x2_t __p1,uint8x8_t __p2)55396 __ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
55397   uint8x8_t __ret;
55398   __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 16);
55399   return __ret;
55400 }
55401 #else
vqtbx2_u8(uint8x8_t __p0,uint8x16x2_t __p1,uint8x8_t __p2)55402 __ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
55403   uint8x8_t __ret;
55404   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55405   uint8x16x2_t __rev1;
55406   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55407   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55408   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55409   __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 16);
55410   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55411   return __ret;
55412 }
55413 #endif
55414 
55415 #ifdef __LITTLE_ENDIAN__
vqtbx2_s8(int8x8_t __p0,int8x16x2_t __p1,uint8x8_t __p2)55416 __ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, uint8x8_t __p2) {
55417   int8x8_t __ret;
55418   __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 0);
55419   return __ret;
55420 }
55421 #else
vqtbx2_s8(int8x8_t __p0,int8x16x2_t __p1,uint8x8_t __p2)55422 __ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, uint8x8_t __p2) {
55423   int8x8_t __ret;
55424   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55425   int8x16x2_t __rev1;
55426   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55427   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55428   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55429   __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 0);
55430   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55431   return __ret;
55432 }
55433 #endif
55434 
55435 #ifdef __LITTLE_ENDIAN__
vqtbx3_p8(poly8x8_t __p0,poly8x16x3_t __p1,uint8x8_t __p2)55436 __ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
55437   poly8x8_t __ret;
55438   __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 4);
55439   return __ret;
55440 }
55441 #else
vqtbx3_p8(poly8x8_t __p0,poly8x16x3_t __p1,uint8x8_t __p2)55442 __ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
55443   poly8x8_t __ret;
55444   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55445   poly8x16x3_t __rev1;
55446   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55447   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55448   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55449   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55450   __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 4);
55451   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55452   return __ret;
55453 }
55454 #endif
55455 
55456 #ifdef __LITTLE_ENDIAN__
vqtbx3q_p8(poly8x16_t __p0,poly8x16x3_t __p1,uint8x16_t __p2)55457 __ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
55458   poly8x16_t __ret;
55459   __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 36);
55460   return __ret;
55461 }
55462 #else
vqtbx3q_p8(poly8x16_t __p0,poly8x16x3_t __p1,uint8x16_t __p2)55463 __ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
55464   poly8x16_t __ret;
55465   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55466   poly8x16x3_t __rev1;
55467   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55468   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55469   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55470   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55471   __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 36);
55472   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55473   return __ret;
55474 }
55475 #endif
55476 
55477 #ifdef __LITTLE_ENDIAN__
vqtbx3q_u8(uint8x16_t __p0,uint8x16x3_t __p1,uint8x16_t __p2)55478 __ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
55479   uint8x16_t __ret;
55480   __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 48);
55481   return __ret;
55482 }
55483 #else
vqtbx3q_u8(uint8x16_t __p0,uint8x16x3_t __p1,uint8x16_t __p2)55484 __ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
55485   uint8x16_t __ret;
55486   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55487   uint8x16x3_t __rev1;
55488   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55489   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55490   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55491   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55492   __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 48);
55493   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55494   return __ret;
55495 }
55496 #endif
55497 
55498 #ifdef __LITTLE_ENDIAN__
vqtbx3q_s8(int8x16_t __p0,int8x16x3_t __p1,uint8x16_t __p2)55499 __ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, uint8x16_t __p2) {
55500   int8x16_t __ret;
55501   __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 32);
55502   return __ret;
55503 }
55504 #else
vqtbx3q_s8(int8x16_t __p0,int8x16x3_t __p1,uint8x16_t __p2)55505 __ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, uint8x16_t __p2) {
55506   int8x16_t __ret;
55507   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55508   int8x16x3_t __rev1;
55509   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55510   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55511   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55512   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55513   __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 32);
55514   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55515   return __ret;
55516 }
55517 #endif
55518 
55519 #ifdef __LITTLE_ENDIAN__
vqtbx3_u8(uint8x8_t __p0,uint8x16x3_t __p1,uint8x8_t __p2)55520 __ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
55521   uint8x8_t __ret;
55522   __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 16);
55523   return __ret;
55524 }
55525 #else
vqtbx3_u8(uint8x8_t __p0,uint8x16x3_t __p1,uint8x8_t __p2)55526 __ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
55527   uint8x8_t __ret;
55528   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55529   uint8x16x3_t __rev1;
55530   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55531   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55532   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55533   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55534   __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 16);
55535   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55536   return __ret;
55537 }
55538 #endif
55539 
55540 #ifdef __LITTLE_ENDIAN__
vqtbx3_s8(int8x8_t __p0,int8x16x3_t __p1,uint8x8_t __p2)55541 __ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, uint8x8_t __p2) {
55542   int8x8_t __ret;
55543   __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 0);
55544   return __ret;
55545 }
55546 #else
vqtbx3_s8(int8x8_t __p0,int8x16x3_t __p1,uint8x8_t __p2)55547 __ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, uint8x8_t __p2) {
55548   int8x8_t __ret;
55549   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55550   int8x16x3_t __rev1;
55551   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55552   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55553   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55554   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55555   __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 0);
55556   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55557   return __ret;
55558 }
55559 #endif
55560 
55561 #ifdef __LITTLE_ENDIAN__
vqtbx4_p8(poly8x8_t __p0,poly8x16x4_t __p1,uint8x8_t __p2)55562 __ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
55563   poly8x8_t __ret;
55564   __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 4);
55565   return __ret;
55566 }
55567 #else
vqtbx4_p8(poly8x8_t __p0,poly8x16x4_t __p1,uint8x8_t __p2)55568 __ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
55569   poly8x8_t __ret;
55570   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55571   poly8x16x4_t __rev1;
55572   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55573   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55574   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55575   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55576   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55577   __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 4);
55578   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55579   return __ret;
55580 }
55581 #endif
55582 
55583 #ifdef __LITTLE_ENDIAN__
vqtbx4q_p8(poly8x16_t __p0,poly8x16x4_t __p1,uint8x16_t __p2)55584 __ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
55585   poly8x16_t __ret;
55586   __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 36);
55587   return __ret;
55588 }
55589 #else
vqtbx4q_p8(poly8x16_t __p0,poly8x16x4_t __p1,uint8x16_t __p2)55590 __ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
55591   poly8x16_t __ret;
55592   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55593   poly8x16x4_t __rev1;
55594   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55595   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55596   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55597   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55598   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55599   __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 36);
55600   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55601   return __ret;
55602 }
55603 #endif
55604 
55605 #ifdef __LITTLE_ENDIAN__
vqtbx4q_u8(uint8x16_t __p0,uint8x16x4_t __p1,uint8x16_t __p2)55606 __ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
55607   uint8x16_t __ret;
55608   __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 48);
55609   return __ret;
55610 }
55611 #else
vqtbx4q_u8(uint8x16_t __p0,uint8x16x4_t __p1,uint8x16_t __p2)55612 __ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
55613   uint8x16_t __ret;
55614   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55615   uint8x16x4_t __rev1;
55616   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55617   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55618   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55619   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55620   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55621   __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 48);
55622   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55623   return __ret;
55624 }
55625 #endif
55626 
55627 #ifdef __LITTLE_ENDIAN__
vqtbx4q_s8(int8x16_t __p0,int8x16x4_t __p1,uint8x16_t __p2)55628 __ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, uint8x16_t __p2) {
55629   int8x16_t __ret;
55630   __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 32);
55631   return __ret;
55632 }
55633 #else
vqtbx4q_s8(int8x16_t __p0,int8x16x4_t __p1,uint8x16_t __p2)55634 __ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, uint8x16_t __p2) {
55635   int8x16_t __ret;
55636   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55637   int8x16x4_t __rev1;
55638   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55639   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55640   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55641   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55642   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55643   __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 32);
55644   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55645   return __ret;
55646 }
55647 #endif
55648 
55649 #ifdef __LITTLE_ENDIAN__
vqtbx4_u8(uint8x8_t __p0,uint8x16x4_t __p1,uint8x8_t __p2)55650 __ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
55651   uint8x8_t __ret;
55652   __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 16);
55653   return __ret;
55654 }
55655 #else
vqtbx4_u8(uint8x8_t __p0,uint8x16x4_t __p1,uint8x8_t __p2)55656 __ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
55657   uint8x8_t __ret;
55658   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55659   uint8x16x4_t __rev1;
55660   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55661   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55662   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55663   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55664   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55665   __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 16);
55666   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55667   return __ret;
55668 }
55669 #endif
55670 
55671 #ifdef __LITTLE_ENDIAN__
vqtbx4_s8(int8x8_t __p0,int8x16x4_t __p1,uint8x8_t __p2)55672 __ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, uint8x8_t __p2) {
55673   int8x8_t __ret;
55674   __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 0);
55675   return __ret;
55676 }
55677 #else
vqtbx4_s8(int8x8_t __p0,int8x16x4_t __p1,uint8x8_t __p2)55678 __ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, uint8x8_t __p2) {
55679   int8x8_t __ret;
55680   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55681   int8x16x4_t __rev1;
55682   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55683   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55684   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55685   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55686   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55687   __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 0);
55688   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55689   return __ret;
55690 }
55691 #endif
55692 
55693 #ifdef __LITTLE_ENDIAN__
vraddhn_high_u32(uint16x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)55694 __ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
55695   uint16x8_t __ret;
55696   __ret = vcombine_u16(__p0, vraddhn_u32(__p1, __p2));
55697   return __ret;
55698 }
55699 #else
vraddhn_high_u32(uint16x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)55700 __ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
55701   uint16x8_t __ret;
55702   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55703   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55704   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
55705   __ret = __noswap_vcombine_u16(__rev0, __noswap_vraddhn_u32(__rev1, __rev2));
55706   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55707   return __ret;
55708 }
55709 #endif
55710 
55711 #ifdef __LITTLE_ENDIAN__
vraddhn_high_u64(uint32x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)55712 __ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
55713   uint32x4_t __ret;
55714   __ret = vcombine_u32(__p0, vraddhn_u64(__p1, __p2));
55715   return __ret;
55716 }
55717 #else
vraddhn_high_u64(uint32x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)55718 __ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
55719   uint32x4_t __ret;
55720   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55721   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55722   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
55723   __ret = __noswap_vcombine_u32(__rev0, __noswap_vraddhn_u64(__rev1, __rev2));
55724   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55725   return __ret;
55726 }
55727 #endif
55728 
55729 #ifdef __LITTLE_ENDIAN__
vraddhn_high_u16(uint8x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)55730 __ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
55731   uint8x16_t __ret;
55732   __ret = vcombine_u8(__p0, vraddhn_u16(__p1, __p2));
55733   return __ret;
55734 }
55735 #else
vraddhn_high_u16(uint8x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)55736 __ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
55737   uint8x16_t __ret;
55738   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55739   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55740   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55741   __ret = __noswap_vcombine_u8(__rev0, __noswap_vraddhn_u16(__rev1, __rev2));
55742   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55743   return __ret;
55744 }
55745 #endif
55746 
55747 #ifdef __LITTLE_ENDIAN__
vraddhn_high_s32(int16x4_t __p0,int32x4_t __p1,int32x4_t __p2)55748 __ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
55749   int16x8_t __ret;
55750   __ret = vcombine_s16(__p0, vraddhn_s32(__p1, __p2));
55751   return __ret;
55752 }
55753 #else
vraddhn_high_s32(int16x4_t __p0,int32x4_t __p1,int32x4_t __p2)55754 __ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
55755   int16x8_t __ret;
55756   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55757   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55758   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
55759   __ret = __noswap_vcombine_s16(__rev0, __noswap_vraddhn_s32(__rev1, __rev2));
55760   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55761   return __ret;
55762 }
55763 #endif
55764 
55765 #ifdef __LITTLE_ENDIAN__
vraddhn_high_s64(int32x2_t __p0,int64x2_t __p1,int64x2_t __p2)55766 __ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
55767   int32x4_t __ret;
55768   __ret = vcombine_s32(__p0, vraddhn_s64(__p1, __p2));
55769   return __ret;
55770 }
55771 #else
vraddhn_high_s64(int32x2_t __p0,int64x2_t __p1,int64x2_t __p2)55772 __ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
55773   int32x4_t __ret;
55774   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55775   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55776   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
55777   __ret = __noswap_vcombine_s32(__rev0, __noswap_vraddhn_s64(__rev1, __rev2));
55778   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55779   return __ret;
55780 }
55781 #endif
55782 
55783 #ifdef __LITTLE_ENDIAN__
vraddhn_high_s16(int8x8_t __p0,int16x8_t __p1,int16x8_t __p2)55784 __ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
55785   int8x16_t __ret;
55786   __ret = vcombine_s8(__p0, vraddhn_s16(__p1, __p2));
55787   return __ret;
55788 }
55789 #else
vraddhn_high_s16(int8x8_t __p0,int16x8_t __p1,int16x8_t __p2)55790 __ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
55791   int8x16_t __ret;
55792   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55793   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55794   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55795   __ret = __noswap_vcombine_s8(__rev0, __noswap_vraddhn_s16(__rev1, __rev2));
55796   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55797   return __ret;
55798 }
55799 #endif
55800 
55801 #ifdef __LITTLE_ENDIAN__
vrbit_p8(poly8x8_t __p0)55802 __ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
55803   poly8x8_t __ret;
55804   __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 4);
55805   return __ret;
55806 }
55807 #else
vrbit_p8(poly8x8_t __p0)55808 __ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
55809   poly8x8_t __ret;
55810   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55811   __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 4);
55812   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55813   return __ret;
55814 }
55815 #endif
55816 
55817 #ifdef __LITTLE_ENDIAN__
vrbitq_p8(poly8x16_t __p0)55818 __ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
55819   poly8x16_t __ret;
55820   __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 36);
55821   return __ret;
55822 }
55823 #else
vrbitq_p8(poly8x16_t __p0)55824 __ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
55825   poly8x16_t __ret;
55826   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55827   __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 36);
55828   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55829   return __ret;
55830 }
55831 #endif
55832 
55833 #ifdef __LITTLE_ENDIAN__
vrbitq_u8(uint8x16_t __p0)55834 __ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
55835   uint8x16_t __ret;
55836   __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 48);
55837   return __ret;
55838 }
55839 #else
vrbitq_u8(uint8x16_t __p0)55840 __ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
55841   uint8x16_t __ret;
55842   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55843   __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 48);
55844   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55845   return __ret;
55846 }
55847 #endif
55848 
55849 #ifdef __LITTLE_ENDIAN__
vrbitq_s8(int8x16_t __p0)55850 __ai int8x16_t vrbitq_s8(int8x16_t __p0) {
55851   int8x16_t __ret;
55852   __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 32);
55853   return __ret;
55854 }
55855 #else
vrbitq_s8(int8x16_t __p0)55856 __ai int8x16_t vrbitq_s8(int8x16_t __p0) {
55857   int8x16_t __ret;
55858   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55859   __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 32);
55860   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55861   return __ret;
55862 }
55863 #endif
55864 
55865 #ifdef __LITTLE_ENDIAN__
vrbit_u8(uint8x8_t __p0)55866 __ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
55867   uint8x8_t __ret;
55868   __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 16);
55869   return __ret;
55870 }
55871 #else
vrbit_u8(uint8x8_t __p0)55872 __ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
55873   uint8x8_t __ret;
55874   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55875   __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 16);
55876   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55877   return __ret;
55878 }
55879 #endif
55880 
55881 #ifdef __LITTLE_ENDIAN__
vrbit_s8(int8x8_t __p0)55882 __ai int8x8_t vrbit_s8(int8x8_t __p0) {
55883   int8x8_t __ret;
55884   __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 0);
55885   return __ret;
55886 }
55887 #else
vrbit_s8(int8x8_t __p0)55888 __ai int8x8_t vrbit_s8(int8x8_t __p0) {
55889   int8x8_t __ret;
55890   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55891   __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 0);
55892   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55893   return __ret;
55894 }
55895 #endif
55896 
55897 #ifdef __LITTLE_ENDIAN__
vrecpeq_f64(float64x2_t __p0)55898 __ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
55899   float64x2_t __ret;
55900   __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 42);
55901   return __ret;
55902 }
55903 #else
vrecpeq_f64(float64x2_t __p0)55904 __ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
55905   float64x2_t __ret;
55906   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55907   __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 42);
55908   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55909   return __ret;
55910 }
55911 #endif
55912 
vrecpe_f64(float64x1_t __p0)55913 __ai float64x1_t vrecpe_f64(float64x1_t __p0) {
55914   float64x1_t __ret;
55915   __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
55916   return __ret;
55917 }
vrecped_f64(float64_t __p0)55918 __ai float64_t vrecped_f64(float64_t __p0) {
55919   float64_t __ret;
55920   __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
55921   return __ret;
55922 }
vrecpes_f32(float32_t __p0)55923 __ai float32_t vrecpes_f32(float32_t __p0) {
55924   float32_t __ret;
55925   __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
55926   return __ret;
55927 }
55928 #ifdef __LITTLE_ENDIAN__
vrecpsq_f64(float64x2_t __p0,float64x2_t __p1)55929 __ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
55930   float64x2_t __ret;
55931   __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55932   return __ret;
55933 }
55934 #else
vrecpsq_f64(float64x2_t __p0,float64x2_t __p1)55935 __ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
55936   float64x2_t __ret;
55937   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55938   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55939   __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55940   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55941   return __ret;
55942 }
55943 #endif
55944 
vrecps_f64(float64x1_t __p0,float64x1_t __p1)55945 __ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
55946   float64x1_t __ret;
55947   __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
55948   return __ret;
55949 }
vrecpsd_f64(float64_t __p0,float64_t __p1)55950 __ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
55951   float64_t __ret;
55952   __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
55953   return __ret;
55954 }
vrecpss_f32(float32_t __p0,float32_t __p1)55955 __ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
55956   float32_t __ret;
55957   __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
55958   return __ret;
55959 }
vrecpxd_f64(float64_t __p0)55960 __ai float64_t vrecpxd_f64(float64_t __p0) {
55961   float64_t __ret;
55962   __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
55963   return __ret;
55964 }
vrecpxs_f32(float32_t __p0)55965 __ai float32_t vrecpxs_f32(float32_t __p0) {
55966   float32_t __ret;
55967   __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
55968   return __ret;
55969 }
vreinterpret_p8_p64(poly64x1_t __p0)55970 __ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
55971   poly8x8_t __ret;
55972   __ret = (poly8x8_t)(__p0);
55973   return __ret;
55974 }
vreinterpret_p8_p16(poly16x4_t __p0)55975 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
55976   poly8x8_t __ret;
55977   __ret = (poly8x8_t)(__p0);
55978   return __ret;
55979 }
vreinterpret_p8_u8(uint8x8_t __p0)55980 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
55981   poly8x8_t __ret;
55982   __ret = (poly8x8_t)(__p0);
55983   return __ret;
55984 }
vreinterpret_p8_u32(uint32x2_t __p0)55985 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
55986   poly8x8_t __ret;
55987   __ret = (poly8x8_t)(__p0);
55988   return __ret;
55989 }
vreinterpret_p8_u64(uint64x1_t __p0)55990 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
55991   poly8x8_t __ret;
55992   __ret = (poly8x8_t)(__p0);
55993   return __ret;
55994 }
vreinterpret_p8_u16(uint16x4_t __p0)55995 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
55996   poly8x8_t __ret;
55997   __ret = (poly8x8_t)(__p0);
55998   return __ret;
55999 }
vreinterpret_p8_s8(int8x8_t __p0)56000 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
56001   poly8x8_t __ret;
56002   __ret = (poly8x8_t)(__p0);
56003   return __ret;
56004 }
vreinterpret_p8_f64(float64x1_t __p0)56005 __ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
56006   poly8x8_t __ret;
56007   __ret = (poly8x8_t)(__p0);
56008   return __ret;
56009 }
vreinterpret_p8_f32(float32x2_t __p0)56010 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
56011   poly8x8_t __ret;
56012   __ret = (poly8x8_t)(__p0);
56013   return __ret;
56014 }
vreinterpret_p8_f16(float16x4_t __p0)56015 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
56016   poly8x8_t __ret;
56017   __ret = (poly8x8_t)(__p0);
56018   return __ret;
56019 }
vreinterpret_p8_s32(int32x2_t __p0)56020 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
56021   poly8x8_t __ret;
56022   __ret = (poly8x8_t)(__p0);
56023   return __ret;
56024 }
vreinterpret_p8_s64(int64x1_t __p0)56025 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
56026   poly8x8_t __ret;
56027   __ret = (poly8x8_t)(__p0);
56028   return __ret;
56029 }
vreinterpret_p8_s16(int16x4_t __p0)56030 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
56031   poly8x8_t __ret;
56032   __ret = (poly8x8_t)(__p0);
56033   return __ret;
56034 }
vreinterpret_p64_p8(poly8x8_t __p0)56035 __ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
56036   poly64x1_t __ret;
56037   __ret = (poly64x1_t)(__p0);
56038   return __ret;
56039 }
vreinterpret_p64_p16(poly16x4_t __p0)56040 __ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
56041   poly64x1_t __ret;
56042   __ret = (poly64x1_t)(__p0);
56043   return __ret;
56044 }
vreinterpret_p64_u8(uint8x8_t __p0)56045 __ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
56046   poly64x1_t __ret;
56047   __ret = (poly64x1_t)(__p0);
56048   return __ret;
56049 }
vreinterpret_p64_u32(uint32x2_t __p0)56050 __ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
56051   poly64x1_t __ret;
56052   __ret = (poly64x1_t)(__p0);
56053   return __ret;
56054 }
vreinterpret_p64_u64(uint64x1_t __p0)56055 __ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
56056   poly64x1_t __ret;
56057   __ret = (poly64x1_t)(__p0);
56058   return __ret;
56059 }
vreinterpret_p64_u16(uint16x4_t __p0)56060 __ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
56061   poly64x1_t __ret;
56062   __ret = (poly64x1_t)(__p0);
56063   return __ret;
56064 }
vreinterpret_p64_s8(int8x8_t __p0)56065 __ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
56066   poly64x1_t __ret;
56067   __ret = (poly64x1_t)(__p0);
56068   return __ret;
56069 }
vreinterpret_p64_f64(float64x1_t __p0)56070 __ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
56071   poly64x1_t __ret;
56072   __ret = (poly64x1_t)(__p0);
56073   return __ret;
56074 }
vreinterpret_p64_f32(float32x2_t __p0)56075 __ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
56076   poly64x1_t __ret;
56077   __ret = (poly64x1_t)(__p0);
56078   return __ret;
56079 }
vreinterpret_p64_f16(float16x4_t __p0)56080 __ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
56081   poly64x1_t __ret;
56082   __ret = (poly64x1_t)(__p0);
56083   return __ret;
56084 }
vreinterpret_p64_s32(int32x2_t __p0)56085 __ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
56086   poly64x1_t __ret;
56087   __ret = (poly64x1_t)(__p0);
56088   return __ret;
56089 }
vreinterpret_p64_s64(int64x1_t __p0)56090 __ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
56091   poly64x1_t __ret;
56092   __ret = (poly64x1_t)(__p0);
56093   return __ret;
56094 }
vreinterpret_p64_s16(int16x4_t __p0)56095 __ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
56096   poly64x1_t __ret;
56097   __ret = (poly64x1_t)(__p0);
56098   return __ret;
56099 }
vreinterpret_p16_p8(poly8x8_t __p0)56100 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
56101   poly16x4_t __ret;
56102   __ret = (poly16x4_t)(__p0);
56103   return __ret;
56104 }
vreinterpret_p16_p64(poly64x1_t __p0)56105 __ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
56106   poly16x4_t __ret;
56107   __ret = (poly16x4_t)(__p0);
56108   return __ret;
56109 }
vreinterpret_p16_u8(uint8x8_t __p0)56110 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
56111   poly16x4_t __ret;
56112   __ret = (poly16x4_t)(__p0);
56113   return __ret;
56114 }
vreinterpret_p16_u32(uint32x2_t __p0)56115 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
56116   poly16x4_t __ret;
56117   __ret = (poly16x4_t)(__p0);
56118   return __ret;
56119 }
vreinterpret_p16_u64(uint64x1_t __p0)56120 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
56121   poly16x4_t __ret;
56122   __ret = (poly16x4_t)(__p0);
56123   return __ret;
56124 }
vreinterpret_p16_u16(uint16x4_t __p0)56125 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
56126   poly16x4_t __ret;
56127   __ret = (poly16x4_t)(__p0);
56128   return __ret;
56129 }
vreinterpret_p16_s8(int8x8_t __p0)56130 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
56131   poly16x4_t __ret;
56132   __ret = (poly16x4_t)(__p0);
56133   return __ret;
56134 }
vreinterpret_p16_f64(float64x1_t __p0)56135 __ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
56136   poly16x4_t __ret;
56137   __ret = (poly16x4_t)(__p0);
56138   return __ret;
56139 }
vreinterpret_p16_f32(float32x2_t __p0)56140 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
56141   poly16x4_t __ret;
56142   __ret = (poly16x4_t)(__p0);
56143   return __ret;
56144 }
vreinterpret_p16_f16(float16x4_t __p0)56145 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
56146   poly16x4_t __ret;
56147   __ret = (poly16x4_t)(__p0);
56148   return __ret;
56149 }
vreinterpret_p16_s32(int32x2_t __p0)56150 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
56151   poly16x4_t __ret;
56152   __ret = (poly16x4_t)(__p0);
56153   return __ret;
56154 }
vreinterpret_p16_s64(int64x1_t __p0)56155 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
56156   poly16x4_t __ret;
56157   __ret = (poly16x4_t)(__p0);
56158   return __ret;
56159 }
vreinterpret_p16_s16(int16x4_t __p0)56160 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
56161   poly16x4_t __ret;
56162   __ret = (poly16x4_t)(__p0);
56163   return __ret;
56164 }
vreinterpretq_p8_p128(poly128_t __p0)56165 __ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
56166   poly8x16_t __ret;
56167   __ret = (poly8x16_t)(__p0);
56168   return __ret;
56169 }
vreinterpretq_p8_p64(poly64x2_t __p0)56170 __ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
56171   poly8x16_t __ret;
56172   __ret = (poly8x16_t)(__p0);
56173   return __ret;
56174 }
vreinterpretq_p8_p16(poly16x8_t __p0)56175 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
56176   poly8x16_t __ret;
56177   __ret = (poly8x16_t)(__p0);
56178   return __ret;
56179 }
vreinterpretq_p8_u8(uint8x16_t __p0)56180 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
56181   poly8x16_t __ret;
56182   __ret = (poly8x16_t)(__p0);
56183   return __ret;
56184 }
vreinterpretq_p8_u32(uint32x4_t __p0)56185 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
56186   poly8x16_t __ret;
56187   __ret = (poly8x16_t)(__p0);
56188   return __ret;
56189 }
vreinterpretq_p8_u64(uint64x2_t __p0)56190 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
56191   poly8x16_t __ret;
56192   __ret = (poly8x16_t)(__p0);
56193   return __ret;
56194 }
vreinterpretq_p8_u16(uint16x8_t __p0)56195 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
56196   poly8x16_t __ret;
56197   __ret = (poly8x16_t)(__p0);
56198   return __ret;
56199 }
vreinterpretq_p8_s8(int8x16_t __p0)56200 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
56201   poly8x16_t __ret;
56202   __ret = (poly8x16_t)(__p0);
56203   return __ret;
56204 }
vreinterpretq_p8_f64(float64x2_t __p0)56205 __ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
56206   poly8x16_t __ret;
56207   __ret = (poly8x16_t)(__p0);
56208   return __ret;
56209 }
vreinterpretq_p8_f32(float32x4_t __p0)56210 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
56211   poly8x16_t __ret;
56212   __ret = (poly8x16_t)(__p0);
56213   return __ret;
56214 }
vreinterpretq_p8_f16(float16x8_t __p0)56215 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
56216   poly8x16_t __ret;
56217   __ret = (poly8x16_t)(__p0);
56218   return __ret;
56219 }
vreinterpretq_p8_s32(int32x4_t __p0)56220 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
56221   poly8x16_t __ret;
56222   __ret = (poly8x16_t)(__p0);
56223   return __ret;
56224 }
vreinterpretq_p8_s64(int64x2_t __p0)56225 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
56226   poly8x16_t __ret;
56227   __ret = (poly8x16_t)(__p0);
56228   return __ret;
56229 }
vreinterpretq_p8_s16(int16x8_t __p0)56230 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
56231   poly8x16_t __ret;
56232   __ret = (poly8x16_t)(__p0);
56233   return __ret;
56234 }
vreinterpretq_p128_p8(poly8x16_t __p0)56235 __ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
56236   poly128_t __ret;
56237   __ret = (poly128_t)(__p0);
56238   return __ret;
56239 }
vreinterpretq_p128_p64(poly64x2_t __p0)56240 __ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
56241   poly128_t __ret;
56242   __ret = (poly128_t)(__p0);
56243   return __ret;
56244 }
vreinterpretq_p128_p16(poly16x8_t __p0)56245 __ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
56246   poly128_t __ret;
56247   __ret = (poly128_t)(__p0);
56248   return __ret;
56249 }
vreinterpretq_p128_u8(uint8x16_t __p0)56250 __ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
56251   poly128_t __ret;
56252   __ret = (poly128_t)(__p0);
56253   return __ret;
56254 }
vreinterpretq_p128_u32(uint32x4_t __p0)56255 __ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
56256   poly128_t __ret;
56257   __ret = (poly128_t)(__p0);
56258   return __ret;
56259 }
vreinterpretq_p128_u64(uint64x2_t __p0)56260 __ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
56261   poly128_t __ret;
56262   __ret = (poly128_t)(__p0);
56263   return __ret;
56264 }
vreinterpretq_p128_u16(uint16x8_t __p0)56265 __ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
56266   poly128_t __ret;
56267   __ret = (poly128_t)(__p0);
56268   return __ret;
56269 }
vreinterpretq_p128_s8(int8x16_t __p0)56270 __ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
56271   poly128_t __ret;
56272   __ret = (poly128_t)(__p0);
56273   return __ret;
56274 }
vreinterpretq_p128_f64(float64x2_t __p0)56275 __ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
56276   poly128_t __ret;
56277   __ret = (poly128_t)(__p0);
56278   return __ret;
56279 }
vreinterpretq_p128_f32(float32x4_t __p0)56280 __ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
56281   poly128_t __ret;
56282   __ret = (poly128_t)(__p0);
56283   return __ret;
56284 }
vreinterpretq_p128_f16(float16x8_t __p0)56285 __ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
56286   poly128_t __ret;
56287   __ret = (poly128_t)(__p0);
56288   return __ret;
56289 }
vreinterpretq_p128_s32(int32x4_t __p0)56290 __ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
56291   poly128_t __ret;
56292   __ret = (poly128_t)(__p0);
56293   return __ret;
56294 }
vreinterpretq_p128_s64(int64x2_t __p0)56295 __ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
56296   poly128_t __ret;
56297   __ret = (poly128_t)(__p0);
56298   return __ret;
56299 }
vreinterpretq_p128_s16(int16x8_t __p0)56300 __ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
56301   poly128_t __ret;
56302   __ret = (poly128_t)(__p0);
56303   return __ret;
56304 }
vreinterpretq_p64_p8(poly8x16_t __p0)56305 __ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
56306   poly64x2_t __ret;
56307   __ret = (poly64x2_t)(__p0);
56308   return __ret;
56309 }
vreinterpretq_p64_p128(poly128_t __p0)56310 __ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
56311   poly64x2_t __ret;
56312   __ret = (poly64x2_t)(__p0);
56313   return __ret;
56314 }
vreinterpretq_p64_p16(poly16x8_t __p0)56315 __ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
56316   poly64x2_t __ret;
56317   __ret = (poly64x2_t)(__p0);
56318   return __ret;
56319 }
vreinterpretq_p64_u8(uint8x16_t __p0)56320 __ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
56321   poly64x2_t __ret;
56322   __ret = (poly64x2_t)(__p0);
56323   return __ret;
56324 }
vreinterpretq_p64_u32(uint32x4_t __p0)56325 __ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
56326   poly64x2_t __ret;
56327   __ret = (poly64x2_t)(__p0);
56328   return __ret;
56329 }
vreinterpretq_p64_u64(uint64x2_t __p0)56330 __ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
56331   poly64x2_t __ret;
56332   __ret = (poly64x2_t)(__p0);
56333   return __ret;
56334 }
vreinterpretq_p64_u16(uint16x8_t __p0)56335 __ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
56336   poly64x2_t __ret;
56337   __ret = (poly64x2_t)(__p0);
56338   return __ret;
56339 }
vreinterpretq_p64_s8(int8x16_t __p0)56340 __ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
56341   poly64x2_t __ret;
56342   __ret = (poly64x2_t)(__p0);
56343   return __ret;
56344 }
vreinterpretq_p64_f64(float64x2_t __p0)56345 __ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
56346   poly64x2_t __ret;
56347   __ret = (poly64x2_t)(__p0);
56348   return __ret;
56349 }
vreinterpretq_p64_f32(float32x4_t __p0)56350 __ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
56351   poly64x2_t __ret;
56352   __ret = (poly64x2_t)(__p0);
56353   return __ret;
56354 }
vreinterpretq_p64_f16(float16x8_t __p0)56355 __ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
56356   poly64x2_t __ret;
56357   __ret = (poly64x2_t)(__p0);
56358   return __ret;
56359 }
vreinterpretq_p64_s32(int32x4_t __p0)56360 __ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
56361   poly64x2_t __ret;
56362   __ret = (poly64x2_t)(__p0);
56363   return __ret;
56364 }
vreinterpretq_p64_s64(int64x2_t __p0)56365 __ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
56366   poly64x2_t __ret;
56367   __ret = (poly64x2_t)(__p0);
56368   return __ret;
56369 }
vreinterpretq_p64_s16(int16x8_t __p0)56370 __ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
56371   poly64x2_t __ret;
56372   __ret = (poly64x2_t)(__p0);
56373   return __ret;
56374 }
vreinterpretq_p16_p8(poly8x16_t __p0)56375 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
56376   poly16x8_t __ret;
56377   __ret = (poly16x8_t)(__p0);
56378   return __ret;
56379 }
vreinterpretq_p16_p128(poly128_t __p0)56380 __ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
56381   poly16x8_t __ret;
56382   __ret = (poly16x8_t)(__p0);
56383   return __ret;
56384 }
vreinterpretq_p16_p64(poly64x2_t __p0)56385 __ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
56386   poly16x8_t __ret;
56387   __ret = (poly16x8_t)(__p0);
56388   return __ret;
56389 }
vreinterpretq_p16_u8(uint8x16_t __p0)56390 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
56391   poly16x8_t __ret;
56392   __ret = (poly16x8_t)(__p0);
56393   return __ret;
56394 }
vreinterpretq_p16_u32(uint32x4_t __p0)56395 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
56396   poly16x8_t __ret;
56397   __ret = (poly16x8_t)(__p0);
56398   return __ret;
56399 }
vreinterpretq_p16_u64(uint64x2_t __p0)56400 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
56401   poly16x8_t __ret;
56402   __ret = (poly16x8_t)(__p0);
56403   return __ret;
56404 }
vreinterpretq_p16_u16(uint16x8_t __p0)56405 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
56406   poly16x8_t __ret;
56407   __ret = (poly16x8_t)(__p0);
56408   return __ret;
56409 }
vreinterpretq_p16_s8(int8x16_t __p0)56410 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
56411   poly16x8_t __ret;
56412   __ret = (poly16x8_t)(__p0);
56413   return __ret;
56414 }
vreinterpretq_p16_f64(float64x2_t __p0)56415 __ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
56416   poly16x8_t __ret;
56417   __ret = (poly16x8_t)(__p0);
56418   return __ret;
56419 }
vreinterpretq_p16_f32(float32x4_t __p0)56420 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
56421   poly16x8_t __ret;
56422   __ret = (poly16x8_t)(__p0);
56423   return __ret;
56424 }
vreinterpretq_p16_f16(float16x8_t __p0)56425 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
56426   poly16x8_t __ret;
56427   __ret = (poly16x8_t)(__p0);
56428   return __ret;
56429 }
vreinterpretq_p16_s32(int32x4_t __p0)56430 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
56431   poly16x8_t __ret;
56432   __ret = (poly16x8_t)(__p0);
56433   return __ret;
56434 }
vreinterpretq_p16_s64(int64x2_t __p0)56435 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
56436   poly16x8_t __ret;
56437   __ret = (poly16x8_t)(__p0);
56438   return __ret;
56439 }
vreinterpretq_p16_s16(int16x8_t __p0)56440 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
56441   poly16x8_t __ret;
56442   __ret = (poly16x8_t)(__p0);
56443   return __ret;
56444 }
vreinterpretq_u8_p8(poly8x16_t __p0)56445 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
56446   uint8x16_t __ret;
56447   __ret = (uint8x16_t)(__p0);
56448   return __ret;
56449 }
vreinterpretq_u8_p128(poly128_t __p0)56450 __ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
56451   uint8x16_t __ret;
56452   __ret = (uint8x16_t)(__p0);
56453   return __ret;
56454 }
vreinterpretq_u8_p64(poly64x2_t __p0)56455 __ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
56456   uint8x16_t __ret;
56457   __ret = (uint8x16_t)(__p0);
56458   return __ret;
56459 }
vreinterpretq_u8_p16(poly16x8_t __p0)56460 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
56461   uint8x16_t __ret;
56462   __ret = (uint8x16_t)(__p0);
56463   return __ret;
56464 }
vreinterpretq_u8_u32(uint32x4_t __p0)56465 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
56466   uint8x16_t __ret;
56467   __ret = (uint8x16_t)(__p0);
56468   return __ret;
56469 }
vreinterpretq_u8_u64(uint64x2_t __p0)56470 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
56471   uint8x16_t __ret;
56472   __ret = (uint8x16_t)(__p0);
56473   return __ret;
56474 }
vreinterpretq_u8_u16(uint16x8_t __p0)56475 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
56476   uint8x16_t __ret;
56477   __ret = (uint8x16_t)(__p0);
56478   return __ret;
56479 }
vreinterpretq_u8_s8(int8x16_t __p0)56480 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
56481   uint8x16_t __ret;
56482   __ret = (uint8x16_t)(__p0);
56483   return __ret;
56484 }
vreinterpretq_u8_f64(float64x2_t __p0)56485 __ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
56486   uint8x16_t __ret;
56487   __ret = (uint8x16_t)(__p0);
56488   return __ret;
56489 }
vreinterpretq_u8_f32(float32x4_t __p0)56490 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
56491   uint8x16_t __ret;
56492   __ret = (uint8x16_t)(__p0);
56493   return __ret;
56494 }
vreinterpretq_u8_f16(float16x8_t __p0)56495 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
56496   uint8x16_t __ret;
56497   __ret = (uint8x16_t)(__p0);
56498   return __ret;
56499 }
vreinterpretq_u8_s32(int32x4_t __p0)56500 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
56501   uint8x16_t __ret;
56502   __ret = (uint8x16_t)(__p0);
56503   return __ret;
56504 }
vreinterpretq_u8_s64(int64x2_t __p0)56505 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
56506   uint8x16_t __ret;
56507   __ret = (uint8x16_t)(__p0);
56508   return __ret;
56509 }
vreinterpretq_u8_s16(int16x8_t __p0)56510 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
56511   uint8x16_t __ret;
56512   __ret = (uint8x16_t)(__p0);
56513   return __ret;
56514 }
vreinterpretq_u32_p8(poly8x16_t __p0)56515 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
56516   uint32x4_t __ret;
56517   __ret = (uint32x4_t)(__p0);
56518   return __ret;
56519 }
vreinterpretq_u32_p128(poly128_t __p0)56520 __ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
56521   uint32x4_t __ret;
56522   __ret = (uint32x4_t)(__p0);
56523   return __ret;
56524 }
vreinterpretq_u32_p64(poly64x2_t __p0)56525 __ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
56526   uint32x4_t __ret;
56527   __ret = (uint32x4_t)(__p0);
56528   return __ret;
56529 }
vreinterpretq_u32_p16(poly16x8_t __p0)56530 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
56531   uint32x4_t __ret;
56532   __ret = (uint32x4_t)(__p0);
56533   return __ret;
56534 }
vreinterpretq_u32_u8(uint8x16_t __p0)56535 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
56536   uint32x4_t __ret;
56537   __ret = (uint32x4_t)(__p0);
56538   return __ret;
56539 }
vreinterpretq_u32_u64(uint64x2_t __p0)56540 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
56541   uint32x4_t __ret;
56542   __ret = (uint32x4_t)(__p0);
56543   return __ret;
56544 }
vreinterpretq_u32_u16(uint16x8_t __p0)56545 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
56546   uint32x4_t __ret;
56547   __ret = (uint32x4_t)(__p0);
56548   return __ret;
56549 }
vreinterpretq_u32_s8(int8x16_t __p0)56550 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
56551   uint32x4_t __ret;
56552   __ret = (uint32x4_t)(__p0);
56553   return __ret;
56554 }
vreinterpretq_u32_f64(float64x2_t __p0)56555 __ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
56556   uint32x4_t __ret;
56557   __ret = (uint32x4_t)(__p0);
56558   return __ret;
56559 }
vreinterpretq_u32_f32(float32x4_t __p0)56560 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
56561   uint32x4_t __ret;
56562   __ret = (uint32x4_t)(__p0);
56563   return __ret;
56564 }
vreinterpretq_u32_f16(float16x8_t __p0)56565 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
56566   uint32x4_t __ret;
56567   __ret = (uint32x4_t)(__p0);
56568   return __ret;
56569 }
vreinterpretq_u32_s32(int32x4_t __p0)56570 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
56571   uint32x4_t __ret;
56572   __ret = (uint32x4_t)(__p0);
56573   return __ret;
56574 }
vreinterpretq_u32_s64(int64x2_t __p0)56575 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
56576   uint32x4_t __ret;
56577   __ret = (uint32x4_t)(__p0);
56578   return __ret;
56579 }
vreinterpretq_u32_s16(int16x8_t __p0)56580 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
56581   uint32x4_t __ret;
56582   __ret = (uint32x4_t)(__p0);
56583   return __ret;
56584 }
vreinterpretq_u64_p8(poly8x16_t __p0)56585 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
56586   uint64x2_t __ret;
56587   __ret = (uint64x2_t)(__p0);
56588   return __ret;
56589 }
vreinterpretq_u64_p128(poly128_t __p0)56590 __ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
56591   uint64x2_t __ret;
56592   __ret = (uint64x2_t)(__p0);
56593   return __ret;
56594 }
vreinterpretq_u64_p64(poly64x2_t __p0)56595 __ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
56596   uint64x2_t __ret;
56597   __ret = (uint64x2_t)(__p0);
56598   return __ret;
56599 }
vreinterpretq_u64_p16(poly16x8_t __p0)56600 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
56601   uint64x2_t __ret;
56602   __ret = (uint64x2_t)(__p0);
56603   return __ret;
56604 }
vreinterpretq_u64_u8(uint8x16_t __p0)56605 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
56606   uint64x2_t __ret;
56607   __ret = (uint64x2_t)(__p0);
56608   return __ret;
56609 }
vreinterpretq_u64_u32(uint32x4_t __p0)56610 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
56611   uint64x2_t __ret;
56612   __ret = (uint64x2_t)(__p0);
56613   return __ret;
56614 }
vreinterpretq_u64_u16(uint16x8_t __p0)56615 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
56616   uint64x2_t __ret;
56617   __ret = (uint64x2_t)(__p0);
56618   return __ret;
56619 }
vreinterpretq_u64_s8(int8x16_t __p0)56620 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
56621   uint64x2_t __ret;
56622   __ret = (uint64x2_t)(__p0);
56623   return __ret;
56624 }
vreinterpretq_u64_f64(float64x2_t __p0)56625 __ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
56626   uint64x2_t __ret;
56627   __ret = (uint64x2_t)(__p0);
56628   return __ret;
56629 }
vreinterpretq_u64_f32(float32x4_t __p0)56630 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
56631   uint64x2_t __ret;
56632   __ret = (uint64x2_t)(__p0);
56633   return __ret;
56634 }
vreinterpretq_u64_f16(float16x8_t __p0)56635 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
56636   uint64x2_t __ret;
56637   __ret = (uint64x2_t)(__p0);
56638   return __ret;
56639 }
vreinterpretq_u64_s32(int32x4_t __p0)56640 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
56641   uint64x2_t __ret;
56642   __ret = (uint64x2_t)(__p0);
56643   return __ret;
56644 }
vreinterpretq_u64_s64(int64x2_t __p0)56645 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
56646   uint64x2_t __ret;
56647   __ret = (uint64x2_t)(__p0);
56648   return __ret;
56649 }
vreinterpretq_u64_s16(int16x8_t __p0)56650 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
56651   uint64x2_t __ret;
56652   __ret = (uint64x2_t)(__p0);
56653   return __ret;
56654 }
vreinterpretq_u16_p8(poly8x16_t __p0)56655 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
56656   uint16x8_t __ret;
56657   __ret = (uint16x8_t)(__p0);
56658   return __ret;
56659 }
vreinterpretq_u16_p128(poly128_t __p0)56660 __ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
56661   uint16x8_t __ret;
56662   __ret = (uint16x8_t)(__p0);
56663   return __ret;
56664 }
vreinterpretq_u16_p64(poly64x2_t __p0)56665 __ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
56666   uint16x8_t __ret;
56667   __ret = (uint16x8_t)(__p0);
56668   return __ret;
56669 }
vreinterpretq_u16_p16(poly16x8_t __p0)56670 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
56671   uint16x8_t __ret;
56672   __ret = (uint16x8_t)(__p0);
56673   return __ret;
56674 }
vreinterpretq_u16_u8(uint8x16_t __p0)56675 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
56676   uint16x8_t __ret;
56677   __ret = (uint16x8_t)(__p0);
56678   return __ret;
56679 }
vreinterpretq_u16_u32(uint32x4_t __p0)56680 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
56681   uint16x8_t __ret;
56682   __ret = (uint16x8_t)(__p0);
56683   return __ret;
56684 }
vreinterpretq_u16_u64(uint64x2_t __p0)56685 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
56686   uint16x8_t __ret;
56687   __ret = (uint16x8_t)(__p0);
56688   return __ret;
56689 }
vreinterpretq_u16_s8(int8x16_t __p0)56690 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
56691   uint16x8_t __ret;
56692   __ret = (uint16x8_t)(__p0);
56693   return __ret;
56694 }
vreinterpretq_u16_f64(float64x2_t __p0)56695 __ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
56696   uint16x8_t __ret;
56697   __ret = (uint16x8_t)(__p0);
56698   return __ret;
56699 }
vreinterpretq_u16_f32(float32x4_t __p0)56700 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
56701   uint16x8_t __ret;
56702   __ret = (uint16x8_t)(__p0);
56703   return __ret;
56704 }
vreinterpretq_u16_f16(float16x8_t __p0)56705 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
56706   uint16x8_t __ret;
56707   __ret = (uint16x8_t)(__p0);
56708   return __ret;
56709 }
vreinterpretq_u16_s32(int32x4_t __p0)56710 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
56711   uint16x8_t __ret;
56712   __ret = (uint16x8_t)(__p0);
56713   return __ret;
56714 }
vreinterpretq_u16_s64(int64x2_t __p0)56715 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
56716   uint16x8_t __ret;
56717   __ret = (uint16x8_t)(__p0);
56718   return __ret;
56719 }
vreinterpretq_u16_s16(int16x8_t __p0)56720 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
56721   uint16x8_t __ret;
56722   __ret = (uint16x8_t)(__p0);
56723   return __ret;
56724 }
vreinterpretq_s8_p8(poly8x16_t __p0)56725 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
56726   int8x16_t __ret;
56727   __ret = (int8x16_t)(__p0);
56728   return __ret;
56729 }
vreinterpretq_s8_p128(poly128_t __p0)56730 __ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
56731   int8x16_t __ret;
56732   __ret = (int8x16_t)(__p0);
56733   return __ret;
56734 }
vreinterpretq_s8_p64(poly64x2_t __p0)56735 __ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
56736   int8x16_t __ret;
56737   __ret = (int8x16_t)(__p0);
56738   return __ret;
56739 }
vreinterpretq_s8_p16(poly16x8_t __p0)56740 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
56741   int8x16_t __ret;
56742   __ret = (int8x16_t)(__p0);
56743   return __ret;
56744 }
vreinterpretq_s8_u8(uint8x16_t __p0)56745 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
56746   int8x16_t __ret;
56747   __ret = (int8x16_t)(__p0);
56748   return __ret;
56749 }
vreinterpretq_s8_u32(uint32x4_t __p0)56750 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
56751   int8x16_t __ret;
56752   __ret = (int8x16_t)(__p0);
56753   return __ret;
56754 }
vreinterpretq_s8_u64(uint64x2_t __p0)56755 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
56756   int8x16_t __ret;
56757   __ret = (int8x16_t)(__p0);
56758   return __ret;
56759 }
vreinterpretq_s8_u16(uint16x8_t __p0)56760 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
56761   int8x16_t __ret;
56762   __ret = (int8x16_t)(__p0);
56763   return __ret;
56764 }
vreinterpretq_s8_f64(float64x2_t __p0)56765 __ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
56766   int8x16_t __ret;
56767   __ret = (int8x16_t)(__p0);
56768   return __ret;
56769 }
vreinterpretq_s8_f32(float32x4_t __p0)56770 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
56771   int8x16_t __ret;
56772   __ret = (int8x16_t)(__p0);
56773   return __ret;
56774 }
vreinterpretq_s8_f16(float16x8_t __p0)56775 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
56776   int8x16_t __ret;
56777   __ret = (int8x16_t)(__p0);
56778   return __ret;
56779 }
vreinterpretq_s8_s32(int32x4_t __p0)56780 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
56781   int8x16_t __ret;
56782   __ret = (int8x16_t)(__p0);
56783   return __ret;
56784 }
vreinterpretq_s8_s64(int64x2_t __p0)56785 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
56786   int8x16_t __ret;
56787   __ret = (int8x16_t)(__p0);
56788   return __ret;
56789 }
vreinterpretq_s8_s16(int16x8_t __p0)56790 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
56791   int8x16_t __ret;
56792   __ret = (int8x16_t)(__p0);
56793   return __ret;
56794 }
vreinterpretq_f64_p8(poly8x16_t __p0)56795 __ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
56796   float64x2_t __ret;
56797   __ret = (float64x2_t)(__p0);
56798   return __ret;
56799 }
vreinterpretq_f64_p128(poly128_t __p0)56800 __ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
56801   float64x2_t __ret;
56802   __ret = (float64x2_t)(__p0);
56803   return __ret;
56804 }
vreinterpretq_f64_p64(poly64x2_t __p0)56805 __ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
56806   float64x2_t __ret;
56807   __ret = (float64x2_t)(__p0);
56808   return __ret;
56809 }
vreinterpretq_f64_p16(poly16x8_t __p0)56810 __ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
56811   float64x2_t __ret;
56812   __ret = (float64x2_t)(__p0);
56813   return __ret;
56814 }
vreinterpretq_f64_u8(uint8x16_t __p0)56815 __ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
56816   float64x2_t __ret;
56817   __ret = (float64x2_t)(__p0);
56818   return __ret;
56819 }
vreinterpretq_f64_u32(uint32x4_t __p0)56820 __ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
56821   float64x2_t __ret;
56822   __ret = (float64x2_t)(__p0);
56823   return __ret;
56824 }
vreinterpretq_f64_u64(uint64x2_t __p0)56825 __ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
56826   float64x2_t __ret;
56827   __ret = (float64x2_t)(__p0);
56828   return __ret;
56829 }
vreinterpretq_f64_u16(uint16x8_t __p0)56830 __ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
56831   float64x2_t __ret;
56832   __ret = (float64x2_t)(__p0);
56833   return __ret;
56834 }
vreinterpretq_f64_s8(int8x16_t __p0)56835 __ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
56836   float64x2_t __ret;
56837   __ret = (float64x2_t)(__p0);
56838   return __ret;
56839 }
vreinterpretq_f64_f32(float32x4_t __p0)56840 __ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
56841   float64x2_t __ret;
56842   __ret = (float64x2_t)(__p0);
56843   return __ret;
56844 }
vreinterpretq_f64_f16(float16x8_t __p0)56845 __ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
56846   float64x2_t __ret;
56847   __ret = (float64x2_t)(__p0);
56848   return __ret;
56849 }
vreinterpretq_f64_s32(int32x4_t __p0)56850 __ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
56851   float64x2_t __ret;
56852   __ret = (float64x2_t)(__p0);
56853   return __ret;
56854 }
vreinterpretq_f64_s64(int64x2_t __p0)56855 __ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
56856   float64x2_t __ret;
56857   __ret = (float64x2_t)(__p0);
56858   return __ret;
56859 }
vreinterpretq_f64_s16(int16x8_t __p0)56860 __ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
56861   float64x2_t __ret;
56862   __ret = (float64x2_t)(__p0);
56863   return __ret;
56864 }
vreinterpretq_f32_p8(poly8x16_t __p0)56865 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
56866   float32x4_t __ret;
56867   __ret = (float32x4_t)(__p0);
56868   return __ret;
56869 }
vreinterpretq_f32_p128(poly128_t __p0)56870 __ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
56871   float32x4_t __ret;
56872   __ret = (float32x4_t)(__p0);
56873   return __ret;
56874 }
vreinterpretq_f32_p64(poly64x2_t __p0)56875 __ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
56876   float32x4_t __ret;
56877   __ret = (float32x4_t)(__p0);
56878   return __ret;
56879 }
vreinterpretq_f32_p16(poly16x8_t __p0)56880 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
56881   float32x4_t __ret;
56882   __ret = (float32x4_t)(__p0);
56883   return __ret;
56884 }
vreinterpretq_f32_u8(uint8x16_t __p0)56885 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
56886   float32x4_t __ret;
56887   __ret = (float32x4_t)(__p0);
56888   return __ret;
56889 }
vreinterpretq_f32_u32(uint32x4_t __p0)56890 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
56891   float32x4_t __ret;
56892   __ret = (float32x4_t)(__p0);
56893   return __ret;
56894 }
vreinterpretq_f32_u64(uint64x2_t __p0)56895 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
56896   float32x4_t __ret;
56897   __ret = (float32x4_t)(__p0);
56898   return __ret;
56899 }
vreinterpretq_f32_u16(uint16x8_t __p0)56900 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
56901   float32x4_t __ret;
56902   __ret = (float32x4_t)(__p0);
56903   return __ret;
56904 }
vreinterpretq_f32_s8(int8x16_t __p0)56905 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
56906   float32x4_t __ret;
56907   __ret = (float32x4_t)(__p0);
56908   return __ret;
56909 }
vreinterpretq_f32_f64(float64x2_t __p0)56910 __ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
56911   float32x4_t __ret;
56912   __ret = (float32x4_t)(__p0);
56913   return __ret;
56914 }
vreinterpretq_f32_f16(float16x8_t __p0)56915 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
56916   float32x4_t __ret;
56917   __ret = (float32x4_t)(__p0);
56918   return __ret;
56919 }
vreinterpretq_f32_s32(int32x4_t __p0)56920 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
56921   float32x4_t __ret;
56922   __ret = (float32x4_t)(__p0);
56923   return __ret;
56924 }
vreinterpretq_f32_s64(int64x2_t __p0)56925 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
56926   float32x4_t __ret;
56927   __ret = (float32x4_t)(__p0);
56928   return __ret;
56929 }
vreinterpretq_f32_s16(int16x8_t __p0)56930 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
56931   float32x4_t __ret;
56932   __ret = (float32x4_t)(__p0);
56933   return __ret;
56934 }
vreinterpretq_f16_p8(poly8x16_t __p0)56935 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
56936   float16x8_t __ret;
56937   __ret = (float16x8_t)(__p0);
56938   return __ret;
56939 }
vreinterpretq_f16_p128(poly128_t __p0)56940 __ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
56941   float16x8_t __ret;
56942   __ret = (float16x8_t)(__p0);
56943   return __ret;
56944 }
vreinterpretq_f16_p64(poly64x2_t __p0)56945 __ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
56946   float16x8_t __ret;
56947   __ret = (float16x8_t)(__p0);
56948   return __ret;
56949 }
vreinterpretq_f16_p16(poly16x8_t __p0)56950 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
56951   float16x8_t __ret;
56952   __ret = (float16x8_t)(__p0);
56953   return __ret;
56954 }
vreinterpretq_f16_u8(uint8x16_t __p0)56955 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
56956   float16x8_t __ret;
56957   __ret = (float16x8_t)(__p0);
56958   return __ret;
56959 }
vreinterpretq_f16_u32(uint32x4_t __p0)56960 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
56961   float16x8_t __ret;
56962   __ret = (float16x8_t)(__p0);
56963   return __ret;
56964 }
vreinterpretq_f16_u64(uint64x2_t __p0)56965 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
56966   float16x8_t __ret;
56967   __ret = (float16x8_t)(__p0);
56968   return __ret;
56969 }
vreinterpretq_f16_u16(uint16x8_t __p0)56970 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
56971   float16x8_t __ret;
56972   __ret = (float16x8_t)(__p0);
56973   return __ret;
56974 }
vreinterpretq_f16_s8(int8x16_t __p0)56975 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
56976   float16x8_t __ret;
56977   __ret = (float16x8_t)(__p0);
56978   return __ret;
56979 }
vreinterpretq_f16_f64(float64x2_t __p0)56980 __ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
56981   float16x8_t __ret;
56982   __ret = (float16x8_t)(__p0);
56983   return __ret;
56984 }
vreinterpretq_f16_f32(float32x4_t __p0)56985 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
56986   float16x8_t __ret;
56987   __ret = (float16x8_t)(__p0);
56988   return __ret;
56989 }
vreinterpretq_f16_s32(int32x4_t __p0)56990 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
56991   float16x8_t __ret;
56992   __ret = (float16x8_t)(__p0);
56993   return __ret;
56994 }
vreinterpretq_f16_s64(int64x2_t __p0)56995 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
56996   float16x8_t __ret;
56997   __ret = (float16x8_t)(__p0);
56998   return __ret;
56999 }
vreinterpretq_f16_s16(int16x8_t __p0)57000 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
57001   float16x8_t __ret;
57002   __ret = (float16x8_t)(__p0);
57003   return __ret;
57004 }
vreinterpretq_s32_p8(poly8x16_t __p0)57005 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
57006   int32x4_t __ret;
57007   __ret = (int32x4_t)(__p0);
57008   return __ret;
57009 }
vreinterpretq_s32_p128(poly128_t __p0)57010 __ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
57011   int32x4_t __ret;
57012   __ret = (int32x4_t)(__p0);
57013   return __ret;
57014 }
vreinterpretq_s32_p64(poly64x2_t __p0)57015 __ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
57016   int32x4_t __ret;
57017   __ret = (int32x4_t)(__p0);
57018   return __ret;
57019 }
vreinterpretq_s32_p16(poly16x8_t __p0)57020 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
57021   int32x4_t __ret;
57022   __ret = (int32x4_t)(__p0);
57023   return __ret;
57024 }
vreinterpretq_s32_u8(uint8x16_t __p0)57025 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
57026   int32x4_t __ret;
57027   __ret = (int32x4_t)(__p0);
57028   return __ret;
57029 }
vreinterpretq_s32_u32(uint32x4_t __p0)57030 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
57031   int32x4_t __ret;
57032   __ret = (int32x4_t)(__p0);
57033   return __ret;
57034 }
vreinterpretq_s32_u64(uint64x2_t __p0)57035 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
57036   int32x4_t __ret;
57037   __ret = (int32x4_t)(__p0);
57038   return __ret;
57039 }
vreinterpretq_s32_u16(uint16x8_t __p0)57040 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
57041   int32x4_t __ret;
57042   __ret = (int32x4_t)(__p0);
57043   return __ret;
57044 }
vreinterpretq_s32_s8(int8x16_t __p0)57045 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
57046   int32x4_t __ret;
57047   __ret = (int32x4_t)(__p0);
57048   return __ret;
57049 }
vreinterpretq_s32_f64(float64x2_t __p0)57050 __ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
57051   int32x4_t __ret;
57052   __ret = (int32x4_t)(__p0);
57053   return __ret;
57054 }
vreinterpretq_s32_f32(float32x4_t __p0)57055 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
57056   int32x4_t __ret;
57057   __ret = (int32x4_t)(__p0);
57058   return __ret;
57059 }
vreinterpretq_s32_f16(float16x8_t __p0)57060 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
57061   int32x4_t __ret;
57062   __ret = (int32x4_t)(__p0);
57063   return __ret;
57064 }
vreinterpretq_s32_s64(int64x2_t __p0)57065 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
57066   int32x4_t __ret;
57067   __ret = (int32x4_t)(__p0);
57068   return __ret;
57069 }
vreinterpretq_s32_s16(int16x8_t __p0)57070 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
57071   int32x4_t __ret;
57072   __ret = (int32x4_t)(__p0);
57073   return __ret;
57074 }
vreinterpretq_s64_p8(poly8x16_t __p0)57075 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
57076   int64x2_t __ret;
57077   __ret = (int64x2_t)(__p0);
57078   return __ret;
57079 }
vreinterpretq_s64_p128(poly128_t __p0)57080 __ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
57081   int64x2_t __ret;
57082   __ret = (int64x2_t)(__p0);
57083   return __ret;
57084 }
vreinterpretq_s64_p64(poly64x2_t __p0)57085 __ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
57086   int64x2_t __ret;
57087   __ret = (int64x2_t)(__p0);
57088   return __ret;
57089 }
vreinterpretq_s64_p16(poly16x8_t __p0)57090 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
57091   int64x2_t __ret;
57092   __ret = (int64x2_t)(__p0);
57093   return __ret;
57094 }
vreinterpretq_s64_u8(uint8x16_t __p0)57095 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
57096   int64x2_t __ret;
57097   __ret = (int64x2_t)(__p0);
57098   return __ret;
57099 }
vreinterpretq_s64_u32(uint32x4_t __p0)57100 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
57101   int64x2_t __ret;
57102   __ret = (int64x2_t)(__p0);
57103   return __ret;
57104 }
vreinterpretq_s64_u64(uint64x2_t __p0)57105 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
57106   int64x2_t __ret;
57107   __ret = (int64x2_t)(__p0);
57108   return __ret;
57109 }
vreinterpretq_s64_u16(uint16x8_t __p0)57110 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
57111   int64x2_t __ret;
57112   __ret = (int64x2_t)(__p0);
57113   return __ret;
57114 }
vreinterpretq_s64_s8(int8x16_t __p0)57115 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
57116   int64x2_t __ret;
57117   __ret = (int64x2_t)(__p0);
57118   return __ret;
57119 }
vreinterpretq_s64_f64(float64x2_t __p0)57120 __ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
57121   int64x2_t __ret;
57122   __ret = (int64x2_t)(__p0);
57123   return __ret;
57124 }
vreinterpretq_s64_f32(float32x4_t __p0)57125 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
57126   int64x2_t __ret;
57127   __ret = (int64x2_t)(__p0);
57128   return __ret;
57129 }
vreinterpretq_s64_f16(float16x8_t __p0)57130 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
57131   int64x2_t __ret;
57132   __ret = (int64x2_t)(__p0);
57133   return __ret;
57134 }
vreinterpretq_s64_s32(int32x4_t __p0)57135 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
57136   int64x2_t __ret;
57137   __ret = (int64x2_t)(__p0);
57138   return __ret;
57139 }
vreinterpretq_s64_s16(int16x8_t __p0)57140 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
57141   int64x2_t __ret;
57142   __ret = (int64x2_t)(__p0);
57143   return __ret;
57144 }
vreinterpretq_s16_p8(poly8x16_t __p0)57145 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
57146   int16x8_t __ret;
57147   __ret = (int16x8_t)(__p0);
57148   return __ret;
57149 }
vreinterpretq_s16_p128(poly128_t __p0)57150 __ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
57151   int16x8_t __ret;
57152   __ret = (int16x8_t)(__p0);
57153   return __ret;
57154 }
vreinterpretq_s16_p64(poly64x2_t __p0)57155 __ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
57156   int16x8_t __ret;
57157   __ret = (int16x8_t)(__p0);
57158   return __ret;
57159 }
vreinterpretq_s16_p16(poly16x8_t __p0)57160 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
57161   int16x8_t __ret;
57162   __ret = (int16x8_t)(__p0);
57163   return __ret;
57164 }
vreinterpretq_s16_u8(uint8x16_t __p0)57165 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
57166   int16x8_t __ret;
57167   __ret = (int16x8_t)(__p0);
57168   return __ret;
57169 }
vreinterpretq_s16_u32(uint32x4_t __p0)57170 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
57171   int16x8_t __ret;
57172   __ret = (int16x8_t)(__p0);
57173   return __ret;
57174 }
vreinterpretq_s16_u64(uint64x2_t __p0)57175 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
57176   int16x8_t __ret;
57177   __ret = (int16x8_t)(__p0);
57178   return __ret;
57179 }
vreinterpretq_s16_u16(uint16x8_t __p0)57180 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
57181   int16x8_t __ret;
57182   __ret = (int16x8_t)(__p0);
57183   return __ret;
57184 }
vreinterpretq_s16_s8(int8x16_t __p0)57185 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
57186   int16x8_t __ret;
57187   __ret = (int16x8_t)(__p0);
57188   return __ret;
57189 }
vreinterpretq_s16_f64(float64x2_t __p0)57190 __ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
57191   int16x8_t __ret;
57192   __ret = (int16x8_t)(__p0);
57193   return __ret;
57194 }
vreinterpretq_s16_f32(float32x4_t __p0)57195 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
57196   int16x8_t __ret;
57197   __ret = (int16x8_t)(__p0);
57198   return __ret;
57199 }
vreinterpretq_s16_f16(float16x8_t __p0)57200 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
57201   int16x8_t __ret;
57202   __ret = (int16x8_t)(__p0);
57203   return __ret;
57204 }
vreinterpretq_s16_s32(int32x4_t __p0)57205 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
57206   int16x8_t __ret;
57207   __ret = (int16x8_t)(__p0);
57208   return __ret;
57209 }
vreinterpretq_s16_s64(int64x2_t __p0)57210 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
57211   int16x8_t __ret;
57212   __ret = (int16x8_t)(__p0);
57213   return __ret;
57214 }
vreinterpret_u8_p8(poly8x8_t __p0)57215 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
57216   uint8x8_t __ret;
57217   __ret = (uint8x8_t)(__p0);
57218   return __ret;
57219 }
vreinterpret_u8_p64(poly64x1_t __p0)57220 __ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
57221   uint8x8_t __ret;
57222   __ret = (uint8x8_t)(__p0);
57223   return __ret;
57224 }
vreinterpret_u8_p16(poly16x4_t __p0)57225 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
57226   uint8x8_t __ret;
57227   __ret = (uint8x8_t)(__p0);
57228   return __ret;
57229 }
vreinterpret_u8_u32(uint32x2_t __p0)57230 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
57231   uint8x8_t __ret;
57232   __ret = (uint8x8_t)(__p0);
57233   return __ret;
57234 }
vreinterpret_u8_u64(uint64x1_t __p0)57235 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
57236   uint8x8_t __ret;
57237   __ret = (uint8x8_t)(__p0);
57238   return __ret;
57239 }
vreinterpret_u8_u16(uint16x4_t __p0)57240 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
57241   uint8x8_t __ret;
57242   __ret = (uint8x8_t)(__p0);
57243   return __ret;
57244 }
vreinterpret_u8_s8(int8x8_t __p0)57245 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
57246   uint8x8_t __ret;
57247   __ret = (uint8x8_t)(__p0);
57248   return __ret;
57249 }
vreinterpret_u8_f64(float64x1_t __p0)57250 __ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
57251   uint8x8_t __ret;
57252   __ret = (uint8x8_t)(__p0);
57253   return __ret;
57254 }
vreinterpret_u8_f32(float32x2_t __p0)57255 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
57256   uint8x8_t __ret;
57257   __ret = (uint8x8_t)(__p0);
57258   return __ret;
57259 }
vreinterpret_u8_f16(float16x4_t __p0)57260 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
57261   uint8x8_t __ret;
57262   __ret = (uint8x8_t)(__p0);
57263   return __ret;
57264 }
vreinterpret_u8_s32(int32x2_t __p0)57265 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
57266   uint8x8_t __ret;
57267   __ret = (uint8x8_t)(__p0);
57268   return __ret;
57269 }
vreinterpret_u8_s64(int64x1_t __p0)57270 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
57271   uint8x8_t __ret;
57272   __ret = (uint8x8_t)(__p0);
57273   return __ret;
57274 }
vreinterpret_u8_s16(int16x4_t __p0)57275 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
57276   uint8x8_t __ret;
57277   __ret = (uint8x8_t)(__p0);
57278   return __ret;
57279 }
vreinterpret_u32_p8(poly8x8_t __p0)57280 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
57281   uint32x2_t __ret;
57282   __ret = (uint32x2_t)(__p0);
57283   return __ret;
57284 }
vreinterpret_u32_p64(poly64x1_t __p0)57285 __ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
57286   uint32x2_t __ret;
57287   __ret = (uint32x2_t)(__p0);
57288   return __ret;
57289 }
vreinterpret_u32_p16(poly16x4_t __p0)57290 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
57291   uint32x2_t __ret;
57292   __ret = (uint32x2_t)(__p0);
57293   return __ret;
57294 }
vreinterpret_u32_u8(uint8x8_t __p0)57295 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
57296   uint32x2_t __ret;
57297   __ret = (uint32x2_t)(__p0);
57298   return __ret;
57299 }
vreinterpret_u32_u64(uint64x1_t __p0)57300 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
57301   uint32x2_t __ret;
57302   __ret = (uint32x2_t)(__p0);
57303   return __ret;
57304 }
vreinterpret_u32_u16(uint16x4_t __p0)57305 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
57306   uint32x2_t __ret;
57307   __ret = (uint32x2_t)(__p0);
57308   return __ret;
57309 }
vreinterpret_u32_s8(int8x8_t __p0)57310 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
57311   uint32x2_t __ret;
57312   __ret = (uint32x2_t)(__p0);
57313   return __ret;
57314 }
vreinterpret_u32_f64(float64x1_t __p0)57315 __ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
57316   uint32x2_t __ret;
57317   __ret = (uint32x2_t)(__p0);
57318   return __ret;
57319 }
vreinterpret_u32_f32(float32x2_t __p0)57320 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
57321   uint32x2_t __ret;
57322   __ret = (uint32x2_t)(__p0);
57323   return __ret;
57324 }
vreinterpret_u32_f16(float16x4_t __p0)57325 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
57326   uint32x2_t __ret;
57327   __ret = (uint32x2_t)(__p0);
57328   return __ret;
57329 }
vreinterpret_u32_s32(int32x2_t __p0)57330 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
57331   uint32x2_t __ret;
57332   __ret = (uint32x2_t)(__p0);
57333   return __ret;
57334 }
vreinterpret_u32_s64(int64x1_t __p0)57335 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
57336   uint32x2_t __ret;
57337   __ret = (uint32x2_t)(__p0);
57338   return __ret;
57339 }
vreinterpret_u32_s16(int16x4_t __p0)57340 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
57341   uint32x2_t __ret;
57342   __ret = (uint32x2_t)(__p0);
57343   return __ret;
57344 }
vreinterpret_u64_p8(poly8x8_t __p0)57345 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
57346   uint64x1_t __ret;
57347   __ret = (uint64x1_t)(__p0);
57348   return __ret;
57349 }
vreinterpret_u64_p64(poly64x1_t __p0)57350 __ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
57351   uint64x1_t __ret;
57352   __ret = (uint64x1_t)(__p0);
57353   return __ret;
57354 }
vreinterpret_u64_p16(poly16x4_t __p0)57355 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
57356   uint64x1_t __ret;
57357   __ret = (uint64x1_t)(__p0);
57358   return __ret;
57359 }
vreinterpret_u64_u8(uint8x8_t __p0)57360 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
57361   uint64x1_t __ret;
57362   __ret = (uint64x1_t)(__p0);
57363   return __ret;
57364 }
vreinterpret_u64_u32(uint32x2_t __p0)57365 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
57366   uint64x1_t __ret;
57367   __ret = (uint64x1_t)(__p0);
57368   return __ret;
57369 }
vreinterpret_u64_u16(uint16x4_t __p0)57370 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
57371   uint64x1_t __ret;
57372   __ret = (uint64x1_t)(__p0);
57373   return __ret;
57374 }
vreinterpret_u64_s8(int8x8_t __p0)57375 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
57376   uint64x1_t __ret;
57377   __ret = (uint64x1_t)(__p0);
57378   return __ret;
57379 }
vreinterpret_u64_f64(float64x1_t __p0)57380 __ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
57381   uint64x1_t __ret;
57382   __ret = (uint64x1_t)(__p0);
57383   return __ret;
57384 }
vreinterpret_u64_f32(float32x2_t __p0)57385 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
57386   uint64x1_t __ret;
57387   __ret = (uint64x1_t)(__p0);
57388   return __ret;
57389 }
vreinterpret_u64_f16(float16x4_t __p0)57390 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
57391   uint64x1_t __ret;
57392   __ret = (uint64x1_t)(__p0);
57393   return __ret;
57394 }
vreinterpret_u64_s32(int32x2_t __p0)57395 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
57396   uint64x1_t __ret;
57397   __ret = (uint64x1_t)(__p0);
57398   return __ret;
57399 }
vreinterpret_u64_s64(int64x1_t __p0)57400 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
57401   uint64x1_t __ret;
57402   __ret = (uint64x1_t)(__p0);
57403   return __ret;
57404 }
vreinterpret_u64_s16(int16x4_t __p0)57405 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
57406   uint64x1_t __ret;
57407   __ret = (uint64x1_t)(__p0);
57408   return __ret;
57409 }
vreinterpret_u16_p8(poly8x8_t __p0)57410 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
57411   uint16x4_t __ret;
57412   __ret = (uint16x4_t)(__p0);
57413   return __ret;
57414 }
vreinterpret_u16_p64(poly64x1_t __p0)57415 __ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
57416   uint16x4_t __ret;
57417   __ret = (uint16x4_t)(__p0);
57418   return __ret;
57419 }
vreinterpret_u16_p16(poly16x4_t __p0)57420 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
57421   uint16x4_t __ret;
57422   __ret = (uint16x4_t)(__p0);
57423   return __ret;
57424 }
vreinterpret_u16_u8(uint8x8_t __p0)57425 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
57426   uint16x4_t __ret;
57427   __ret = (uint16x4_t)(__p0);
57428   return __ret;
57429 }
vreinterpret_u16_u32(uint32x2_t __p0)57430 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
57431   uint16x4_t __ret;
57432   __ret = (uint16x4_t)(__p0);
57433   return __ret;
57434 }
vreinterpret_u16_u64(uint64x1_t __p0)57435 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
57436   uint16x4_t __ret;
57437   __ret = (uint16x4_t)(__p0);
57438   return __ret;
57439 }
vreinterpret_u16_s8(int8x8_t __p0)57440 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
57441   uint16x4_t __ret;
57442   __ret = (uint16x4_t)(__p0);
57443   return __ret;
57444 }
vreinterpret_u16_f64(float64x1_t __p0)57445 __ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
57446   uint16x4_t __ret;
57447   __ret = (uint16x4_t)(__p0);
57448   return __ret;
57449 }
vreinterpret_u16_f32(float32x2_t __p0)57450 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
57451   uint16x4_t __ret;
57452   __ret = (uint16x4_t)(__p0);
57453   return __ret;
57454 }
vreinterpret_u16_f16(float16x4_t __p0)57455 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
57456   uint16x4_t __ret;
57457   __ret = (uint16x4_t)(__p0);
57458   return __ret;
57459 }
vreinterpret_u16_s32(int32x2_t __p0)57460 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
57461   uint16x4_t __ret;
57462   __ret = (uint16x4_t)(__p0);
57463   return __ret;
57464 }
vreinterpret_u16_s64(int64x1_t __p0)57465 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
57466   uint16x4_t __ret;
57467   __ret = (uint16x4_t)(__p0);
57468   return __ret;
57469 }
vreinterpret_u16_s16(int16x4_t __p0)57470 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
57471   uint16x4_t __ret;
57472   __ret = (uint16x4_t)(__p0);
57473   return __ret;
57474 }
vreinterpret_s8_p8(poly8x8_t __p0)57475 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
57476   int8x8_t __ret;
57477   __ret = (int8x8_t)(__p0);
57478   return __ret;
57479 }
vreinterpret_s8_p64(poly64x1_t __p0)57480 __ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
57481   int8x8_t __ret;
57482   __ret = (int8x8_t)(__p0);
57483   return __ret;
57484 }
vreinterpret_s8_p16(poly16x4_t __p0)57485 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
57486   int8x8_t __ret;
57487   __ret = (int8x8_t)(__p0);
57488   return __ret;
57489 }
vreinterpret_s8_u8(uint8x8_t __p0)57490 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
57491   int8x8_t __ret;
57492   __ret = (int8x8_t)(__p0);
57493   return __ret;
57494 }
vreinterpret_s8_u32(uint32x2_t __p0)57495 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
57496   int8x8_t __ret;
57497   __ret = (int8x8_t)(__p0);
57498   return __ret;
57499 }
vreinterpret_s8_u64(uint64x1_t __p0)57500 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
57501   int8x8_t __ret;
57502   __ret = (int8x8_t)(__p0);
57503   return __ret;
57504 }
vreinterpret_s8_u16(uint16x4_t __p0)57505 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
57506   int8x8_t __ret;
57507   __ret = (int8x8_t)(__p0);
57508   return __ret;
57509 }
vreinterpret_s8_f64(float64x1_t __p0)57510 __ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
57511   int8x8_t __ret;
57512   __ret = (int8x8_t)(__p0);
57513   return __ret;
57514 }
vreinterpret_s8_f32(float32x2_t __p0)57515 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
57516   int8x8_t __ret;
57517   __ret = (int8x8_t)(__p0);
57518   return __ret;
57519 }
vreinterpret_s8_f16(float16x4_t __p0)57520 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
57521   int8x8_t __ret;
57522   __ret = (int8x8_t)(__p0);
57523   return __ret;
57524 }
vreinterpret_s8_s32(int32x2_t __p0)57525 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
57526   int8x8_t __ret;
57527   __ret = (int8x8_t)(__p0);
57528   return __ret;
57529 }
vreinterpret_s8_s64(int64x1_t __p0)57530 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
57531   int8x8_t __ret;
57532   __ret = (int8x8_t)(__p0);
57533   return __ret;
57534 }
vreinterpret_s8_s16(int16x4_t __p0)57535 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
57536   int8x8_t __ret;
57537   __ret = (int8x8_t)(__p0);
57538   return __ret;
57539 }
vreinterpret_f64_p8(poly8x8_t __p0)57540 __ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
57541   float64x1_t __ret;
57542   __ret = (float64x1_t)(__p0);
57543   return __ret;
57544 }
vreinterpret_f64_p64(poly64x1_t __p0)57545 __ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
57546   float64x1_t __ret;
57547   __ret = (float64x1_t)(__p0);
57548   return __ret;
57549 }
vreinterpret_f64_p16(poly16x4_t __p0)57550 __ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
57551   float64x1_t __ret;
57552   __ret = (float64x1_t)(__p0);
57553   return __ret;
57554 }
vreinterpret_f64_u8(uint8x8_t __p0)57555 __ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
57556   float64x1_t __ret;
57557   __ret = (float64x1_t)(__p0);
57558   return __ret;
57559 }
vreinterpret_f64_u32(uint32x2_t __p0)57560 __ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
57561   float64x1_t __ret;
57562   __ret = (float64x1_t)(__p0);
57563   return __ret;
57564 }
vreinterpret_f64_u64(uint64x1_t __p0)57565 __ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
57566   float64x1_t __ret;
57567   __ret = (float64x1_t)(__p0);
57568   return __ret;
57569 }
vreinterpret_f64_u16(uint16x4_t __p0)57570 __ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
57571   float64x1_t __ret;
57572   __ret = (float64x1_t)(__p0);
57573   return __ret;
57574 }
vreinterpret_f64_s8(int8x8_t __p0)57575 __ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
57576   float64x1_t __ret;
57577   __ret = (float64x1_t)(__p0);
57578   return __ret;
57579 }
vreinterpret_f64_f32(float32x2_t __p0)57580 __ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
57581   float64x1_t __ret;
57582   __ret = (float64x1_t)(__p0);
57583   return __ret;
57584 }
vreinterpret_f64_f16(float16x4_t __p0)57585 __ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
57586   float64x1_t __ret;
57587   __ret = (float64x1_t)(__p0);
57588   return __ret;
57589 }
vreinterpret_f64_s32(int32x2_t __p0)57590 __ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
57591   float64x1_t __ret;
57592   __ret = (float64x1_t)(__p0);
57593   return __ret;
57594 }
vreinterpret_f64_s64(int64x1_t __p0)57595 __ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
57596   float64x1_t __ret;
57597   __ret = (float64x1_t)(__p0);
57598   return __ret;
57599 }
vreinterpret_f64_s16(int16x4_t __p0)57600 __ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
57601   float64x1_t __ret;
57602   __ret = (float64x1_t)(__p0);
57603   return __ret;
57604 }
vreinterpret_f32_p8(poly8x8_t __p0)57605 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
57606   float32x2_t __ret;
57607   __ret = (float32x2_t)(__p0);
57608   return __ret;
57609 }
vreinterpret_f32_p64(poly64x1_t __p0)57610 __ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
57611   float32x2_t __ret;
57612   __ret = (float32x2_t)(__p0);
57613   return __ret;
57614 }
vreinterpret_f32_p16(poly16x4_t __p0)57615 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
57616   float32x2_t __ret;
57617   __ret = (float32x2_t)(__p0);
57618   return __ret;
57619 }
vreinterpret_f32_u8(uint8x8_t __p0)57620 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
57621   float32x2_t __ret;
57622   __ret = (float32x2_t)(__p0);
57623   return __ret;
57624 }
vreinterpret_f32_u32(uint32x2_t __p0)57625 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
57626   float32x2_t __ret;
57627   __ret = (float32x2_t)(__p0);
57628   return __ret;
57629 }
vreinterpret_f32_u64(uint64x1_t __p0)57630 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
57631   float32x2_t __ret;
57632   __ret = (float32x2_t)(__p0);
57633   return __ret;
57634 }
vreinterpret_f32_u16(uint16x4_t __p0)57635 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
57636   float32x2_t __ret;
57637   __ret = (float32x2_t)(__p0);
57638   return __ret;
57639 }
vreinterpret_f32_s8(int8x8_t __p0)57640 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
57641   float32x2_t __ret;
57642   __ret = (float32x2_t)(__p0);
57643   return __ret;
57644 }
vreinterpret_f32_f64(float64x1_t __p0)57645 __ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
57646   float32x2_t __ret;
57647   __ret = (float32x2_t)(__p0);
57648   return __ret;
57649 }
vreinterpret_f32_f16(float16x4_t __p0)57650 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
57651   float32x2_t __ret;
57652   __ret = (float32x2_t)(__p0);
57653   return __ret;
57654 }
vreinterpret_f32_s32(int32x2_t __p0)57655 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
57656   float32x2_t __ret;
57657   __ret = (float32x2_t)(__p0);
57658   return __ret;
57659 }
vreinterpret_f32_s64(int64x1_t __p0)57660 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
57661   float32x2_t __ret;
57662   __ret = (float32x2_t)(__p0);
57663   return __ret;
57664 }
vreinterpret_f32_s16(int16x4_t __p0)57665 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
57666   float32x2_t __ret;
57667   __ret = (float32x2_t)(__p0);
57668   return __ret;
57669 }
vreinterpret_f16_p8(poly8x8_t __p0)57670 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
57671   float16x4_t __ret;
57672   __ret = (float16x4_t)(__p0);
57673   return __ret;
57674 }
vreinterpret_f16_p64(poly64x1_t __p0)57675 __ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
57676   float16x4_t __ret;
57677   __ret = (float16x4_t)(__p0);
57678   return __ret;
57679 }
vreinterpret_f16_p16(poly16x4_t __p0)57680 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
57681   float16x4_t __ret;
57682   __ret = (float16x4_t)(__p0);
57683   return __ret;
57684 }
vreinterpret_f16_u8(uint8x8_t __p0)57685 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
57686   float16x4_t __ret;
57687   __ret = (float16x4_t)(__p0);
57688   return __ret;
57689 }
vreinterpret_f16_u32(uint32x2_t __p0)57690 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
57691   float16x4_t __ret;
57692   __ret = (float16x4_t)(__p0);
57693   return __ret;
57694 }
vreinterpret_f16_u64(uint64x1_t __p0)57695 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
57696   float16x4_t __ret;
57697   __ret = (float16x4_t)(__p0);
57698   return __ret;
57699 }
vreinterpret_f16_u16(uint16x4_t __p0)57700 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
57701   float16x4_t __ret;
57702   __ret = (float16x4_t)(__p0);
57703   return __ret;
57704 }
vreinterpret_f16_s8(int8x8_t __p0)57705 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
57706   float16x4_t __ret;
57707   __ret = (float16x4_t)(__p0);
57708   return __ret;
57709 }
vreinterpret_f16_f64(float64x1_t __p0)57710 __ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
57711   float16x4_t __ret;
57712   __ret = (float16x4_t)(__p0);
57713   return __ret;
57714 }
vreinterpret_f16_f32(float32x2_t __p0)57715 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
57716   float16x4_t __ret;
57717   __ret = (float16x4_t)(__p0);
57718   return __ret;
57719 }
vreinterpret_f16_s32(int32x2_t __p0)57720 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
57721   float16x4_t __ret;
57722   __ret = (float16x4_t)(__p0);
57723   return __ret;
57724 }
vreinterpret_f16_s64(int64x1_t __p0)57725 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
57726   float16x4_t __ret;
57727   __ret = (float16x4_t)(__p0);
57728   return __ret;
57729 }
vreinterpret_f16_s16(int16x4_t __p0)57730 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
57731   float16x4_t __ret;
57732   __ret = (float16x4_t)(__p0);
57733   return __ret;
57734 }
vreinterpret_s32_p8(poly8x8_t __p0)57735 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
57736   int32x2_t __ret;
57737   __ret = (int32x2_t)(__p0);
57738   return __ret;
57739 }
vreinterpret_s32_p64(poly64x1_t __p0)57740 __ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
57741   int32x2_t __ret;
57742   __ret = (int32x2_t)(__p0);
57743   return __ret;
57744 }
vreinterpret_s32_p16(poly16x4_t __p0)57745 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
57746   int32x2_t __ret;
57747   __ret = (int32x2_t)(__p0);
57748   return __ret;
57749 }
vreinterpret_s32_u8(uint8x8_t __p0)57750 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
57751   int32x2_t __ret;
57752   __ret = (int32x2_t)(__p0);
57753   return __ret;
57754 }
vreinterpret_s32_u32(uint32x2_t __p0)57755 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
57756   int32x2_t __ret;
57757   __ret = (int32x2_t)(__p0);
57758   return __ret;
57759 }
vreinterpret_s32_u64(uint64x1_t __p0)57760 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
57761   int32x2_t __ret;
57762   __ret = (int32x2_t)(__p0);
57763   return __ret;
57764 }
vreinterpret_s32_u16(uint16x4_t __p0)57765 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
57766   int32x2_t __ret;
57767   __ret = (int32x2_t)(__p0);
57768   return __ret;
57769 }
vreinterpret_s32_s8(int8x8_t __p0)57770 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
57771   int32x2_t __ret;
57772   __ret = (int32x2_t)(__p0);
57773   return __ret;
57774 }
vreinterpret_s32_f64(float64x1_t __p0)57775 __ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
57776   int32x2_t __ret;
57777   __ret = (int32x2_t)(__p0);
57778   return __ret;
57779 }
vreinterpret_s32_f32(float32x2_t __p0)57780 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
57781   int32x2_t __ret;
57782   __ret = (int32x2_t)(__p0);
57783   return __ret;
57784 }
vreinterpret_s32_f16(float16x4_t __p0)57785 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
57786   int32x2_t __ret;
57787   __ret = (int32x2_t)(__p0);
57788   return __ret;
57789 }
vreinterpret_s32_s64(int64x1_t __p0)57790 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
57791   int32x2_t __ret;
57792   __ret = (int32x2_t)(__p0);
57793   return __ret;
57794 }
vreinterpret_s32_s16(int16x4_t __p0)57795 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
57796   int32x2_t __ret;
57797   __ret = (int32x2_t)(__p0);
57798   return __ret;
57799 }
vreinterpret_s64_p8(poly8x8_t __p0)57800 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
57801   int64x1_t __ret;
57802   __ret = (int64x1_t)(__p0);
57803   return __ret;
57804 }
vreinterpret_s64_p64(poly64x1_t __p0)57805 __ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
57806   int64x1_t __ret;
57807   __ret = (int64x1_t)(__p0);
57808   return __ret;
57809 }
vreinterpret_s64_p16(poly16x4_t __p0)57810 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
57811   int64x1_t __ret;
57812   __ret = (int64x1_t)(__p0);
57813   return __ret;
57814 }
vreinterpret_s64_u8(uint8x8_t __p0)57815 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
57816   int64x1_t __ret;
57817   __ret = (int64x1_t)(__p0);
57818   return __ret;
57819 }
vreinterpret_s64_u32(uint32x2_t __p0)57820 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
57821   int64x1_t __ret;
57822   __ret = (int64x1_t)(__p0);
57823   return __ret;
57824 }
vreinterpret_s64_u64(uint64x1_t __p0)57825 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
57826   int64x1_t __ret;
57827   __ret = (int64x1_t)(__p0);
57828   return __ret;
57829 }
vreinterpret_s64_u16(uint16x4_t __p0)57830 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
57831   int64x1_t __ret;
57832   __ret = (int64x1_t)(__p0);
57833   return __ret;
57834 }
vreinterpret_s64_s8(int8x8_t __p0)57835 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
57836   int64x1_t __ret;
57837   __ret = (int64x1_t)(__p0);
57838   return __ret;
57839 }
vreinterpret_s64_f64(float64x1_t __p0)57840 __ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
57841   int64x1_t __ret;
57842   __ret = (int64x1_t)(__p0);
57843   return __ret;
57844 }
vreinterpret_s64_f32(float32x2_t __p0)57845 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
57846   int64x1_t __ret;
57847   __ret = (int64x1_t)(__p0);
57848   return __ret;
57849 }
vreinterpret_s64_f16(float16x4_t __p0)57850 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
57851   int64x1_t __ret;
57852   __ret = (int64x1_t)(__p0);
57853   return __ret;
57854 }
vreinterpret_s64_s32(int32x2_t __p0)57855 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
57856   int64x1_t __ret;
57857   __ret = (int64x1_t)(__p0);
57858   return __ret;
57859 }
vreinterpret_s64_s16(int16x4_t __p0)57860 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
57861   int64x1_t __ret;
57862   __ret = (int64x1_t)(__p0);
57863   return __ret;
57864 }
vreinterpret_s16_p8(poly8x8_t __p0)57865 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
57866   int16x4_t __ret;
57867   __ret = (int16x4_t)(__p0);
57868   return __ret;
57869 }
vreinterpret_s16_p64(poly64x1_t __p0)57870 __ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
57871   int16x4_t __ret;
57872   __ret = (int16x4_t)(__p0);
57873   return __ret;
57874 }
vreinterpret_s16_p16(poly16x4_t __p0)57875 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
57876   int16x4_t __ret;
57877   __ret = (int16x4_t)(__p0);
57878   return __ret;
57879 }
vreinterpret_s16_u8(uint8x8_t __p0)57880 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
57881   int16x4_t __ret;
57882   __ret = (int16x4_t)(__p0);
57883   return __ret;
57884 }
vreinterpret_s16_u32(uint32x2_t __p0)57885 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
57886   int16x4_t __ret;
57887   __ret = (int16x4_t)(__p0);
57888   return __ret;
57889 }
vreinterpret_s16_u64(uint64x1_t __p0)57890 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
57891   int16x4_t __ret;
57892   __ret = (int16x4_t)(__p0);
57893   return __ret;
57894 }
vreinterpret_s16_u16(uint16x4_t __p0)57895 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
57896   int16x4_t __ret;
57897   __ret = (int16x4_t)(__p0);
57898   return __ret;
57899 }
vreinterpret_s16_s8(int8x8_t __p0)57900 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
57901   int16x4_t __ret;
57902   __ret = (int16x4_t)(__p0);
57903   return __ret;
57904 }
vreinterpret_s16_f64(float64x1_t __p0)57905 __ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
57906   int16x4_t __ret;
57907   __ret = (int16x4_t)(__p0);
57908   return __ret;
57909 }
vreinterpret_s16_f32(float32x2_t __p0)57910 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
57911   int16x4_t __ret;
57912   __ret = (int16x4_t)(__p0);
57913   return __ret;
57914 }
vreinterpret_s16_f16(float16x4_t __p0)57915 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
57916   int16x4_t __ret;
57917   __ret = (int16x4_t)(__p0);
57918   return __ret;
57919 }
vreinterpret_s16_s32(int32x2_t __p0)57920 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
57921   int16x4_t __ret;
57922   __ret = (int16x4_t)(__p0);
57923   return __ret;
57924 }
vreinterpret_s16_s64(int64x1_t __p0)57925 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
57926   int16x4_t __ret;
57927   __ret = (int16x4_t)(__p0);
57928   return __ret;
57929 }
vrshld_u64(uint64_t __p0,int64_t __p1)57930 __ai uint64_t vrshld_u64(uint64_t __p0, int64_t __p1) {
57931   uint64_t __ret;
57932   __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
57933   return __ret;
57934 }
vrshld_s64(int64_t __p0,int64_t __p1)57935 __ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
57936   int64_t __ret;
57937   __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
57938   return __ret;
57939 }
57940 #define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
57941   uint64_t __ret; \
57942   uint64_t __s0 = __p0; \
57943   __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
57944   __ret; \
57945 })
57946 #define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
57947   int64_t __ret; \
57948   int64_t __s0 = __p0; \
57949   __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
57950   __ret; \
57951 })
57952 #ifdef __LITTLE_ENDIAN__
57953 #define vrshrn_high_n_u32(__p0_703, __p1_703, __p2_703) __extension__ ({ \
57954   uint16x8_t __ret_703; \
57955   uint16x4_t __s0_703 = __p0_703; \
57956   uint32x4_t __s1_703 = __p1_703; \
57957   __ret_703 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_703), (uint16x4_t)(vrshrn_n_u32(__s1_703, __p2_703)))); \
57958   __ret_703; \
57959 })
57960 #else
57961 #define vrshrn_high_n_u32(__p0_704, __p1_704, __p2_704) __extension__ ({ \
57962   uint16x8_t __ret_704; \
57963   uint16x4_t __s0_704 = __p0_704; \
57964   uint32x4_t __s1_704 = __p1_704; \
57965   uint16x4_t __rev0_704;  __rev0_704 = __builtin_shufflevector(__s0_704, __s0_704, 3, 2, 1, 0); \
57966   uint32x4_t __rev1_704;  __rev1_704 = __builtin_shufflevector(__s1_704, __s1_704, 3, 2, 1, 0); \
57967   __ret_704 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_704), (uint16x4_t)(__noswap_vrshrn_n_u32(__rev1_704, __p2_704)))); \
57968   __ret_704 = __builtin_shufflevector(__ret_704, __ret_704, 7, 6, 5, 4, 3, 2, 1, 0); \
57969   __ret_704; \
57970 })
57971 #endif
57972 
57973 #ifdef __LITTLE_ENDIAN__
57974 #define vrshrn_high_n_u64(__p0_705, __p1_705, __p2_705) __extension__ ({ \
57975   uint32x4_t __ret_705; \
57976   uint32x2_t __s0_705 = __p0_705; \
57977   uint64x2_t __s1_705 = __p1_705; \
57978   __ret_705 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_705), (uint32x2_t)(vrshrn_n_u64(__s1_705, __p2_705)))); \
57979   __ret_705; \
57980 })
57981 #else
57982 #define vrshrn_high_n_u64(__p0_706, __p1_706, __p2_706) __extension__ ({ \
57983   uint32x4_t __ret_706; \
57984   uint32x2_t __s0_706 = __p0_706; \
57985   uint64x2_t __s1_706 = __p1_706; \
57986   uint32x2_t __rev0_706;  __rev0_706 = __builtin_shufflevector(__s0_706, __s0_706, 1, 0); \
57987   uint64x2_t __rev1_706;  __rev1_706 = __builtin_shufflevector(__s1_706, __s1_706, 1, 0); \
57988   __ret_706 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_706), (uint32x2_t)(__noswap_vrshrn_n_u64(__rev1_706, __p2_706)))); \
57989   __ret_706 = __builtin_shufflevector(__ret_706, __ret_706, 3, 2, 1, 0); \
57990   __ret_706; \
57991 })
57992 #endif
57993 
57994 #ifdef __LITTLE_ENDIAN__
57995 #define vrshrn_high_n_u16(__p0_707, __p1_707, __p2_707) __extension__ ({ \
57996   uint8x16_t __ret_707; \
57997   uint8x8_t __s0_707 = __p0_707; \
57998   uint16x8_t __s1_707 = __p1_707; \
57999   __ret_707 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_707), (uint8x8_t)(vrshrn_n_u16(__s1_707, __p2_707)))); \
58000   __ret_707; \
58001 })
58002 #else
58003 #define vrshrn_high_n_u16(__p0_708, __p1_708, __p2_708) __extension__ ({ \
58004   uint8x16_t __ret_708; \
58005   uint8x8_t __s0_708 = __p0_708; \
58006   uint16x8_t __s1_708 = __p1_708; \
58007   uint8x8_t __rev0_708;  __rev0_708 = __builtin_shufflevector(__s0_708, __s0_708, 7, 6, 5, 4, 3, 2, 1, 0); \
58008   uint16x8_t __rev1_708;  __rev1_708 = __builtin_shufflevector(__s1_708, __s1_708, 7, 6, 5, 4, 3, 2, 1, 0); \
58009   __ret_708 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_708), (uint8x8_t)(__noswap_vrshrn_n_u16(__rev1_708, __p2_708)))); \
58010   __ret_708 = __builtin_shufflevector(__ret_708, __ret_708, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58011   __ret_708; \
58012 })
58013 #endif
58014 
58015 #ifdef __LITTLE_ENDIAN__
58016 #define vrshrn_high_n_s32(__p0_709, __p1_709, __p2_709) __extension__ ({ \
58017   int16x8_t __ret_709; \
58018   int16x4_t __s0_709 = __p0_709; \
58019   int32x4_t __s1_709 = __p1_709; \
58020   __ret_709 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_709), (int16x4_t)(vrshrn_n_s32(__s1_709, __p2_709)))); \
58021   __ret_709; \
58022 })
58023 #else
58024 #define vrshrn_high_n_s32(__p0_710, __p1_710, __p2_710) __extension__ ({ \
58025   int16x8_t __ret_710; \
58026   int16x4_t __s0_710 = __p0_710; \
58027   int32x4_t __s1_710 = __p1_710; \
58028   int16x4_t __rev0_710;  __rev0_710 = __builtin_shufflevector(__s0_710, __s0_710, 3, 2, 1, 0); \
58029   int32x4_t __rev1_710;  __rev1_710 = __builtin_shufflevector(__s1_710, __s1_710, 3, 2, 1, 0); \
58030   __ret_710 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_710), (int16x4_t)(__noswap_vrshrn_n_s32(__rev1_710, __p2_710)))); \
58031   __ret_710 = __builtin_shufflevector(__ret_710, __ret_710, 7, 6, 5, 4, 3, 2, 1, 0); \
58032   __ret_710; \
58033 })
58034 #endif
58035 
58036 #ifdef __LITTLE_ENDIAN__
58037 #define vrshrn_high_n_s64(__p0_711, __p1_711, __p2_711) __extension__ ({ \
58038   int32x4_t __ret_711; \
58039   int32x2_t __s0_711 = __p0_711; \
58040   int64x2_t __s1_711 = __p1_711; \
58041   __ret_711 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_711), (int32x2_t)(vrshrn_n_s64(__s1_711, __p2_711)))); \
58042   __ret_711; \
58043 })
58044 #else
58045 #define vrshrn_high_n_s64(__p0_712, __p1_712, __p2_712) __extension__ ({ \
58046   int32x4_t __ret_712; \
58047   int32x2_t __s0_712 = __p0_712; \
58048   int64x2_t __s1_712 = __p1_712; \
58049   int32x2_t __rev0_712;  __rev0_712 = __builtin_shufflevector(__s0_712, __s0_712, 1, 0); \
58050   int64x2_t __rev1_712;  __rev1_712 = __builtin_shufflevector(__s1_712, __s1_712, 1, 0); \
58051   __ret_712 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_712), (int32x2_t)(__noswap_vrshrn_n_s64(__rev1_712, __p2_712)))); \
58052   __ret_712 = __builtin_shufflevector(__ret_712, __ret_712, 3, 2, 1, 0); \
58053   __ret_712; \
58054 })
58055 #endif
58056 
58057 #ifdef __LITTLE_ENDIAN__
58058 #define vrshrn_high_n_s16(__p0_713, __p1_713, __p2_713) __extension__ ({ \
58059   int8x16_t __ret_713; \
58060   int8x8_t __s0_713 = __p0_713; \
58061   int16x8_t __s1_713 = __p1_713; \
58062   __ret_713 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_713), (int8x8_t)(vrshrn_n_s16(__s1_713, __p2_713)))); \
58063   __ret_713; \
58064 })
58065 #else
58066 #define vrshrn_high_n_s16(__p0_714, __p1_714, __p2_714) __extension__ ({ \
58067   int8x16_t __ret_714; \
58068   int8x8_t __s0_714 = __p0_714; \
58069   int16x8_t __s1_714 = __p1_714; \
58070   int8x8_t __rev0_714;  __rev0_714 = __builtin_shufflevector(__s0_714, __s0_714, 7, 6, 5, 4, 3, 2, 1, 0); \
58071   int16x8_t __rev1_714;  __rev1_714 = __builtin_shufflevector(__s1_714, __s1_714, 7, 6, 5, 4, 3, 2, 1, 0); \
58072   __ret_714 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_714), (int8x8_t)(__noswap_vrshrn_n_s16(__rev1_714, __p2_714)))); \
58073   __ret_714 = __builtin_shufflevector(__ret_714, __ret_714, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58074   __ret_714; \
58075 })
58076 #endif
58077 
58078 #ifdef __LITTLE_ENDIAN__
vrsqrteq_f64(float64x2_t __p0)58079 __ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
58080   float64x2_t __ret;
58081   __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 42);
58082   return __ret;
58083 }
58084 #else
vrsqrteq_f64(float64x2_t __p0)58085 __ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
58086   float64x2_t __ret;
58087   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58088   __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 42);
58089   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
58090   return __ret;
58091 }
58092 #endif
58093 
vrsqrte_f64(float64x1_t __p0)58094 __ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
58095   float64x1_t __ret;
58096   __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
58097   return __ret;
58098 }
vrsqrted_f64(float64_t __p0)58099 __ai float64_t vrsqrted_f64(float64_t __p0) {
58100   float64_t __ret;
58101   __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
58102   return __ret;
58103 }
vrsqrtes_f32(float32_t __p0)58104 __ai float32_t vrsqrtes_f32(float32_t __p0) {
58105   float32_t __ret;
58106   __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
58107   return __ret;
58108 }
58109 #ifdef __LITTLE_ENDIAN__
vrsqrtsq_f64(float64x2_t __p0,float64x2_t __p1)58110 __ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
58111   float64x2_t __ret;
58112   __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
58113   return __ret;
58114 }
58115 #else
vrsqrtsq_f64(float64x2_t __p0,float64x2_t __p1)58116 __ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
58117   float64x2_t __ret;
58118   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58119   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
58120   __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
58121   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
58122   return __ret;
58123 }
58124 #endif
58125 
vrsqrts_f64(float64x1_t __p0,float64x1_t __p1)58126 __ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
58127   float64x1_t __ret;
58128   __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
58129   return __ret;
58130 }
vrsqrtsd_f64(float64_t __p0,float64_t __p1)58131 __ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
58132   float64_t __ret;
58133   __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
58134   return __ret;
58135 }
vrsqrtss_f32(float32_t __p0,float32_t __p1)58136 __ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
58137   float32_t __ret;
58138   __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
58139   return __ret;
58140 }
58141 #define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
58142   uint64_t __ret; \
58143   uint64_t __s0 = __p0; \
58144   uint64_t __s1 = __p1; \
58145   __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
58146   __ret; \
58147 })
58148 #define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
58149   int64_t __ret; \
58150   int64_t __s0 = __p0; \
58151   int64_t __s1 = __p1; \
58152   __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
58153   __ret; \
58154 })
58155 #ifdef __LITTLE_ENDIAN__
vrsubhn_high_u32(uint16x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)58156 __ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
58157   uint16x8_t __ret;
58158   __ret = vcombine_u16(__p0, vrsubhn_u32(__p1, __p2));
58159   return __ret;
58160 }
58161 #else
vrsubhn_high_u32(uint16x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)58162 __ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
58163   uint16x8_t __ret;
58164   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
58165   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
58166   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
58167   __ret = __noswap_vcombine_u16(__rev0, __noswap_vrsubhn_u32(__rev1, __rev2));
58168   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58169   return __ret;
58170 }
58171 #endif
58172 
58173 #ifdef __LITTLE_ENDIAN__
vrsubhn_high_u64(uint32x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)58174 __ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
58175   uint32x4_t __ret;
58176   __ret = vcombine_u32(__p0, vrsubhn_u64(__p1, __p2));
58177   return __ret;
58178 }
58179 #else
vrsubhn_high_u64(uint32x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)58180 __ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
58181   uint32x4_t __ret;
58182   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58183   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
58184   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
58185   __ret = __noswap_vcombine_u32(__rev0, __noswap_vrsubhn_u64(__rev1, __rev2));
58186   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
58187   return __ret;
58188 }
58189 #endif
58190 
58191 #ifdef __LITTLE_ENDIAN__
vrsubhn_high_u16(uint8x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)58192 __ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
58193   uint8x16_t __ret;
58194   __ret = vcombine_u8(__p0, vrsubhn_u16(__p1, __p2));
58195   return __ret;
58196 }
58197 #else
vrsubhn_high_u16(uint8x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)58198 __ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
58199   uint8x16_t __ret;
58200   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58201   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58202   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58203   __ret = __noswap_vcombine_u8(__rev0, __noswap_vrsubhn_u16(__rev1, __rev2));
58204   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58205   return __ret;
58206 }
58207 #endif
58208 
58209 #ifdef __LITTLE_ENDIAN__
vrsubhn_high_s32(int16x4_t __p0,int32x4_t __p1,int32x4_t __p2)58210 __ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
58211   int16x8_t __ret;
58212   __ret = vcombine_s16(__p0, vrsubhn_s32(__p1, __p2));
58213   return __ret;
58214 }
58215 #else
vrsubhn_high_s32(int16x4_t __p0,int32x4_t __p1,int32x4_t __p2)58216 __ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
58217   int16x8_t __ret;
58218   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
58219   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
58220   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
58221   __ret = __noswap_vcombine_s16(__rev0, __noswap_vrsubhn_s32(__rev1, __rev2));
58222   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58223   return __ret;
58224 }
58225 #endif
58226 
58227 #ifdef __LITTLE_ENDIAN__
vrsubhn_high_s64(int32x2_t __p0,int64x2_t __p1,int64x2_t __p2)58228 __ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
58229   int32x4_t __ret;
58230   __ret = vcombine_s32(__p0, vrsubhn_s64(__p1, __p2));
58231   return __ret;
58232 }
58233 #else
vrsubhn_high_s64(int32x2_t __p0,int64x2_t __p1,int64x2_t __p2)58234 __ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
58235   int32x4_t __ret;
58236   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58237   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
58238   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
58239   __ret = __noswap_vcombine_s32(__rev0, __noswap_vrsubhn_s64(__rev1, __rev2));
58240   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
58241   return __ret;
58242 }
58243 #endif
58244 
58245 #ifdef __LITTLE_ENDIAN__
vrsubhn_high_s16(int8x8_t __p0,int16x8_t __p1,int16x8_t __p2)58246 __ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
58247   int8x16_t __ret;
58248   __ret = vcombine_s8(__p0, vrsubhn_s16(__p1, __p2));
58249   return __ret;
58250 }
58251 #else
vrsubhn_high_s16(int8x8_t __p0,int16x8_t __p1,int16x8_t __p2)58252 __ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
58253   int8x16_t __ret;
58254   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58255   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58256   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58257   __ret = __noswap_vcombine_s8(__rev0, __noswap_vrsubhn_s16(__rev1, __rev2));
58258   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58259   return __ret;
58260 }
58261 #endif
58262 
58263 #define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
58264   poly64x1_t __ret; \
58265   poly64_t __s0 = __p0; \
58266   poly64x1_t __s1 = __p1; \
58267   __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (poly64x1_t)__s1, __p2); \
58268   __ret; \
58269 })
58270 #ifdef __LITTLE_ENDIAN__
58271 #define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
58272   poly64x2_t __ret; \
58273   poly64_t __s0 = __p0; \
58274   poly64x2_t __s1 = __p1; \
58275   __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (poly64x2_t)__s1, __p2); \
58276   __ret; \
58277 })
58278 #else
58279 #define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
58280   poly64x2_t __ret; \
58281   poly64_t __s0 = __p0; \
58282   poly64x2_t __s1 = __p1; \
58283   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
58284   __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (poly64x2_t)__rev1, __p2); \
58285   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
58286   __ret; \
58287 })
58288 #define __noswap_vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
58289   poly64x2_t __ret; \
58290   poly64_t __s0 = __p0; \
58291   poly64x2_t __s1 = __p1; \
58292   __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (poly64x2_t)__s1, __p2); \
58293   __ret; \
58294 })
58295 #endif
58296 
58297 #ifdef __LITTLE_ENDIAN__
58298 #define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
58299   float64x2_t __ret; \
58300   float64_t __s0 = __p0; \
58301   float64x2_t __s1 = __p1; \
58302   __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (float64x2_t)__s1, __p2); \
58303   __ret; \
58304 })
58305 #else
58306 #define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
58307   float64x2_t __ret; \
58308   float64_t __s0 = __p0; \
58309   float64x2_t __s1 = __p1; \
58310   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
58311   __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (float64x2_t)__rev1, __p2); \
58312   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
58313   __ret; \
58314 })
58315 #define __noswap_vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
58316   float64x2_t __ret; \
58317   float64_t __s0 = __p0; \
58318   float64x2_t __s1 = __p1; \
58319   __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (float64x2_t)__s1, __p2); \
58320   __ret; \
58321 })
58322 #endif
58323 
58324 #define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
58325   float64x1_t __ret; \
58326   float64_t __s0 = __p0; \
58327   float64x1_t __s1 = __p1; \
58328   __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (float64x1_t)__s1, __p2); \
58329   __ret; \
58330 })
vshld_u64(uint64_t __p0,int64_t __p1)58331 __ai uint64_t vshld_u64(uint64_t __p0, int64_t __p1) {
58332   uint64_t __ret;
58333   __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
58334   return __ret;
58335 }
vshld_s64(int64_t __p0,int64_t __p1)58336 __ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
58337   int64_t __ret;
58338   __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
58339   return __ret;
58340 }
58341 #define vshld_n_u64(__p0, __p1) __extension__ ({ \
58342   uint64_t __ret; \
58343   uint64_t __s0 = __p0; \
58344   __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
58345   __ret; \
58346 })
58347 #define vshld_n_s64(__p0, __p1) __extension__ ({ \
58348   int64_t __ret; \
58349   int64_t __s0 = __p0; \
58350   __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
58351   __ret; \
58352 })
58353 #ifdef __LITTLE_ENDIAN__
58354 #define vshll_high_n_u8(__p0_715, __p1_715) __extension__ ({ \
58355   uint16x8_t __ret_715; \
58356   uint8x16_t __s0_715 = __p0_715; \
58357   __ret_715 = (uint16x8_t)(vshll_n_u8(vget_high_u8(__s0_715), __p1_715)); \
58358   __ret_715; \
58359 })
58360 #else
58361 #define vshll_high_n_u8(__p0_716, __p1_716) __extension__ ({ \
58362   uint16x8_t __ret_716; \
58363   uint8x16_t __s0_716 = __p0_716; \
58364   uint8x16_t __rev0_716;  __rev0_716 = __builtin_shufflevector(__s0_716, __s0_716, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58365   __ret_716 = (uint16x8_t)(__noswap_vshll_n_u8(__noswap_vget_high_u8(__rev0_716), __p1_716)); \
58366   __ret_716 = __builtin_shufflevector(__ret_716, __ret_716, 7, 6, 5, 4, 3, 2, 1, 0); \
58367   __ret_716; \
58368 })
58369 #endif
58370 
58371 #ifdef __LITTLE_ENDIAN__
58372 #define vshll_high_n_u32(__p0_717, __p1_717) __extension__ ({ \
58373   uint64x2_t __ret_717; \
58374   uint32x4_t __s0_717 = __p0_717; \
58375   __ret_717 = (uint64x2_t)(vshll_n_u32(vget_high_u32(__s0_717), __p1_717)); \
58376   __ret_717; \
58377 })
58378 #else
58379 #define vshll_high_n_u32(__p0_718, __p1_718) __extension__ ({ \
58380   uint64x2_t __ret_718; \
58381   uint32x4_t __s0_718 = __p0_718; \
58382   uint32x4_t __rev0_718;  __rev0_718 = __builtin_shufflevector(__s0_718, __s0_718, 3, 2, 1, 0); \
58383   __ret_718 = (uint64x2_t)(__noswap_vshll_n_u32(__noswap_vget_high_u32(__rev0_718), __p1_718)); \
58384   __ret_718 = __builtin_shufflevector(__ret_718, __ret_718, 1, 0); \
58385   __ret_718; \
58386 })
58387 #endif
58388 
58389 #ifdef __LITTLE_ENDIAN__
58390 #define vshll_high_n_u16(__p0_719, __p1_719) __extension__ ({ \
58391   uint32x4_t __ret_719; \
58392   uint16x8_t __s0_719 = __p0_719; \
58393   __ret_719 = (uint32x4_t)(vshll_n_u16(vget_high_u16(__s0_719), __p1_719)); \
58394   __ret_719; \
58395 })
58396 #else
58397 #define vshll_high_n_u16(__p0_720, __p1_720) __extension__ ({ \
58398   uint32x4_t __ret_720; \
58399   uint16x8_t __s0_720 = __p0_720; \
58400   uint16x8_t __rev0_720;  __rev0_720 = __builtin_shufflevector(__s0_720, __s0_720, 7, 6, 5, 4, 3, 2, 1, 0); \
58401   __ret_720 = (uint32x4_t)(__noswap_vshll_n_u16(__noswap_vget_high_u16(__rev0_720), __p1_720)); \
58402   __ret_720 = __builtin_shufflevector(__ret_720, __ret_720, 3, 2, 1, 0); \
58403   __ret_720; \
58404 })
58405 #endif
58406 
58407 #ifdef __LITTLE_ENDIAN__
58408 #define vshll_high_n_s8(__p0_721, __p1_721) __extension__ ({ \
58409   int16x8_t __ret_721; \
58410   int8x16_t __s0_721 = __p0_721; \
58411   __ret_721 = (int16x8_t)(vshll_n_s8(vget_high_s8(__s0_721), __p1_721)); \
58412   __ret_721; \
58413 })
58414 #else
58415 #define vshll_high_n_s8(__p0_722, __p1_722) __extension__ ({ \
58416   int16x8_t __ret_722; \
58417   int8x16_t __s0_722 = __p0_722; \
58418   int8x16_t __rev0_722;  __rev0_722 = __builtin_shufflevector(__s0_722, __s0_722, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58419   __ret_722 = (int16x8_t)(__noswap_vshll_n_s8(__noswap_vget_high_s8(__rev0_722), __p1_722)); \
58420   __ret_722 = __builtin_shufflevector(__ret_722, __ret_722, 7, 6, 5, 4, 3, 2, 1, 0); \
58421   __ret_722; \
58422 })
58423 #endif
58424 
58425 #ifdef __LITTLE_ENDIAN__
58426 #define vshll_high_n_s32(__p0_723, __p1_723) __extension__ ({ \
58427   int64x2_t __ret_723; \
58428   int32x4_t __s0_723 = __p0_723; \
58429   __ret_723 = (int64x2_t)(vshll_n_s32(vget_high_s32(__s0_723), __p1_723)); \
58430   __ret_723; \
58431 })
58432 #else
58433 #define vshll_high_n_s32(__p0_724, __p1_724) __extension__ ({ \
58434   int64x2_t __ret_724; \
58435   int32x4_t __s0_724 = __p0_724; \
58436   int32x4_t __rev0_724;  __rev0_724 = __builtin_shufflevector(__s0_724, __s0_724, 3, 2, 1, 0); \
58437   __ret_724 = (int64x2_t)(__noswap_vshll_n_s32(__noswap_vget_high_s32(__rev0_724), __p1_724)); \
58438   __ret_724 = __builtin_shufflevector(__ret_724, __ret_724, 1, 0); \
58439   __ret_724; \
58440 })
58441 #endif
58442 
58443 #ifdef __LITTLE_ENDIAN__
58444 #define vshll_high_n_s16(__p0_725, __p1_725) __extension__ ({ \
58445   int32x4_t __ret_725; \
58446   int16x8_t __s0_725 = __p0_725; \
58447   __ret_725 = (int32x4_t)(vshll_n_s16(vget_high_s16(__s0_725), __p1_725)); \
58448   __ret_725; \
58449 })
58450 #else
58451 #define vshll_high_n_s16(__p0_726, __p1_726) __extension__ ({ \
58452   int32x4_t __ret_726; \
58453   int16x8_t __s0_726 = __p0_726; \
58454   int16x8_t __rev0_726;  __rev0_726 = __builtin_shufflevector(__s0_726, __s0_726, 7, 6, 5, 4, 3, 2, 1, 0); \
58455   __ret_726 = (int32x4_t)(__noswap_vshll_n_s16(__noswap_vget_high_s16(__rev0_726), __p1_726)); \
58456   __ret_726 = __builtin_shufflevector(__ret_726, __ret_726, 3, 2, 1, 0); \
58457   __ret_726; \
58458 })
58459 #endif
58460 
58461 #define vshrd_n_u64(__p0, __p1) __extension__ ({ \
58462   uint64_t __ret; \
58463   uint64_t __s0 = __p0; \
58464   __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
58465   __ret; \
58466 })
58467 #define vshrd_n_s64(__p0, __p1) __extension__ ({ \
58468   int64_t __ret; \
58469   int64_t __s0 = __p0; \
58470   __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
58471   __ret; \
58472 })
58473 #ifdef __LITTLE_ENDIAN__
58474 #define vshrn_high_n_u32(__p0_727, __p1_727, __p2_727) __extension__ ({ \
58475   uint16x8_t __ret_727; \
58476   uint16x4_t __s0_727 = __p0_727; \
58477   uint32x4_t __s1_727 = __p1_727; \
58478   __ret_727 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_727), (uint16x4_t)(vshrn_n_u32(__s1_727, __p2_727)))); \
58479   __ret_727; \
58480 })
58481 #else
58482 #define vshrn_high_n_u32(__p0_728, __p1_728, __p2_728) __extension__ ({ \
58483   uint16x8_t __ret_728; \
58484   uint16x4_t __s0_728 = __p0_728; \
58485   uint32x4_t __s1_728 = __p1_728; \
58486   uint16x4_t __rev0_728;  __rev0_728 = __builtin_shufflevector(__s0_728, __s0_728, 3, 2, 1, 0); \
58487   uint32x4_t __rev1_728;  __rev1_728 = __builtin_shufflevector(__s1_728, __s1_728, 3, 2, 1, 0); \
58488   __ret_728 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_728), (uint16x4_t)(__noswap_vshrn_n_u32(__rev1_728, __p2_728)))); \
58489   __ret_728 = __builtin_shufflevector(__ret_728, __ret_728, 7, 6, 5, 4, 3, 2, 1, 0); \
58490   __ret_728; \
58491 })
58492 #endif
58493 
58494 #ifdef __LITTLE_ENDIAN__
58495 #define vshrn_high_n_u64(__p0_729, __p1_729, __p2_729) __extension__ ({ \
58496   uint32x4_t __ret_729; \
58497   uint32x2_t __s0_729 = __p0_729; \
58498   uint64x2_t __s1_729 = __p1_729; \
58499   __ret_729 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_729), (uint32x2_t)(vshrn_n_u64(__s1_729, __p2_729)))); \
58500   __ret_729; \
58501 })
58502 #else
58503 #define vshrn_high_n_u64(__p0_730, __p1_730, __p2_730) __extension__ ({ \
58504   uint32x4_t __ret_730; \
58505   uint32x2_t __s0_730 = __p0_730; \
58506   uint64x2_t __s1_730 = __p1_730; \
58507   uint32x2_t __rev0_730;  __rev0_730 = __builtin_shufflevector(__s0_730, __s0_730, 1, 0); \
58508   uint64x2_t __rev1_730;  __rev1_730 = __builtin_shufflevector(__s1_730, __s1_730, 1, 0); \
58509   __ret_730 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_730), (uint32x2_t)(__noswap_vshrn_n_u64(__rev1_730, __p2_730)))); \
58510   __ret_730 = __builtin_shufflevector(__ret_730, __ret_730, 3, 2, 1, 0); \
58511   __ret_730; \
58512 })
58513 #endif
58514 
58515 #ifdef __LITTLE_ENDIAN__
58516 #define vshrn_high_n_u16(__p0_731, __p1_731, __p2_731) __extension__ ({ \
58517   uint8x16_t __ret_731; \
58518   uint8x8_t __s0_731 = __p0_731; \
58519   uint16x8_t __s1_731 = __p1_731; \
58520   __ret_731 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_731), (uint8x8_t)(vshrn_n_u16(__s1_731, __p2_731)))); \
58521   __ret_731; \
58522 })
58523 #else
58524 #define vshrn_high_n_u16(__p0_732, __p1_732, __p2_732) __extension__ ({ \
58525   uint8x16_t __ret_732; \
58526   uint8x8_t __s0_732 = __p0_732; \
58527   uint16x8_t __s1_732 = __p1_732; \
58528   uint8x8_t __rev0_732;  __rev0_732 = __builtin_shufflevector(__s0_732, __s0_732, 7, 6, 5, 4, 3, 2, 1, 0); \
58529   uint16x8_t __rev1_732;  __rev1_732 = __builtin_shufflevector(__s1_732, __s1_732, 7, 6, 5, 4, 3, 2, 1, 0); \
58530   __ret_732 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_732), (uint8x8_t)(__noswap_vshrn_n_u16(__rev1_732, __p2_732)))); \
58531   __ret_732 = __builtin_shufflevector(__ret_732, __ret_732, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58532   __ret_732; \
58533 })
58534 #endif
58535 
58536 #ifdef __LITTLE_ENDIAN__
58537 #define vshrn_high_n_s32(__p0_733, __p1_733, __p2_733) __extension__ ({ \
58538   int16x8_t __ret_733; \
58539   int16x4_t __s0_733 = __p0_733; \
58540   int32x4_t __s1_733 = __p1_733; \
58541   __ret_733 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_733), (int16x4_t)(vshrn_n_s32(__s1_733, __p2_733)))); \
58542   __ret_733; \
58543 })
58544 #else
58545 #define vshrn_high_n_s32(__p0_734, __p1_734, __p2_734) __extension__ ({ \
58546   int16x8_t __ret_734; \
58547   int16x4_t __s0_734 = __p0_734; \
58548   int32x4_t __s1_734 = __p1_734; \
58549   int16x4_t __rev0_734;  __rev0_734 = __builtin_shufflevector(__s0_734, __s0_734, 3, 2, 1, 0); \
58550   int32x4_t __rev1_734;  __rev1_734 = __builtin_shufflevector(__s1_734, __s1_734, 3, 2, 1, 0); \
58551   __ret_734 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_734), (int16x4_t)(__noswap_vshrn_n_s32(__rev1_734, __p2_734)))); \
58552   __ret_734 = __builtin_shufflevector(__ret_734, __ret_734, 7, 6, 5, 4, 3, 2, 1, 0); \
58553   __ret_734; \
58554 })
58555 #endif
58556 
58557 #ifdef __LITTLE_ENDIAN__
58558 #define vshrn_high_n_s64(__p0_735, __p1_735, __p2_735) __extension__ ({ \
58559   int32x4_t __ret_735; \
58560   int32x2_t __s0_735 = __p0_735; \
58561   int64x2_t __s1_735 = __p1_735; \
58562   __ret_735 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_735), (int32x2_t)(vshrn_n_s64(__s1_735, __p2_735)))); \
58563   __ret_735; \
58564 })
58565 #else
58566 #define vshrn_high_n_s64(__p0_736, __p1_736, __p2_736) __extension__ ({ \
58567   int32x4_t __ret_736; \
58568   int32x2_t __s0_736 = __p0_736; \
58569   int64x2_t __s1_736 = __p1_736; \
58570   int32x2_t __rev0_736;  __rev0_736 = __builtin_shufflevector(__s0_736, __s0_736, 1, 0); \
58571   int64x2_t __rev1_736;  __rev1_736 = __builtin_shufflevector(__s1_736, __s1_736, 1, 0); \
58572   __ret_736 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_736), (int32x2_t)(__noswap_vshrn_n_s64(__rev1_736, __p2_736)))); \
58573   __ret_736 = __builtin_shufflevector(__ret_736, __ret_736, 3, 2, 1, 0); \
58574   __ret_736; \
58575 })
58576 #endif
58577 
58578 #ifdef __LITTLE_ENDIAN__
58579 #define vshrn_high_n_s16(__p0_737, __p1_737, __p2_737) __extension__ ({ \
58580   int8x16_t __ret_737; \
58581   int8x8_t __s0_737 = __p0_737; \
58582   int16x8_t __s1_737 = __p1_737; \
58583   __ret_737 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_737), (int8x8_t)(vshrn_n_s16(__s1_737, __p2_737)))); \
58584   __ret_737; \
58585 })
58586 #else
58587 #define vshrn_high_n_s16(__p0_738, __p1_738, __p2_738) __extension__ ({ \
58588   int8x16_t __ret_738; \
58589   int8x8_t __s0_738 = __p0_738; \
58590   int16x8_t __s1_738 = __p1_738; \
58591   int8x8_t __rev0_738;  __rev0_738 = __builtin_shufflevector(__s0_738, __s0_738, 7, 6, 5, 4, 3, 2, 1, 0); \
58592   int16x8_t __rev1_738;  __rev1_738 = __builtin_shufflevector(__s1_738, __s1_738, 7, 6, 5, 4, 3, 2, 1, 0); \
58593   __ret_738 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_738), (int8x8_t)(__noswap_vshrn_n_s16(__rev1_738, __p2_738)))); \
58594   __ret_738 = __builtin_shufflevector(__ret_738, __ret_738, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58595   __ret_738; \
58596 })
58597 #endif
58598 
58599 #define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
58600   uint64_t __ret; \
58601   uint64_t __s0 = __p0; \
58602   uint64_t __s1 = __p1; \
58603   __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
58604   __ret; \
58605 })
58606 #define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
58607   int64_t __ret; \
58608   int64_t __s0 = __p0; \
58609   int64_t __s1 = __p1; \
58610   __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
58611   __ret; \
58612 })
58613 #define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
58614   poly64x1_t __ret; \
58615   poly64x1_t __s0 = __p0; \
58616   poly64x1_t __s1 = __p1; \
58617   __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
58618   __ret; \
58619 })
58620 #ifdef __LITTLE_ENDIAN__
58621 #define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
58622   poly64x2_t __ret; \
58623   poly64x2_t __s0 = __p0; \
58624   poly64x2_t __s1 = __p1; \
58625   __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
58626   __ret; \
58627 })
58628 #else
58629 #define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
58630   poly64x2_t __ret; \
58631   poly64x2_t __s0 = __p0; \
58632   poly64x2_t __s1 = __p1; \
58633   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
58634   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
58635   __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
58636   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
58637   __ret; \
58638 })
58639 #endif
58640 
vsqaddb_u8(uint8_t __p0,int8_t __p1)58641 __ai uint8_t vsqaddb_u8(uint8_t __p0, int8_t __p1) {
58642   uint8_t __ret;
58643   __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
58644   return __ret;
58645 }
vsqadds_u32(uint32_t __p0,int32_t __p1)58646 __ai uint32_t vsqadds_u32(uint32_t __p0, int32_t __p1) {
58647   uint32_t __ret;
58648   __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
58649   return __ret;
58650 }
vsqaddd_u64(uint64_t __p0,int64_t __p1)58651 __ai uint64_t vsqaddd_u64(uint64_t __p0, int64_t __p1) {
58652   uint64_t __ret;
58653   __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
58654   return __ret;
58655 }
vsqaddh_u16(uint16_t __p0,int16_t __p1)58656 __ai uint16_t vsqaddh_u16(uint16_t __p0, int16_t __p1) {
58657   uint16_t __ret;
58658   __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
58659   return __ret;
58660 }
58661 #ifdef __LITTLE_ENDIAN__
vsqaddq_u8(uint8x16_t __p0,int8x16_t __p1)58662 __ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, int8x16_t __p1) {
58663   uint8x16_t __ret;
58664   __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
58665   return __ret;
58666 }
58667 #else
vsqaddq_u8(uint8x16_t __p0,int8x16_t __p1)58668 __ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, int8x16_t __p1) {
58669   uint8x16_t __ret;
58670   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58671   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58672   __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
58673   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58674   return __ret;
58675 }
58676 #endif
58677 
58678 #ifdef __LITTLE_ENDIAN__
vsqaddq_u32(uint32x4_t __p0,int32x4_t __p1)58679 __ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, int32x4_t __p1) {
58680   uint32x4_t __ret;
58681   __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
58682   return __ret;
58683 }
58684 #else
vsqaddq_u32(uint32x4_t __p0,int32x4_t __p1)58685 __ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, int32x4_t __p1) {
58686   uint32x4_t __ret;
58687   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
58688   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
58689   __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
58690   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
58691   return __ret;
58692 }
58693 #endif
58694 
58695 #ifdef __LITTLE_ENDIAN__
vsqaddq_u64(uint64x2_t __p0,int64x2_t __p1)58696 __ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, int64x2_t __p1) {
58697   uint64x2_t __ret;
58698   __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
58699   return __ret;
58700 }
58701 #else
vsqaddq_u64(uint64x2_t __p0,int64x2_t __p1)58702 __ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, int64x2_t __p1) {
58703   uint64x2_t __ret;
58704   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58705   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
58706   __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
58707   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
58708   return __ret;
58709 }
58710 #endif
58711 
58712 #ifdef __LITTLE_ENDIAN__
vsqaddq_u16(uint16x8_t __p0,int16x8_t __p1)58713 __ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, int16x8_t __p1) {
58714   uint16x8_t __ret;
58715   __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
58716   return __ret;
58717 }
58718 #else
vsqaddq_u16(uint16x8_t __p0,int16x8_t __p1)58719 __ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, int16x8_t __p1) {
58720   uint16x8_t __ret;
58721   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58722   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58723   __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
58724   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58725   return __ret;
58726 }
58727 #endif
58728 
58729 #ifdef __LITTLE_ENDIAN__
vsqadd_u8(uint8x8_t __p0,int8x8_t __p1)58730 __ai uint8x8_t vsqadd_u8(uint8x8_t __p0, int8x8_t __p1) {
58731   uint8x8_t __ret;
58732   __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
58733   return __ret;
58734 }
58735 #else
vsqadd_u8(uint8x8_t __p0,int8x8_t __p1)58736 __ai uint8x8_t vsqadd_u8(uint8x8_t __p0, int8x8_t __p1) {
58737   uint8x8_t __ret;
58738   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58739   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58740   __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
58741   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58742   return __ret;
58743 }
58744 #endif
58745 
58746 #ifdef __LITTLE_ENDIAN__
vsqadd_u32(uint32x2_t __p0,int32x2_t __p1)58747 __ai uint32x2_t vsqadd_u32(uint32x2_t __p0, int32x2_t __p1) {
58748   uint32x2_t __ret;
58749   __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
58750   return __ret;
58751 }
58752 #else
vsqadd_u32(uint32x2_t __p0,int32x2_t __p1)58753 __ai uint32x2_t vsqadd_u32(uint32x2_t __p0, int32x2_t __p1) {
58754   uint32x2_t __ret;
58755   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58756   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
58757   __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
58758   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
58759   return __ret;
58760 }
58761 #endif
58762 
vsqadd_u64(uint64x1_t __p0,int64x1_t __p1)58763 __ai uint64x1_t vsqadd_u64(uint64x1_t __p0, int64x1_t __p1) {
58764   uint64x1_t __ret;
58765   __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
58766   return __ret;
58767 }
58768 #ifdef __LITTLE_ENDIAN__
vsqadd_u16(uint16x4_t __p0,int16x4_t __p1)58769 __ai uint16x4_t vsqadd_u16(uint16x4_t __p0, int16x4_t __p1) {
58770   uint16x4_t __ret;
58771   __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
58772   return __ret;
58773 }
58774 #else
vsqadd_u16(uint16x4_t __p0,int16x4_t __p1)58775 __ai uint16x4_t vsqadd_u16(uint16x4_t __p0, int16x4_t __p1) {
58776   uint16x4_t __ret;
58777   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
58778   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
58779   __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
58780   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
58781   return __ret;
58782 }
58783 #endif
58784 
58785 #ifdef __LITTLE_ENDIAN__
vsqrtq_f64(float64x2_t __p0)58786 __ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
58787   float64x2_t __ret;
58788   __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 42);
58789   return __ret;
58790 }
58791 #else
vsqrtq_f64(float64x2_t __p0)58792 __ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
58793   float64x2_t __ret;
58794   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58795   __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 42);
58796   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
58797   return __ret;
58798 }
58799 #endif
58800 
58801 #ifdef __LITTLE_ENDIAN__
vsqrtq_f32(float32x4_t __p0)58802 __ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
58803   float32x4_t __ret;
58804   __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 41);
58805   return __ret;
58806 }
58807 #else
vsqrtq_f32(float32x4_t __p0)58808 __ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
58809   float32x4_t __ret;
58810   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
58811   __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 41);
58812   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
58813   return __ret;
58814 }
58815 #endif
58816 
vsqrt_f64(float64x1_t __p0)58817 __ai float64x1_t vsqrt_f64(float64x1_t __p0) {
58818   float64x1_t __ret;
58819   __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
58820   return __ret;
58821 }
58822 #ifdef __LITTLE_ENDIAN__
vsqrt_f32(float32x2_t __p0)58823 __ai float32x2_t vsqrt_f32(float32x2_t __p0) {
58824   float32x2_t __ret;
58825   __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 9);
58826   return __ret;
58827 }
58828 #else
vsqrt_f32(float32x2_t __p0)58829 __ai float32x2_t vsqrt_f32(float32x2_t __p0) {
58830   float32x2_t __ret;
58831   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58832   __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__rev0, 9);
58833   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
58834   return __ret;
58835 }
58836 #endif
58837 
58838 #define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
58839   uint64_t __ret; \
58840   uint64_t __s0 = __p0; \
58841   uint64_t __s1 = __p1; \
58842   __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
58843   __ret; \
58844 })
58845 #define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
58846   int64_t __ret; \
58847   int64_t __s0 = __p0; \
58848   int64_t __s1 = __p1; \
58849   __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
58850   __ret; \
58851 })
58852 #define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
58853   uint64_t __ret; \
58854   uint64_t __s0 = __p0; \
58855   uint64_t __s1 = __p1; \
58856   __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
58857   __ret; \
58858 })
58859 #define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
58860   int64_t __ret; \
58861   int64_t __s0 = __p0; \
58862   int64_t __s1 = __p1; \
58863   __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
58864   __ret; \
58865 })
58866 #define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
58867   poly64x1_t __ret; \
58868   poly64x1_t __s0 = __p0; \
58869   poly64x1_t __s1 = __p1; \
58870   __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
58871   __ret; \
58872 })
58873 #ifdef __LITTLE_ENDIAN__
58874 #define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
58875   poly64x2_t __ret; \
58876   poly64x2_t __s0 = __p0; \
58877   poly64x2_t __s1 = __p1; \
58878   __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
58879   __ret; \
58880 })
58881 #else
58882 #define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
58883   poly64x2_t __ret; \
58884   poly64x2_t __s0 = __p0; \
58885   poly64x2_t __s1 = __p1; \
58886   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
58887   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
58888   __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
58889   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
58890   __ret; \
58891 })
58892 #endif
58893 
58894 #define vst1_p64(__p0, __p1) __extension__ ({ \
58895   poly64x1_t __s1 = __p1; \
58896   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
58897 })
58898 #ifdef __LITTLE_ENDIAN__
58899 #define vst1q_p64(__p0, __p1) __extension__ ({ \
58900   poly64x2_t __s1 = __p1; \
58901   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 38); \
58902 })
58903 #else
58904 #define vst1q_p64(__p0, __p1) __extension__ ({ \
58905   poly64x2_t __s1 = __p1; \
58906   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
58907   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 38); \
58908 })
58909 #endif
58910 
58911 #ifdef __LITTLE_ENDIAN__
58912 #define vst1q_f64(__p0, __p1) __extension__ ({ \
58913   float64x2_t __s1 = __p1; \
58914   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 42); \
58915 })
58916 #else
58917 #define vst1q_f64(__p0, __p1) __extension__ ({ \
58918   float64x2_t __s1 = __p1; \
58919   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
58920   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 42); \
58921 })
58922 #endif
58923 
58924 #define vst1_f64(__p0, __p1) __extension__ ({ \
58925   float64x1_t __s1 = __p1; \
58926   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
58927 })
58928 #define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
58929   poly64x1_t __s1 = __p1; \
58930   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
58931 })
58932 #ifdef __LITTLE_ENDIAN__
58933 #define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
58934   poly64x2_t __s1 = __p1; \
58935   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
58936 })
58937 #else
58938 #define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
58939   poly64x2_t __s1 = __p1; \
58940   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
58941   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
58942 })
58943 #endif
58944 
58945 #ifdef __LITTLE_ENDIAN__
58946 #define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
58947   float64x2_t __s1 = __p1; \
58948   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
58949 })
58950 #else
58951 #define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
58952   float64x2_t __s1 = __p1; \
58953   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
58954   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
58955 })
58956 #endif
58957 
58958 #define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
58959   float64x1_t __s1 = __p1; \
58960   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
58961 })
58962 #define vst1_p64_x2(__p0, __p1) __extension__ ({ \
58963   poly64x1x2_t __s1 = __p1; \
58964   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
58965 })
58966 #ifdef __LITTLE_ENDIAN__
58967 #define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
58968   poly64x2x2_t __s1 = __p1; \
58969   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
58970 })
58971 #else
58972 #define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
58973   poly64x2x2_t __s1 = __p1; \
58974   poly64x2x2_t __rev1; \
58975   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
58976   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
58977   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
58978 })
58979 #endif
58980 
58981 #ifdef __LITTLE_ENDIAN__
58982 #define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
58983   float64x2x2_t __s1 = __p1; \
58984   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 42); \
58985 })
58986 #else
58987 #define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
58988   float64x2x2_t __s1 = __p1; \
58989   float64x2x2_t __rev1; \
58990   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
58991   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
58992   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 42); \
58993 })
58994 #endif
58995 
58996 #define vst1_f64_x2(__p0, __p1) __extension__ ({ \
58997   float64x1x2_t __s1 = __p1; \
58998   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 10); \
58999 })
59000 #define vst1_p64_x3(__p0, __p1) __extension__ ({ \
59001   poly64x1x3_t __s1 = __p1; \
59002   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
59003 })
59004 #ifdef __LITTLE_ENDIAN__
59005 #define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
59006   poly64x2x3_t __s1 = __p1; \
59007   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
59008 })
59009 #else
59010 #define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
59011   poly64x2x3_t __s1 = __p1; \
59012   poly64x2x3_t __rev1; \
59013   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59014   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59015   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59016   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
59017 })
59018 #endif
59019 
59020 #ifdef __LITTLE_ENDIAN__
59021 #define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
59022   float64x2x3_t __s1 = __p1; \
59023   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 42); \
59024 })
59025 #else
59026 #define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
59027   float64x2x3_t __s1 = __p1; \
59028   float64x2x3_t __rev1; \
59029   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59030   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59031   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59032   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 42); \
59033 })
59034 #endif
59035 
59036 #define vst1_f64_x3(__p0, __p1) __extension__ ({ \
59037   float64x1x3_t __s1 = __p1; \
59038   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 10); \
59039 })
59040 #define vst1_p64_x4(__p0, __p1) __extension__ ({ \
59041   poly64x1x4_t __s1 = __p1; \
59042   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
59043 })
59044 #ifdef __LITTLE_ENDIAN__
59045 #define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
59046   poly64x2x4_t __s1 = __p1; \
59047   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
59048 })
59049 #else
59050 #define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
59051   poly64x2x4_t __s1 = __p1; \
59052   poly64x2x4_t __rev1; \
59053   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59054   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59055   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59056   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59057   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
59058 })
59059 #endif
59060 
59061 #ifdef __LITTLE_ENDIAN__
59062 #define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
59063   float64x2x4_t __s1 = __p1; \
59064   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 42); \
59065 })
59066 #else
59067 #define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
59068   float64x2x4_t __s1 = __p1; \
59069   float64x2x4_t __rev1; \
59070   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59071   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59072   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59073   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59074   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 42); \
59075 })
59076 #endif
59077 
59078 #define vst1_f64_x4(__p0, __p1) __extension__ ({ \
59079   float64x1x4_t __s1 = __p1; \
59080   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 10); \
59081 })
59082 #define vst2_p64(__p0, __p1) __extension__ ({ \
59083   poly64x1x2_t __s1 = __p1; \
59084   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
59085 })
59086 #ifdef __LITTLE_ENDIAN__
59087 #define vst2q_p64(__p0, __p1) __extension__ ({ \
59088   poly64x2x2_t __s1 = __p1; \
59089   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
59090 })
59091 #else
59092 #define vst2q_p64(__p0, __p1) __extension__ ({ \
59093   poly64x2x2_t __s1 = __p1; \
59094   poly64x2x2_t __rev1; \
59095   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59096   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59097   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
59098 })
59099 #endif
59100 
59101 #ifdef __LITTLE_ENDIAN__
59102 #define vst2q_u64(__p0, __p1) __extension__ ({ \
59103   uint64x2x2_t __s1 = __p1; \
59104   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
59105 })
59106 #else
59107 #define vst2q_u64(__p0, __p1) __extension__ ({ \
59108   uint64x2x2_t __s1 = __p1; \
59109   uint64x2x2_t __rev1; \
59110   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59111   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59112   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
59113 })
59114 #endif
59115 
59116 #ifdef __LITTLE_ENDIAN__
59117 #define vst2q_f64(__p0, __p1) __extension__ ({ \
59118   float64x2x2_t __s1 = __p1; \
59119   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 42); \
59120 })
59121 #else
59122 #define vst2q_f64(__p0, __p1) __extension__ ({ \
59123   float64x2x2_t __s1 = __p1; \
59124   float64x2x2_t __rev1; \
59125   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59126   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59127   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 42); \
59128 })
59129 #endif
59130 
59131 #ifdef __LITTLE_ENDIAN__
59132 #define vst2q_s64(__p0, __p1) __extension__ ({ \
59133   int64x2x2_t __s1 = __p1; \
59134   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 35); \
59135 })
59136 #else
59137 #define vst2q_s64(__p0, __p1) __extension__ ({ \
59138   int64x2x2_t __s1 = __p1; \
59139   int64x2x2_t __rev1; \
59140   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59141   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59142   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 35); \
59143 })
59144 #endif
59145 
59146 #define vst2_f64(__p0, __p1) __extension__ ({ \
59147   float64x1x2_t __s1 = __p1; \
59148   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 10); \
59149 })
59150 #define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59151   poly64x1x2_t __s1 = __p1; \
59152   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
59153 })
59154 #ifdef __LITTLE_ENDIAN__
59155 #define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
59156   poly8x16x2_t __s1 = __p1; \
59157   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
59158 })
59159 #else
59160 #define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
59161   poly8x16x2_t __s1 = __p1; \
59162   poly8x16x2_t __rev1; \
59163   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59164   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59165   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
59166 })
59167 #endif
59168 
59169 #ifdef __LITTLE_ENDIAN__
59170 #define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59171   poly64x2x2_t __s1 = __p1; \
59172   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
59173 })
59174 #else
59175 #define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59176   poly64x2x2_t __s1 = __p1; \
59177   poly64x2x2_t __rev1; \
59178   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59179   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59180   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
59181 })
59182 #endif
59183 
59184 #ifdef __LITTLE_ENDIAN__
59185 #define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
59186   uint8x16x2_t __s1 = __p1; \
59187   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
59188 })
59189 #else
59190 #define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
59191   uint8x16x2_t __s1 = __p1; \
59192   uint8x16x2_t __rev1; \
59193   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59194   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59195   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
59196 })
59197 #endif
59198 
59199 #ifdef __LITTLE_ENDIAN__
59200 #define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
59201   uint64x2x2_t __s1 = __p1; \
59202   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
59203 })
59204 #else
59205 #define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
59206   uint64x2x2_t __s1 = __p1; \
59207   uint64x2x2_t __rev1; \
59208   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59209   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59210   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
59211 })
59212 #endif
59213 
59214 #ifdef __LITTLE_ENDIAN__
59215 #define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
59216   int8x16x2_t __s1 = __p1; \
59217   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
59218 })
59219 #else
59220 #define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
59221   int8x16x2_t __s1 = __p1; \
59222   int8x16x2_t __rev1; \
59223   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59224   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59225   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
59226 })
59227 #endif
59228 
59229 #ifdef __LITTLE_ENDIAN__
59230 #define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59231   float64x2x2_t __s1 = __p1; \
59232   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 42); \
59233 })
59234 #else
59235 #define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59236   float64x2x2_t __s1 = __p1; \
59237   float64x2x2_t __rev1; \
59238   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59239   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59240   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 42); \
59241 })
59242 #endif
59243 
59244 #ifdef __LITTLE_ENDIAN__
59245 #define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
59246   int64x2x2_t __s1 = __p1; \
59247   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 35); \
59248 })
59249 #else
59250 #define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
59251   int64x2x2_t __s1 = __p1; \
59252   int64x2x2_t __rev1; \
59253   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59254   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59255   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 35); \
59256 })
59257 #endif
59258 
59259 #define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
59260   uint64x1x2_t __s1 = __p1; \
59261   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
59262 })
59263 #define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59264   float64x1x2_t __s1 = __p1; \
59265   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 10); \
59266 })
59267 #define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
59268   int64x1x2_t __s1 = __p1; \
59269   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 3); \
59270 })
59271 #define vst3_p64(__p0, __p1) __extension__ ({ \
59272   poly64x1x3_t __s1 = __p1; \
59273   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
59274 })
59275 #ifdef __LITTLE_ENDIAN__
59276 #define vst3q_p64(__p0, __p1) __extension__ ({ \
59277   poly64x2x3_t __s1 = __p1; \
59278   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
59279 })
59280 #else
59281 #define vst3q_p64(__p0, __p1) __extension__ ({ \
59282   poly64x2x3_t __s1 = __p1; \
59283   poly64x2x3_t __rev1; \
59284   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59285   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59286   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59287   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
59288 })
59289 #endif
59290 
59291 #ifdef __LITTLE_ENDIAN__
59292 #define vst3q_u64(__p0, __p1) __extension__ ({ \
59293   uint64x2x3_t __s1 = __p1; \
59294   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
59295 })
59296 #else
59297 #define vst3q_u64(__p0, __p1) __extension__ ({ \
59298   uint64x2x3_t __s1 = __p1; \
59299   uint64x2x3_t __rev1; \
59300   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59301   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59302   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59303   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
59304 })
59305 #endif
59306 
59307 #ifdef __LITTLE_ENDIAN__
59308 #define vst3q_f64(__p0, __p1) __extension__ ({ \
59309   float64x2x3_t __s1 = __p1; \
59310   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 42); \
59311 })
59312 #else
59313 #define vst3q_f64(__p0, __p1) __extension__ ({ \
59314   float64x2x3_t __s1 = __p1; \
59315   float64x2x3_t __rev1; \
59316   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59317   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59318   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59319   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 42); \
59320 })
59321 #endif
59322 
59323 #ifdef __LITTLE_ENDIAN__
59324 #define vst3q_s64(__p0, __p1) __extension__ ({ \
59325   int64x2x3_t __s1 = __p1; \
59326   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 35); \
59327 })
59328 #else
59329 #define vst3q_s64(__p0, __p1) __extension__ ({ \
59330   int64x2x3_t __s1 = __p1; \
59331   int64x2x3_t __rev1; \
59332   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59333   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59334   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59335   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 35); \
59336 })
59337 #endif
59338 
59339 #define vst3_f64(__p0, __p1) __extension__ ({ \
59340   float64x1x3_t __s1 = __p1; \
59341   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 10); \
59342 })
59343 #define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59344   poly64x1x3_t __s1 = __p1; \
59345   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
59346 })
59347 #ifdef __LITTLE_ENDIAN__
59348 #define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
59349   poly8x16x3_t __s1 = __p1; \
59350   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
59351 })
59352 #else
59353 #define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
59354   poly8x16x3_t __s1 = __p1; \
59355   poly8x16x3_t __rev1; \
59356   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59357   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59358   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59359   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
59360 })
59361 #endif
59362 
59363 #ifdef __LITTLE_ENDIAN__
59364 #define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59365   poly64x2x3_t __s1 = __p1; \
59366   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
59367 })
59368 #else
59369 #define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59370   poly64x2x3_t __s1 = __p1; \
59371   poly64x2x3_t __rev1; \
59372   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59373   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59374   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59375   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
59376 })
59377 #endif
59378 
59379 #ifdef __LITTLE_ENDIAN__
59380 #define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
59381   uint8x16x3_t __s1 = __p1; \
59382   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
59383 })
59384 #else
59385 #define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
59386   uint8x16x3_t __s1 = __p1; \
59387   uint8x16x3_t __rev1; \
59388   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59389   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59390   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59391   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
59392 })
59393 #endif
59394 
59395 #ifdef __LITTLE_ENDIAN__
59396 #define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
59397   uint64x2x3_t __s1 = __p1; \
59398   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
59399 })
59400 #else
59401 #define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
59402   uint64x2x3_t __s1 = __p1; \
59403   uint64x2x3_t __rev1; \
59404   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59405   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59406   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59407   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
59408 })
59409 #endif
59410 
59411 #ifdef __LITTLE_ENDIAN__
59412 #define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
59413   int8x16x3_t __s1 = __p1; \
59414   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
59415 })
59416 #else
59417 #define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
59418   int8x16x3_t __s1 = __p1; \
59419   int8x16x3_t __rev1; \
59420   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59421   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59422   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59423   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
59424 })
59425 #endif
59426 
59427 #ifdef __LITTLE_ENDIAN__
59428 #define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59429   float64x2x3_t __s1 = __p1; \
59430   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 42); \
59431 })
59432 #else
59433 #define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59434   float64x2x3_t __s1 = __p1; \
59435   float64x2x3_t __rev1; \
59436   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59437   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59438   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59439   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 42); \
59440 })
59441 #endif
59442 
59443 #ifdef __LITTLE_ENDIAN__
59444 #define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
59445   int64x2x3_t __s1 = __p1; \
59446   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 35); \
59447 })
59448 #else
59449 #define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
59450   int64x2x3_t __s1 = __p1; \
59451   int64x2x3_t __rev1; \
59452   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59453   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59454   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59455   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 35); \
59456 })
59457 #endif
59458 
59459 #define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
59460   uint64x1x3_t __s1 = __p1; \
59461   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
59462 })
59463 #define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59464   float64x1x3_t __s1 = __p1; \
59465   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 10); \
59466 })
59467 #define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
59468   int64x1x3_t __s1 = __p1; \
59469   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 3); \
59470 })
59471 #define vst4_p64(__p0, __p1) __extension__ ({ \
59472   poly64x1x4_t __s1 = __p1; \
59473   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
59474 })
59475 #ifdef __LITTLE_ENDIAN__
59476 #define vst4q_p64(__p0, __p1) __extension__ ({ \
59477   poly64x2x4_t __s1 = __p1; \
59478   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
59479 })
59480 #else
59481 #define vst4q_p64(__p0, __p1) __extension__ ({ \
59482   poly64x2x4_t __s1 = __p1; \
59483   poly64x2x4_t __rev1; \
59484   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59485   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59486   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59487   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59488   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
59489 })
59490 #endif
59491 
59492 #ifdef __LITTLE_ENDIAN__
59493 #define vst4q_u64(__p0, __p1) __extension__ ({ \
59494   uint64x2x4_t __s1 = __p1; \
59495   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
59496 })
59497 #else
59498 #define vst4q_u64(__p0, __p1) __extension__ ({ \
59499   uint64x2x4_t __s1 = __p1; \
59500   uint64x2x4_t __rev1; \
59501   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59502   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59503   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59504   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59505   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
59506 })
59507 #endif
59508 
59509 #ifdef __LITTLE_ENDIAN__
59510 #define vst4q_f64(__p0, __p1) __extension__ ({ \
59511   float64x2x4_t __s1 = __p1; \
59512   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 42); \
59513 })
59514 #else
59515 #define vst4q_f64(__p0, __p1) __extension__ ({ \
59516   float64x2x4_t __s1 = __p1; \
59517   float64x2x4_t __rev1; \
59518   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59519   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59520   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59521   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59522   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 42); \
59523 })
59524 #endif
59525 
59526 #ifdef __LITTLE_ENDIAN__
59527 #define vst4q_s64(__p0, __p1) __extension__ ({ \
59528   int64x2x4_t __s1 = __p1; \
59529   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 35); \
59530 })
59531 #else
59532 #define vst4q_s64(__p0, __p1) __extension__ ({ \
59533   int64x2x4_t __s1 = __p1; \
59534   int64x2x4_t __rev1; \
59535   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59536   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59537   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59538   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59539   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 35); \
59540 })
59541 #endif
59542 
59543 #define vst4_f64(__p0, __p1) __extension__ ({ \
59544   float64x1x4_t __s1 = __p1; \
59545   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 10); \
59546 })
59547 #define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59548   poly64x1x4_t __s1 = __p1; \
59549   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
59550 })
59551 #ifdef __LITTLE_ENDIAN__
59552 #define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
59553   poly8x16x4_t __s1 = __p1; \
59554   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
59555 })
59556 #else
59557 #define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
59558   poly8x16x4_t __s1 = __p1; \
59559   poly8x16x4_t __rev1; \
59560   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59561   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59562   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59563   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59564   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
59565 })
59566 #endif
59567 
59568 #ifdef __LITTLE_ENDIAN__
59569 #define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59570   poly64x2x4_t __s1 = __p1; \
59571   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
59572 })
59573 #else
59574 #define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59575   poly64x2x4_t __s1 = __p1; \
59576   poly64x2x4_t __rev1; \
59577   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59578   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59579   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59580   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59581   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
59582 })
59583 #endif
59584 
59585 #ifdef __LITTLE_ENDIAN__
59586 #define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
59587   uint8x16x4_t __s1 = __p1; \
59588   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
59589 })
59590 #else
59591 #define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
59592   uint8x16x4_t __s1 = __p1; \
59593   uint8x16x4_t __rev1; \
59594   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59595   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59596   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59597   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59598   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
59599 })
59600 #endif
59601 
59602 #ifdef __LITTLE_ENDIAN__
59603 #define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
59604   uint64x2x4_t __s1 = __p1; \
59605   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
59606 })
59607 #else
59608 #define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
59609   uint64x2x4_t __s1 = __p1; \
59610   uint64x2x4_t __rev1; \
59611   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59612   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59613   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59614   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59615   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
59616 })
59617 #endif
59618 
59619 #ifdef __LITTLE_ENDIAN__
59620 #define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
59621   int8x16x4_t __s1 = __p1; \
59622   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
59623 })
59624 #else
59625 #define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
59626   int8x16x4_t __s1 = __p1; \
59627   int8x16x4_t __rev1; \
59628   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59629   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59630   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59631   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59632   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
59633 })
59634 #endif
59635 
59636 #ifdef __LITTLE_ENDIAN__
59637 #define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59638   float64x2x4_t __s1 = __p1; \
59639   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 42); \
59640 })
59641 #else
59642 #define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59643   float64x2x4_t __s1 = __p1; \
59644   float64x2x4_t __rev1; \
59645   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59646   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59647   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59648   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59649   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 42); \
59650 })
59651 #endif
59652 
59653 #ifdef __LITTLE_ENDIAN__
59654 #define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
59655   int64x2x4_t __s1 = __p1; \
59656   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 35); \
59657 })
59658 #else
59659 #define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
59660   int64x2x4_t __s1 = __p1; \
59661   int64x2x4_t __rev1; \
59662   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
59663   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
59664   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
59665   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
59666   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 35); \
59667 })
59668 #endif
59669 
59670 #define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
59671   uint64x1x4_t __s1 = __p1; \
59672   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
59673 })
59674 #define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59675   float64x1x4_t __s1 = __p1; \
59676   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 10); \
59677 })
59678 #define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
59679   int64x1x4_t __s1 = __p1; \
59680   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 3); \
59681 })
59682 #define vstrq_p128(__p0, __p1) __extension__ ({ \
59683   poly128_t __s1 = __p1; \
59684   __builtin_neon_vstrq_p128(__p0, __s1); \
59685 })
vsubd_u64(uint64_t __p0,uint64_t __p1)59686 __ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
59687   uint64_t __ret;
59688   __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
59689   return __ret;
59690 }
vsubd_s64(int64_t __p0,int64_t __p1)59691 __ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
59692   int64_t __ret;
59693   __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
59694   return __ret;
59695 }
59696 #ifdef __LITTLE_ENDIAN__
vsubq_f64(float64x2_t __p0,float64x2_t __p1)59697 __ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
59698   float64x2_t __ret;
59699   __ret = __p0 - __p1;
59700   return __ret;
59701 }
59702 #else
vsubq_f64(float64x2_t __p0,float64x2_t __p1)59703 __ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
59704   float64x2_t __ret;
59705   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59706   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59707   __ret = __rev0 - __rev1;
59708   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59709   return __ret;
59710 }
59711 #endif
59712 
vsub_f64(float64x1_t __p0,float64x1_t __p1)59713 __ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
59714   float64x1_t __ret;
59715   __ret = __p0 - __p1;
59716   return __ret;
59717 }
59718 #ifdef __LITTLE_ENDIAN__
vsubhn_high_u32(uint16x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)59719 __ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59720   uint16x8_t __ret;
59721   __ret = vcombine_u16(__p0, vsubhn_u32(__p1, __p2));
59722   return __ret;
59723 }
59724 #else
vsubhn_high_u32(uint16x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)59725 __ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59726   uint16x8_t __ret;
59727   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59728   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59729   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59730   __ret = __noswap_vcombine_u16(__rev0, __noswap_vsubhn_u32(__rev1, __rev2));
59731   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59732   return __ret;
59733 }
59734 #endif
59735 
59736 #ifdef __LITTLE_ENDIAN__
vsubhn_high_u64(uint32x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)59737 __ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59738   uint32x4_t __ret;
59739   __ret = vcombine_u32(__p0, vsubhn_u64(__p1, __p2));
59740   return __ret;
59741 }
59742 #else
vsubhn_high_u64(uint32x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)59743 __ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59744   uint32x4_t __ret;
59745   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59746   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59747   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59748   __ret = __noswap_vcombine_u32(__rev0, __noswap_vsubhn_u64(__rev1, __rev2));
59749   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59750   return __ret;
59751 }
59752 #endif
59753 
59754 #ifdef __LITTLE_ENDIAN__
vsubhn_high_u16(uint8x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)59755 __ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59756   uint8x16_t __ret;
59757   __ret = vcombine_u8(__p0, vsubhn_u16(__p1, __p2));
59758   return __ret;
59759 }
59760 #else
vsubhn_high_u16(uint8x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)59761 __ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59762   uint8x16_t __ret;
59763   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59764   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59765   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59766   __ret = __noswap_vcombine_u8(__rev0, __noswap_vsubhn_u16(__rev1, __rev2));
59767   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59768   return __ret;
59769 }
59770 #endif
59771 
59772 #ifdef __LITTLE_ENDIAN__
vsubhn_high_s32(int16x4_t __p0,int32x4_t __p1,int32x4_t __p2)59773 __ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59774   int16x8_t __ret;
59775   __ret = vcombine_s16(__p0, vsubhn_s32(__p1, __p2));
59776   return __ret;
59777 }
59778 #else
vsubhn_high_s32(int16x4_t __p0,int32x4_t __p1,int32x4_t __p2)59779 __ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59780   int16x8_t __ret;
59781   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59782   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59783   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59784   __ret = __noswap_vcombine_s16(__rev0, __noswap_vsubhn_s32(__rev1, __rev2));
59785   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59786   return __ret;
59787 }
59788 #endif
59789 
59790 #ifdef __LITTLE_ENDIAN__
vsubhn_high_s64(int32x2_t __p0,int64x2_t __p1,int64x2_t __p2)59791 __ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59792   int32x4_t __ret;
59793   __ret = vcombine_s32(__p0, vsubhn_s64(__p1, __p2));
59794   return __ret;
59795 }
59796 #else
vsubhn_high_s64(int32x2_t __p0,int64x2_t __p1,int64x2_t __p2)59797 __ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59798   int32x4_t __ret;
59799   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59800   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59801   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59802   __ret = __noswap_vcombine_s32(__rev0, __noswap_vsubhn_s64(__rev1, __rev2));
59803   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59804   return __ret;
59805 }
59806 #endif
59807 
59808 #ifdef __LITTLE_ENDIAN__
vsubhn_high_s16(int8x8_t __p0,int16x8_t __p1,int16x8_t __p2)59809 __ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59810   int8x16_t __ret;
59811   __ret = vcombine_s8(__p0, vsubhn_s16(__p1, __p2));
59812   return __ret;
59813 }
59814 #else
vsubhn_high_s16(int8x8_t __p0,int16x8_t __p1,int16x8_t __p2)59815 __ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59816   int8x16_t __ret;
59817   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59818   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59819   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59820   __ret = __noswap_vcombine_s8(__rev0, __noswap_vsubhn_s16(__rev1, __rev2));
59821   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59822   return __ret;
59823 }
59824 #endif
59825 
59826 #ifdef __LITTLE_ENDIAN__
vsubl_high_u8(uint8x16_t __p0,uint8x16_t __p1)59827 __ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
59828   uint16x8_t __ret;
59829   __ret = vmovl_high_u8(__p0) - vmovl_high_u8(__p1);
59830   return __ret;
59831 }
59832 #else
vsubl_high_u8(uint8x16_t __p0,uint8x16_t __p1)59833 __ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
59834   uint16x8_t __ret;
59835   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59836   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59837   __ret = __noswap_vmovl_high_u8(__rev0) - __noswap_vmovl_high_u8(__rev1);
59838   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59839   return __ret;
59840 }
59841 #endif
59842 
59843 #ifdef __LITTLE_ENDIAN__
vsubl_high_u32(uint32x4_t __p0,uint32x4_t __p1)59844 __ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
59845   uint64x2_t __ret;
59846   __ret = vmovl_high_u32(__p0) - vmovl_high_u32(__p1);
59847   return __ret;
59848 }
59849 #else
vsubl_high_u32(uint32x4_t __p0,uint32x4_t __p1)59850 __ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
59851   uint64x2_t __ret;
59852   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59853   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59854   __ret = __noswap_vmovl_high_u32(__rev0) - __noswap_vmovl_high_u32(__rev1);
59855   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59856   return __ret;
59857 }
59858 #endif
59859 
59860 #ifdef __LITTLE_ENDIAN__
vsubl_high_u16(uint16x8_t __p0,uint16x8_t __p1)59861 __ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
59862   uint32x4_t __ret;
59863   __ret = vmovl_high_u16(__p0) - vmovl_high_u16(__p1);
59864   return __ret;
59865 }
59866 #else
vsubl_high_u16(uint16x8_t __p0,uint16x8_t __p1)59867 __ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
59868   uint32x4_t __ret;
59869   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59870   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59871   __ret = __noswap_vmovl_high_u16(__rev0) - __noswap_vmovl_high_u16(__rev1);
59872   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59873   return __ret;
59874 }
59875 #endif
59876 
59877 #ifdef __LITTLE_ENDIAN__
vsubl_high_s8(int8x16_t __p0,int8x16_t __p1)59878 __ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
59879   int16x8_t __ret;
59880   __ret = vmovl_high_s8(__p0) - vmovl_high_s8(__p1);
59881   return __ret;
59882 }
59883 #else
vsubl_high_s8(int8x16_t __p0,int8x16_t __p1)59884 __ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
59885   int16x8_t __ret;
59886   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59887   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59888   __ret = __noswap_vmovl_high_s8(__rev0) - __noswap_vmovl_high_s8(__rev1);
59889   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59890   return __ret;
59891 }
59892 #endif
59893 
59894 #ifdef __LITTLE_ENDIAN__
vsubl_high_s32(int32x4_t __p0,int32x4_t __p1)59895 __ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
59896   int64x2_t __ret;
59897   __ret = vmovl_high_s32(__p0) - vmovl_high_s32(__p1);
59898   return __ret;
59899 }
59900 #else
vsubl_high_s32(int32x4_t __p0,int32x4_t __p1)59901 __ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
59902   int64x2_t __ret;
59903   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59904   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59905   __ret = __noswap_vmovl_high_s32(__rev0) - __noswap_vmovl_high_s32(__rev1);
59906   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59907   return __ret;
59908 }
59909 #endif
59910 
59911 #ifdef __LITTLE_ENDIAN__
vsubl_high_s16(int16x8_t __p0,int16x8_t __p1)59912 __ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
59913   int32x4_t __ret;
59914   __ret = vmovl_high_s16(__p0) - vmovl_high_s16(__p1);
59915   return __ret;
59916 }
59917 #else
vsubl_high_s16(int16x8_t __p0,int16x8_t __p1)59918 __ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
59919   int32x4_t __ret;
59920   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59921   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59922   __ret = __noswap_vmovl_high_s16(__rev0) - __noswap_vmovl_high_s16(__rev1);
59923   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59924   return __ret;
59925 }
59926 #endif
59927 
59928 #ifdef __LITTLE_ENDIAN__
vsubw_high_u8(uint16x8_t __p0,uint8x16_t __p1)59929 __ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
59930   uint16x8_t __ret;
59931   __ret = __p0 - vmovl_high_u8(__p1);
59932   return __ret;
59933 }
59934 #else
vsubw_high_u8(uint16x8_t __p0,uint8x16_t __p1)59935 __ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
59936   uint16x8_t __ret;
59937   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59938   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59939   __ret = __rev0 - __noswap_vmovl_high_u8(__rev1);
59940   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59941   return __ret;
59942 }
59943 #endif
59944 
59945 #ifdef __LITTLE_ENDIAN__
vsubw_high_u32(uint64x2_t __p0,uint32x4_t __p1)59946 __ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
59947   uint64x2_t __ret;
59948   __ret = __p0 - vmovl_high_u32(__p1);
59949   return __ret;
59950 }
59951 #else
vsubw_high_u32(uint64x2_t __p0,uint32x4_t __p1)59952 __ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
59953   uint64x2_t __ret;
59954   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59955   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59956   __ret = __rev0 - __noswap_vmovl_high_u32(__rev1);
59957   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59958   return __ret;
59959 }
59960 #endif
59961 
59962 #ifdef __LITTLE_ENDIAN__
vsubw_high_u16(uint32x4_t __p0,uint16x8_t __p1)59963 __ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
59964   uint32x4_t __ret;
59965   __ret = __p0 - vmovl_high_u16(__p1);
59966   return __ret;
59967 }
59968 #else
vsubw_high_u16(uint32x4_t __p0,uint16x8_t __p1)59969 __ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
59970   uint32x4_t __ret;
59971   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59972   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59973   __ret = __rev0 - __noswap_vmovl_high_u16(__rev1);
59974   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59975   return __ret;
59976 }
59977 #endif
59978 
59979 #ifdef __LITTLE_ENDIAN__
vsubw_high_s8(int16x8_t __p0,int8x16_t __p1)59980 __ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
59981   int16x8_t __ret;
59982   __ret = __p0 - vmovl_high_s8(__p1);
59983   return __ret;
59984 }
59985 #else
vsubw_high_s8(int16x8_t __p0,int8x16_t __p1)59986 __ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
59987   int16x8_t __ret;
59988   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59989   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59990   __ret = __rev0 - __noswap_vmovl_high_s8(__rev1);
59991   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59992   return __ret;
59993 }
59994 #endif
59995 
59996 #ifdef __LITTLE_ENDIAN__
vsubw_high_s32(int64x2_t __p0,int32x4_t __p1)59997 __ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
59998   int64x2_t __ret;
59999   __ret = __p0 - vmovl_high_s32(__p1);
60000   return __ret;
60001 }
60002 #else
vsubw_high_s32(int64x2_t __p0,int32x4_t __p1)60003 __ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
60004   int64x2_t __ret;
60005   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60006   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60007   __ret = __rev0 - __noswap_vmovl_high_s32(__rev1);
60008   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60009   return __ret;
60010 }
60011 #endif
60012 
60013 #ifdef __LITTLE_ENDIAN__
vsubw_high_s16(int32x4_t __p0,int16x8_t __p1)60014 __ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
60015   int32x4_t __ret;
60016   __ret = __p0 - vmovl_high_s16(__p1);
60017   return __ret;
60018 }
60019 #else
vsubw_high_s16(int32x4_t __p0,int16x8_t __p1)60020 __ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
60021   int32x4_t __ret;
60022   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60023   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60024   __ret = __rev0 - __noswap_vmovl_high_s16(__rev1);
60025   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60026   return __ret;
60027 }
60028 #endif
60029 
60030 #ifdef __LITTLE_ENDIAN__
vtrn1_p8(poly8x8_t __p0,poly8x8_t __p1)60031 __ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
60032   poly8x8_t __ret;
60033   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
60034   return __ret;
60035 }
60036 #else
vtrn1_p8(poly8x8_t __p0,poly8x8_t __p1)60037 __ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
60038   poly8x8_t __ret;
60039   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60040   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60041   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
60042   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60043   return __ret;
60044 }
60045 #endif
60046 
60047 #ifdef __LITTLE_ENDIAN__
vtrn1_p16(poly16x4_t __p0,poly16x4_t __p1)60048 __ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
60049   poly16x4_t __ret;
60050   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
60051   return __ret;
60052 }
60053 #else
vtrn1_p16(poly16x4_t __p0,poly16x4_t __p1)60054 __ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
60055   poly16x4_t __ret;
60056   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60057   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60058   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
60059   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60060   return __ret;
60061 }
60062 #endif
60063 
60064 #ifdef __LITTLE_ENDIAN__
vtrn1q_p8(poly8x16_t __p0,poly8x16_t __p1)60065 __ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
60066   poly8x16_t __ret;
60067   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
60068   return __ret;
60069 }
60070 #else
vtrn1q_p8(poly8x16_t __p0,poly8x16_t __p1)60071 __ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
60072   poly8x16_t __ret;
60073   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60074   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60075   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
60076   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60077   return __ret;
60078 }
60079 #endif
60080 
60081 #ifdef __LITTLE_ENDIAN__
vtrn1q_p64(poly64x2_t __p0,poly64x2_t __p1)60082 __ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
60083   poly64x2_t __ret;
60084   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
60085   return __ret;
60086 }
60087 #else
vtrn1q_p64(poly64x2_t __p0,poly64x2_t __p1)60088 __ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
60089   poly64x2_t __ret;
60090   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60091   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60092   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
60093   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60094   return __ret;
60095 }
60096 #endif
60097 
60098 #ifdef __LITTLE_ENDIAN__
vtrn1q_p16(poly16x8_t __p0,poly16x8_t __p1)60099 __ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
60100   poly16x8_t __ret;
60101   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
60102   return __ret;
60103 }
60104 #else
vtrn1q_p16(poly16x8_t __p0,poly16x8_t __p1)60105 __ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
60106   poly16x8_t __ret;
60107   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60108   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60109   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
60110   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60111   return __ret;
60112 }
60113 #endif
60114 
60115 #ifdef __LITTLE_ENDIAN__
vtrn1q_u8(uint8x16_t __p0,uint8x16_t __p1)60116 __ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
60117   uint8x16_t __ret;
60118   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
60119   return __ret;
60120 }
60121 #else
vtrn1q_u8(uint8x16_t __p0,uint8x16_t __p1)60122 __ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
60123   uint8x16_t __ret;
60124   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60125   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60126   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
60127   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60128   return __ret;
60129 }
60130 #endif
60131 
60132 #ifdef __LITTLE_ENDIAN__
vtrn1q_u32(uint32x4_t __p0,uint32x4_t __p1)60133 __ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
60134   uint32x4_t __ret;
60135   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
60136   return __ret;
60137 }
60138 #else
vtrn1q_u32(uint32x4_t __p0,uint32x4_t __p1)60139 __ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
60140   uint32x4_t __ret;
60141   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60142   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60143   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
60144   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60145   return __ret;
60146 }
60147 #endif
60148 
60149 #ifdef __LITTLE_ENDIAN__
vtrn1q_u64(uint64x2_t __p0,uint64x2_t __p1)60150 __ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
60151   uint64x2_t __ret;
60152   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
60153   return __ret;
60154 }
60155 #else
vtrn1q_u64(uint64x2_t __p0,uint64x2_t __p1)60156 __ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
60157   uint64x2_t __ret;
60158   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60159   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60160   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
60161   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60162   return __ret;
60163 }
60164 #endif
60165 
60166 #ifdef __LITTLE_ENDIAN__
vtrn1q_u16(uint16x8_t __p0,uint16x8_t __p1)60167 __ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
60168   uint16x8_t __ret;
60169   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
60170   return __ret;
60171 }
60172 #else
vtrn1q_u16(uint16x8_t __p0,uint16x8_t __p1)60173 __ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
60174   uint16x8_t __ret;
60175   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60176   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60177   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
60178   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60179   return __ret;
60180 }
60181 #endif
60182 
60183 #ifdef __LITTLE_ENDIAN__
vtrn1q_s8(int8x16_t __p0,int8x16_t __p1)60184 __ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
60185   int8x16_t __ret;
60186   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
60187   return __ret;
60188 }
60189 #else
vtrn1q_s8(int8x16_t __p0,int8x16_t __p1)60190 __ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
60191   int8x16_t __ret;
60192   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60193   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60194   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
60195   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60196   return __ret;
60197 }
60198 #endif
60199 
60200 #ifdef __LITTLE_ENDIAN__
vtrn1q_f64(float64x2_t __p0,float64x2_t __p1)60201 __ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
60202   float64x2_t __ret;
60203   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
60204   return __ret;
60205 }
60206 #else
vtrn1q_f64(float64x2_t __p0,float64x2_t __p1)60207 __ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
60208   float64x2_t __ret;
60209   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60210   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60211   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
60212   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60213   return __ret;
60214 }
60215 #endif
60216 
60217 #ifdef __LITTLE_ENDIAN__
vtrn1q_f32(float32x4_t __p0,float32x4_t __p1)60218 __ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
60219   float32x4_t __ret;
60220   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
60221   return __ret;
60222 }
60223 #else
vtrn1q_f32(float32x4_t __p0,float32x4_t __p1)60224 __ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
60225   float32x4_t __ret;
60226   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60227   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60228   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
60229   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60230   return __ret;
60231 }
60232 #endif
60233 
60234 #ifdef __LITTLE_ENDIAN__
vtrn1q_s32(int32x4_t __p0,int32x4_t __p1)60235 __ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
60236   int32x4_t __ret;
60237   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
60238   return __ret;
60239 }
60240 #else
vtrn1q_s32(int32x4_t __p0,int32x4_t __p1)60241 __ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
60242   int32x4_t __ret;
60243   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60244   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60245   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
60246   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60247   return __ret;
60248 }
60249 #endif
60250 
60251 #ifdef __LITTLE_ENDIAN__
vtrn1q_s64(int64x2_t __p0,int64x2_t __p1)60252 __ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
60253   int64x2_t __ret;
60254   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
60255   return __ret;
60256 }
60257 #else
vtrn1q_s64(int64x2_t __p0,int64x2_t __p1)60258 __ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
60259   int64x2_t __ret;
60260   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60261   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60262   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
60263   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60264   return __ret;
60265 }
60266 #endif
60267 
60268 #ifdef __LITTLE_ENDIAN__
vtrn1q_s16(int16x8_t __p0,int16x8_t __p1)60269 __ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
60270   int16x8_t __ret;
60271   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
60272   return __ret;
60273 }
60274 #else
vtrn1q_s16(int16x8_t __p0,int16x8_t __p1)60275 __ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
60276   int16x8_t __ret;
60277   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60278   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60279   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
60280   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60281   return __ret;
60282 }
60283 #endif
60284 
60285 #ifdef __LITTLE_ENDIAN__
vtrn1_u8(uint8x8_t __p0,uint8x8_t __p1)60286 __ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
60287   uint8x8_t __ret;
60288   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
60289   return __ret;
60290 }
60291 #else
vtrn1_u8(uint8x8_t __p0,uint8x8_t __p1)60292 __ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
60293   uint8x8_t __ret;
60294   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60295   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60296   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
60297   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60298   return __ret;
60299 }
60300 #endif
60301 
60302 #ifdef __LITTLE_ENDIAN__
vtrn1_u32(uint32x2_t __p0,uint32x2_t __p1)60303 __ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
60304   uint32x2_t __ret;
60305   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
60306   return __ret;
60307 }
60308 #else
vtrn1_u32(uint32x2_t __p0,uint32x2_t __p1)60309 __ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
60310   uint32x2_t __ret;
60311   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60312   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60313   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
60314   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60315   return __ret;
60316 }
60317 #endif
60318 
60319 #ifdef __LITTLE_ENDIAN__
vtrn1_u16(uint16x4_t __p0,uint16x4_t __p1)60320 __ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
60321   uint16x4_t __ret;
60322   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
60323   return __ret;
60324 }
60325 #else
vtrn1_u16(uint16x4_t __p0,uint16x4_t __p1)60326 __ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
60327   uint16x4_t __ret;
60328   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60329   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60330   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
60331   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60332   return __ret;
60333 }
60334 #endif
60335 
60336 #ifdef __LITTLE_ENDIAN__
vtrn1_s8(int8x8_t __p0,int8x8_t __p1)60337 __ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
60338   int8x8_t __ret;
60339   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
60340   return __ret;
60341 }
60342 #else
vtrn1_s8(int8x8_t __p0,int8x8_t __p1)60343 __ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
60344   int8x8_t __ret;
60345   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60346   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60347   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
60348   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60349   return __ret;
60350 }
60351 #endif
60352 
60353 #ifdef __LITTLE_ENDIAN__
vtrn1_f32(float32x2_t __p0,float32x2_t __p1)60354 __ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
60355   float32x2_t __ret;
60356   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
60357   return __ret;
60358 }
60359 #else
vtrn1_f32(float32x2_t __p0,float32x2_t __p1)60360 __ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
60361   float32x2_t __ret;
60362   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60363   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60364   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
60365   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60366   return __ret;
60367 }
60368 #endif
60369 
60370 #ifdef __LITTLE_ENDIAN__
vtrn1_s32(int32x2_t __p0,int32x2_t __p1)60371 __ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
60372   int32x2_t __ret;
60373   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
60374   return __ret;
60375 }
60376 #else
vtrn1_s32(int32x2_t __p0,int32x2_t __p1)60377 __ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
60378   int32x2_t __ret;
60379   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60380   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60381   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
60382   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60383   return __ret;
60384 }
60385 #endif
60386 
60387 #ifdef __LITTLE_ENDIAN__
vtrn1_s16(int16x4_t __p0,int16x4_t __p1)60388 __ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
60389   int16x4_t __ret;
60390   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
60391   return __ret;
60392 }
60393 #else
vtrn1_s16(int16x4_t __p0,int16x4_t __p1)60394 __ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
60395   int16x4_t __ret;
60396   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60397   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60398   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
60399   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60400   return __ret;
60401 }
60402 #endif
60403 
60404 #ifdef __LITTLE_ENDIAN__
vtrn2_p8(poly8x8_t __p0,poly8x8_t __p1)60405 __ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
60406   poly8x8_t __ret;
60407   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
60408   return __ret;
60409 }
60410 #else
vtrn2_p8(poly8x8_t __p0,poly8x8_t __p1)60411 __ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
60412   poly8x8_t __ret;
60413   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60414   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60415   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
60416   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60417   return __ret;
60418 }
60419 #endif
60420 
60421 #ifdef __LITTLE_ENDIAN__
vtrn2_p16(poly16x4_t __p0,poly16x4_t __p1)60422 __ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
60423   poly16x4_t __ret;
60424   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
60425   return __ret;
60426 }
60427 #else
vtrn2_p16(poly16x4_t __p0,poly16x4_t __p1)60428 __ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
60429   poly16x4_t __ret;
60430   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60431   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60432   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
60433   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60434   return __ret;
60435 }
60436 #endif
60437 
60438 #ifdef __LITTLE_ENDIAN__
vtrn2q_p8(poly8x16_t __p0,poly8x16_t __p1)60439 __ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
60440   poly8x16_t __ret;
60441   __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
60442   return __ret;
60443 }
60444 #else
vtrn2q_p8(poly8x16_t __p0,poly8x16_t __p1)60445 __ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
60446   poly8x16_t __ret;
60447   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60448   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60449   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
60450   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60451   return __ret;
60452 }
60453 #endif
60454 
60455 #ifdef __LITTLE_ENDIAN__
vtrn2q_p64(poly64x2_t __p0,poly64x2_t __p1)60456 __ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
60457   poly64x2_t __ret;
60458   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
60459   return __ret;
60460 }
60461 #else
vtrn2q_p64(poly64x2_t __p0,poly64x2_t __p1)60462 __ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
60463   poly64x2_t __ret;
60464   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60465   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60466   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
60467   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60468   return __ret;
60469 }
60470 #endif
60471 
60472 #ifdef __LITTLE_ENDIAN__
vtrn2q_p16(poly16x8_t __p0,poly16x8_t __p1)60473 __ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
60474   poly16x8_t __ret;
60475   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
60476   return __ret;
60477 }
60478 #else
vtrn2q_p16(poly16x8_t __p0,poly16x8_t __p1)60479 __ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
60480   poly16x8_t __ret;
60481   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60482   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60483   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
60484   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60485   return __ret;
60486 }
60487 #endif
60488 
60489 #ifdef __LITTLE_ENDIAN__
vtrn2q_u8(uint8x16_t __p0,uint8x16_t __p1)60490 __ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
60491   uint8x16_t __ret;
60492   __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
60493   return __ret;
60494 }
60495 #else
vtrn2q_u8(uint8x16_t __p0,uint8x16_t __p1)60496 __ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
60497   uint8x16_t __ret;
60498   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60499   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60500   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
60501   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60502   return __ret;
60503 }
60504 #endif
60505 
60506 #ifdef __LITTLE_ENDIAN__
vtrn2q_u32(uint32x4_t __p0,uint32x4_t __p1)60507 __ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
60508   uint32x4_t __ret;
60509   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
60510   return __ret;
60511 }
60512 #else
vtrn2q_u32(uint32x4_t __p0,uint32x4_t __p1)60513 __ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
60514   uint32x4_t __ret;
60515   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60516   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60517   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
60518   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60519   return __ret;
60520 }
60521 #endif
60522 
60523 #ifdef __LITTLE_ENDIAN__
vtrn2q_u64(uint64x2_t __p0,uint64x2_t __p1)60524 __ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
60525   uint64x2_t __ret;
60526   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
60527   return __ret;
60528 }
60529 #else
vtrn2q_u64(uint64x2_t __p0,uint64x2_t __p1)60530 __ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
60531   uint64x2_t __ret;
60532   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60533   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60534   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
60535   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60536   return __ret;
60537 }
60538 #endif
60539 
60540 #ifdef __LITTLE_ENDIAN__
vtrn2q_u16(uint16x8_t __p0,uint16x8_t __p1)60541 __ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
60542   uint16x8_t __ret;
60543   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
60544   return __ret;
60545 }
60546 #else
vtrn2q_u16(uint16x8_t __p0,uint16x8_t __p1)60547 __ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
60548   uint16x8_t __ret;
60549   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60550   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60551   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
60552   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60553   return __ret;
60554 }
60555 #endif
60556 
60557 #ifdef __LITTLE_ENDIAN__
vtrn2q_s8(int8x16_t __p0,int8x16_t __p1)60558 __ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
60559   int8x16_t __ret;
60560   __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
60561   return __ret;
60562 }
60563 #else
vtrn2q_s8(int8x16_t __p0,int8x16_t __p1)60564 __ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
60565   int8x16_t __ret;
60566   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60567   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60568   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
60569   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60570   return __ret;
60571 }
60572 #endif
60573 
60574 #ifdef __LITTLE_ENDIAN__
vtrn2q_f64(float64x2_t __p0,float64x2_t __p1)60575 __ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
60576   float64x2_t __ret;
60577   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
60578   return __ret;
60579 }
60580 #else
vtrn2q_f64(float64x2_t __p0,float64x2_t __p1)60581 __ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
60582   float64x2_t __ret;
60583   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60584   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60585   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
60586   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60587   return __ret;
60588 }
60589 #endif
60590 
60591 #ifdef __LITTLE_ENDIAN__
vtrn2q_f32(float32x4_t __p0,float32x4_t __p1)60592 __ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
60593   float32x4_t __ret;
60594   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
60595   return __ret;
60596 }
60597 #else
vtrn2q_f32(float32x4_t __p0,float32x4_t __p1)60598 __ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
60599   float32x4_t __ret;
60600   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60601   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60602   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
60603   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60604   return __ret;
60605 }
60606 #endif
60607 
60608 #ifdef __LITTLE_ENDIAN__
vtrn2q_s32(int32x4_t __p0,int32x4_t __p1)60609 __ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
60610   int32x4_t __ret;
60611   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
60612   return __ret;
60613 }
60614 #else
vtrn2q_s32(int32x4_t __p0,int32x4_t __p1)60615 __ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
60616   int32x4_t __ret;
60617   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60618   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60619   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
60620   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60621   return __ret;
60622 }
60623 #endif
60624 
60625 #ifdef __LITTLE_ENDIAN__
vtrn2q_s64(int64x2_t __p0,int64x2_t __p1)60626 __ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
60627   int64x2_t __ret;
60628   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
60629   return __ret;
60630 }
60631 #else
vtrn2q_s64(int64x2_t __p0,int64x2_t __p1)60632 __ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
60633   int64x2_t __ret;
60634   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60635   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60636   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
60637   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60638   return __ret;
60639 }
60640 #endif
60641 
60642 #ifdef __LITTLE_ENDIAN__
vtrn2q_s16(int16x8_t __p0,int16x8_t __p1)60643 __ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
60644   int16x8_t __ret;
60645   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
60646   return __ret;
60647 }
60648 #else
vtrn2q_s16(int16x8_t __p0,int16x8_t __p1)60649 __ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
60650   int16x8_t __ret;
60651   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60652   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60653   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
60654   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60655   return __ret;
60656 }
60657 #endif
60658 
60659 #ifdef __LITTLE_ENDIAN__
vtrn2_u8(uint8x8_t __p0,uint8x8_t __p1)60660 __ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
60661   uint8x8_t __ret;
60662   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
60663   return __ret;
60664 }
60665 #else
vtrn2_u8(uint8x8_t __p0,uint8x8_t __p1)60666 __ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
60667   uint8x8_t __ret;
60668   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60669   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60670   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
60671   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60672   return __ret;
60673 }
60674 #endif
60675 
60676 #ifdef __LITTLE_ENDIAN__
vtrn2_u32(uint32x2_t __p0,uint32x2_t __p1)60677 __ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
60678   uint32x2_t __ret;
60679   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
60680   return __ret;
60681 }
60682 #else
vtrn2_u32(uint32x2_t __p0,uint32x2_t __p1)60683 __ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
60684   uint32x2_t __ret;
60685   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60686   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60687   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
60688   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60689   return __ret;
60690 }
60691 #endif
60692 
60693 #ifdef __LITTLE_ENDIAN__
vtrn2_u16(uint16x4_t __p0,uint16x4_t __p1)60694 __ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
60695   uint16x4_t __ret;
60696   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
60697   return __ret;
60698 }
60699 #else
vtrn2_u16(uint16x4_t __p0,uint16x4_t __p1)60700 __ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
60701   uint16x4_t __ret;
60702   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60703   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60704   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
60705   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60706   return __ret;
60707 }
60708 #endif
60709 
60710 #ifdef __LITTLE_ENDIAN__
vtrn2_s8(int8x8_t __p0,int8x8_t __p1)60711 __ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
60712   int8x8_t __ret;
60713   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
60714   return __ret;
60715 }
60716 #else
vtrn2_s8(int8x8_t __p0,int8x8_t __p1)60717 __ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
60718   int8x8_t __ret;
60719   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60720   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60721   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
60722   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60723   return __ret;
60724 }
60725 #endif
60726 
60727 #ifdef __LITTLE_ENDIAN__
vtrn2_f32(float32x2_t __p0,float32x2_t __p1)60728 __ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
60729   float32x2_t __ret;
60730   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
60731   return __ret;
60732 }
60733 #else
vtrn2_f32(float32x2_t __p0,float32x2_t __p1)60734 __ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
60735   float32x2_t __ret;
60736   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60737   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60738   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
60739   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60740   return __ret;
60741 }
60742 #endif
60743 
60744 #ifdef __LITTLE_ENDIAN__
vtrn2_s32(int32x2_t __p0,int32x2_t __p1)60745 __ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
60746   int32x2_t __ret;
60747   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
60748   return __ret;
60749 }
60750 #else
vtrn2_s32(int32x2_t __p0,int32x2_t __p1)60751 __ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
60752   int32x2_t __ret;
60753   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60754   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60755   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
60756   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60757   return __ret;
60758 }
60759 #endif
60760 
60761 #ifdef __LITTLE_ENDIAN__
vtrn2_s16(int16x4_t __p0,int16x4_t __p1)60762 __ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
60763   int16x4_t __ret;
60764   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
60765   return __ret;
60766 }
60767 #else
vtrn2_s16(int16x4_t __p0,int16x4_t __p1)60768 __ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
60769   int16x4_t __ret;
60770   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60771   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60772   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
60773   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60774   return __ret;
60775 }
60776 #endif
60777 
vtst_p64(poly64x1_t __p0,poly64x1_t __p1)60778 __ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
60779   uint64x1_t __ret;
60780   __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
60781   return __ret;
60782 }
60783 #ifdef __LITTLE_ENDIAN__
vtstq_p64(poly64x2_t __p0,poly64x2_t __p1)60784 __ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
60785   uint64x2_t __ret;
60786   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
60787   return __ret;
60788 }
60789 #else
vtstq_p64(poly64x2_t __p0,poly64x2_t __p1)60790 __ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
60791   uint64x2_t __ret;
60792   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60793   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60794   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
60795   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60796   return __ret;
60797 }
60798 #endif
60799 
60800 #ifdef __LITTLE_ENDIAN__
vtstq_u64(uint64x2_t __p0,uint64x2_t __p1)60801 __ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
60802   uint64x2_t __ret;
60803   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
60804   return __ret;
60805 }
60806 #else
vtstq_u64(uint64x2_t __p0,uint64x2_t __p1)60807 __ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
60808   uint64x2_t __ret;
60809   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60810   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60811   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
60812   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60813   return __ret;
60814 }
60815 #endif
60816 
60817 #ifdef __LITTLE_ENDIAN__
vtstq_s64(int64x2_t __p0,int64x2_t __p1)60818 __ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
60819   uint64x2_t __ret;
60820   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
60821   return __ret;
60822 }
60823 #else
vtstq_s64(int64x2_t __p0,int64x2_t __p1)60824 __ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
60825   uint64x2_t __ret;
60826   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60827   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60828   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
60829   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60830   return __ret;
60831 }
60832 #endif
60833 
vtst_u64(uint64x1_t __p0,uint64x1_t __p1)60834 __ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
60835   uint64x1_t __ret;
60836   __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
60837   return __ret;
60838 }
vtst_s64(int64x1_t __p0,int64x1_t __p1)60839 __ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
60840   uint64x1_t __ret;
60841   __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
60842   return __ret;
60843 }
vtstd_u64(uint64_t __p0,uint64_t __p1)60844 __ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
60845   uint64_t __ret;
60846   __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
60847   return __ret;
60848 }
vtstd_s64(int64_t __p0,int64_t __p1)60849 __ai uint64_t vtstd_s64(int64_t __p0, int64_t __p1) {
60850   uint64_t __ret;
60851   __ret = (uint64_t) __builtin_neon_vtstd_s64(__p0, __p1);
60852   return __ret;
60853 }
vuqaddb_s8(int8_t __p0,uint8_t __p1)60854 __ai int8_t vuqaddb_s8(int8_t __p0, uint8_t __p1) {
60855   int8_t __ret;
60856   __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
60857   return __ret;
60858 }
vuqadds_s32(int32_t __p0,uint32_t __p1)60859 __ai int32_t vuqadds_s32(int32_t __p0, uint32_t __p1) {
60860   int32_t __ret;
60861   __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
60862   return __ret;
60863 }
vuqaddd_s64(int64_t __p0,uint64_t __p1)60864 __ai int64_t vuqaddd_s64(int64_t __p0, uint64_t __p1) {
60865   int64_t __ret;
60866   __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
60867   return __ret;
60868 }
vuqaddh_s16(int16_t __p0,uint16_t __p1)60869 __ai int16_t vuqaddh_s16(int16_t __p0, uint16_t __p1) {
60870   int16_t __ret;
60871   __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
60872   return __ret;
60873 }
60874 #ifdef __LITTLE_ENDIAN__
vuqaddq_s8(int8x16_t __p0,uint8x16_t __p1)60875 __ai int8x16_t vuqaddq_s8(int8x16_t __p0, uint8x16_t __p1) {
60876   int8x16_t __ret;
60877   __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
60878   return __ret;
60879 }
60880 #else
vuqaddq_s8(int8x16_t __p0,uint8x16_t __p1)60881 __ai int8x16_t vuqaddq_s8(int8x16_t __p0, uint8x16_t __p1) {
60882   int8x16_t __ret;
60883   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60884   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60885   __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
60886   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60887   return __ret;
60888 }
60889 #endif
60890 
60891 #ifdef __LITTLE_ENDIAN__
vuqaddq_s32(int32x4_t __p0,uint32x4_t __p1)60892 __ai int32x4_t vuqaddq_s32(int32x4_t __p0, uint32x4_t __p1) {
60893   int32x4_t __ret;
60894   __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
60895   return __ret;
60896 }
60897 #else
vuqaddq_s32(int32x4_t __p0,uint32x4_t __p1)60898 __ai int32x4_t vuqaddq_s32(int32x4_t __p0, uint32x4_t __p1) {
60899   int32x4_t __ret;
60900   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60901   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60902   __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
60903   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60904   return __ret;
60905 }
60906 #endif
60907 
60908 #ifdef __LITTLE_ENDIAN__
vuqaddq_s64(int64x2_t __p0,uint64x2_t __p1)60909 __ai int64x2_t vuqaddq_s64(int64x2_t __p0, uint64x2_t __p1) {
60910   int64x2_t __ret;
60911   __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
60912   return __ret;
60913 }
60914 #else
vuqaddq_s64(int64x2_t __p0,uint64x2_t __p1)60915 __ai int64x2_t vuqaddq_s64(int64x2_t __p0, uint64x2_t __p1) {
60916   int64x2_t __ret;
60917   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60918   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60919   __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
60920   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60921   return __ret;
60922 }
60923 #endif
60924 
60925 #ifdef __LITTLE_ENDIAN__
vuqaddq_s16(int16x8_t __p0,uint16x8_t __p1)60926 __ai int16x8_t vuqaddq_s16(int16x8_t __p0, uint16x8_t __p1) {
60927   int16x8_t __ret;
60928   __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
60929   return __ret;
60930 }
60931 #else
vuqaddq_s16(int16x8_t __p0,uint16x8_t __p1)60932 __ai int16x8_t vuqaddq_s16(int16x8_t __p0, uint16x8_t __p1) {
60933   int16x8_t __ret;
60934   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60935   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60936   __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
60937   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60938   return __ret;
60939 }
60940 #endif
60941 
60942 #ifdef __LITTLE_ENDIAN__
vuqadd_s8(int8x8_t __p0,uint8x8_t __p1)60943 __ai int8x8_t vuqadd_s8(int8x8_t __p0, uint8x8_t __p1) {
60944   int8x8_t __ret;
60945   __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
60946   return __ret;
60947 }
60948 #else
vuqadd_s8(int8x8_t __p0,uint8x8_t __p1)60949 __ai int8x8_t vuqadd_s8(int8x8_t __p0, uint8x8_t __p1) {
60950   int8x8_t __ret;
60951   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60952   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60953   __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
60954   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60955   return __ret;
60956 }
60957 #endif
60958 
60959 #ifdef __LITTLE_ENDIAN__
vuqadd_s32(int32x2_t __p0,uint32x2_t __p1)60960 __ai int32x2_t vuqadd_s32(int32x2_t __p0, uint32x2_t __p1) {
60961   int32x2_t __ret;
60962   __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
60963   return __ret;
60964 }
60965 #else
vuqadd_s32(int32x2_t __p0,uint32x2_t __p1)60966 __ai int32x2_t vuqadd_s32(int32x2_t __p0, uint32x2_t __p1) {
60967   int32x2_t __ret;
60968   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60969   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60970   __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
60971   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60972   return __ret;
60973 }
60974 #endif
60975 
vuqadd_s64(int64x1_t __p0,uint64x1_t __p1)60976 __ai int64x1_t vuqadd_s64(int64x1_t __p0, uint64x1_t __p1) {
60977   int64x1_t __ret;
60978   __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
60979   return __ret;
60980 }
60981 #ifdef __LITTLE_ENDIAN__
vuqadd_s16(int16x4_t __p0,uint16x4_t __p1)60982 __ai int16x4_t vuqadd_s16(int16x4_t __p0, uint16x4_t __p1) {
60983   int16x4_t __ret;
60984   __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
60985   return __ret;
60986 }
60987 #else
vuqadd_s16(int16x4_t __p0,uint16x4_t __p1)60988 __ai int16x4_t vuqadd_s16(int16x4_t __p0, uint16x4_t __p1) {
60989   int16x4_t __ret;
60990   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60991   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60992   __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
60993   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60994   return __ret;
60995 }
60996 #endif
60997 
60998 #ifdef __LITTLE_ENDIAN__
vuzp1_p8(poly8x8_t __p0,poly8x8_t __p1)60999 __ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
61000   poly8x8_t __ret;
61001   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
61002   return __ret;
61003 }
61004 #else
vuzp1_p8(poly8x8_t __p0,poly8x8_t __p1)61005 __ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
61006   poly8x8_t __ret;
61007   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61008   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61009   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
61010   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61011   return __ret;
61012 }
61013 #endif
61014 
61015 #ifdef __LITTLE_ENDIAN__
vuzp1_p16(poly16x4_t __p0,poly16x4_t __p1)61016 __ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
61017   poly16x4_t __ret;
61018   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
61019   return __ret;
61020 }
61021 #else
vuzp1_p16(poly16x4_t __p0,poly16x4_t __p1)61022 __ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
61023   poly16x4_t __ret;
61024   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61025   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61026   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
61027   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61028   return __ret;
61029 }
61030 #endif
61031 
61032 #ifdef __LITTLE_ENDIAN__
vuzp1q_p8(poly8x16_t __p0,poly8x16_t __p1)61033 __ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
61034   poly8x16_t __ret;
61035   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
61036   return __ret;
61037 }
61038 #else
vuzp1q_p8(poly8x16_t __p0,poly8x16_t __p1)61039 __ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
61040   poly8x16_t __ret;
61041   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61042   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61043   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
61044   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61045   return __ret;
61046 }
61047 #endif
61048 
61049 #ifdef __LITTLE_ENDIAN__
vuzp1q_p64(poly64x2_t __p0,poly64x2_t __p1)61050 __ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
61051   poly64x2_t __ret;
61052   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61053   return __ret;
61054 }
61055 #else
vuzp1q_p64(poly64x2_t __p0,poly64x2_t __p1)61056 __ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
61057   poly64x2_t __ret;
61058   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61059   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61060   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61061   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61062   return __ret;
61063 }
61064 #endif
61065 
61066 #ifdef __LITTLE_ENDIAN__
vuzp1q_p16(poly16x8_t __p0,poly16x8_t __p1)61067 __ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
61068   poly16x8_t __ret;
61069   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
61070   return __ret;
61071 }
61072 #else
vuzp1q_p16(poly16x8_t __p0,poly16x8_t __p1)61073 __ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
61074   poly16x8_t __ret;
61075   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61076   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61077   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
61078   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61079   return __ret;
61080 }
61081 #endif
61082 
61083 #ifdef __LITTLE_ENDIAN__
vuzp1q_u8(uint8x16_t __p0,uint8x16_t __p1)61084 __ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
61085   uint8x16_t __ret;
61086   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
61087   return __ret;
61088 }
61089 #else
vuzp1q_u8(uint8x16_t __p0,uint8x16_t __p1)61090 __ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
61091   uint8x16_t __ret;
61092   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61093   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61094   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
61095   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61096   return __ret;
61097 }
61098 #endif
61099 
61100 #ifdef __LITTLE_ENDIAN__
vuzp1q_u32(uint32x4_t __p0,uint32x4_t __p1)61101 __ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
61102   uint32x4_t __ret;
61103   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
61104   return __ret;
61105 }
61106 #else
vuzp1q_u32(uint32x4_t __p0,uint32x4_t __p1)61107 __ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
61108   uint32x4_t __ret;
61109   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61110   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61111   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
61112   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61113   return __ret;
61114 }
61115 #endif
61116 
61117 #ifdef __LITTLE_ENDIAN__
vuzp1q_u64(uint64x2_t __p0,uint64x2_t __p1)61118 __ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
61119   uint64x2_t __ret;
61120   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61121   return __ret;
61122 }
61123 #else
vuzp1q_u64(uint64x2_t __p0,uint64x2_t __p1)61124 __ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
61125   uint64x2_t __ret;
61126   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61127   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61128   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61129   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61130   return __ret;
61131 }
61132 #endif
61133 
61134 #ifdef __LITTLE_ENDIAN__
vuzp1q_u16(uint16x8_t __p0,uint16x8_t __p1)61135 __ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
61136   uint16x8_t __ret;
61137   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
61138   return __ret;
61139 }
61140 #else
vuzp1q_u16(uint16x8_t __p0,uint16x8_t __p1)61141 __ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
61142   uint16x8_t __ret;
61143   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61144   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61145   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
61146   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61147   return __ret;
61148 }
61149 #endif
61150 
61151 #ifdef __LITTLE_ENDIAN__
vuzp1q_s8(int8x16_t __p0,int8x16_t __p1)61152 __ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
61153   int8x16_t __ret;
61154   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
61155   return __ret;
61156 }
61157 #else
vuzp1q_s8(int8x16_t __p0,int8x16_t __p1)61158 __ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
61159   int8x16_t __ret;
61160   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61161   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61162   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
61163   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61164   return __ret;
61165 }
61166 #endif
61167 
61168 #ifdef __LITTLE_ENDIAN__
vuzp1q_f64(float64x2_t __p0,float64x2_t __p1)61169 __ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
61170   float64x2_t __ret;
61171   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61172   return __ret;
61173 }
61174 #else
vuzp1q_f64(float64x2_t __p0,float64x2_t __p1)61175 __ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
61176   float64x2_t __ret;
61177   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61178   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61179   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61180   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61181   return __ret;
61182 }
61183 #endif
61184 
61185 #ifdef __LITTLE_ENDIAN__
vuzp1q_f32(float32x4_t __p0,float32x4_t __p1)61186 __ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
61187   float32x4_t __ret;
61188   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
61189   return __ret;
61190 }
61191 #else
vuzp1q_f32(float32x4_t __p0,float32x4_t __p1)61192 __ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
61193   float32x4_t __ret;
61194   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61195   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61196   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
61197   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61198   return __ret;
61199 }
61200 #endif
61201 
61202 #ifdef __LITTLE_ENDIAN__
vuzp1q_s32(int32x4_t __p0,int32x4_t __p1)61203 __ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
61204   int32x4_t __ret;
61205   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
61206   return __ret;
61207 }
61208 #else
vuzp1q_s32(int32x4_t __p0,int32x4_t __p1)61209 __ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
61210   int32x4_t __ret;
61211   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61212   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61213   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
61214   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61215   return __ret;
61216 }
61217 #endif
61218 
61219 #ifdef __LITTLE_ENDIAN__
vuzp1q_s64(int64x2_t __p0,int64x2_t __p1)61220 __ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
61221   int64x2_t __ret;
61222   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61223   return __ret;
61224 }
61225 #else
vuzp1q_s64(int64x2_t __p0,int64x2_t __p1)61226 __ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
61227   int64x2_t __ret;
61228   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61229   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61230   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61231   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61232   return __ret;
61233 }
61234 #endif
61235 
61236 #ifdef __LITTLE_ENDIAN__
vuzp1q_s16(int16x8_t __p0,int16x8_t __p1)61237 __ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
61238   int16x8_t __ret;
61239   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
61240   return __ret;
61241 }
61242 #else
vuzp1q_s16(int16x8_t __p0,int16x8_t __p1)61243 __ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
61244   int16x8_t __ret;
61245   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61246   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61247   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
61248   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61249   return __ret;
61250 }
61251 #endif
61252 
61253 #ifdef __LITTLE_ENDIAN__
vuzp1_u8(uint8x8_t __p0,uint8x8_t __p1)61254 __ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
61255   uint8x8_t __ret;
61256   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
61257   return __ret;
61258 }
61259 #else
vuzp1_u8(uint8x8_t __p0,uint8x8_t __p1)61260 __ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
61261   uint8x8_t __ret;
61262   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61263   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61264   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
61265   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61266   return __ret;
61267 }
61268 #endif
61269 
61270 #ifdef __LITTLE_ENDIAN__
vuzp1_u32(uint32x2_t __p0,uint32x2_t __p1)61271 __ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
61272   uint32x2_t __ret;
61273   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61274   return __ret;
61275 }
61276 #else
vuzp1_u32(uint32x2_t __p0,uint32x2_t __p1)61277 __ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
61278   uint32x2_t __ret;
61279   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61280   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61281   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61282   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61283   return __ret;
61284 }
61285 #endif
61286 
61287 #ifdef __LITTLE_ENDIAN__
vuzp1_u16(uint16x4_t __p0,uint16x4_t __p1)61288 __ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
61289   uint16x4_t __ret;
61290   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
61291   return __ret;
61292 }
61293 #else
vuzp1_u16(uint16x4_t __p0,uint16x4_t __p1)61294 __ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
61295   uint16x4_t __ret;
61296   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61297   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61298   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
61299   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61300   return __ret;
61301 }
61302 #endif
61303 
61304 #ifdef __LITTLE_ENDIAN__
vuzp1_s8(int8x8_t __p0,int8x8_t __p1)61305 __ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
61306   int8x8_t __ret;
61307   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
61308   return __ret;
61309 }
61310 #else
vuzp1_s8(int8x8_t __p0,int8x8_t __p1)61311 __ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
61312   int8x8_t __ret;
61313   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61314   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61315   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
61316   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61317   return __ret;
61318 }
61319 #endif
61320 
61321 #ifdef __LITTLE_ENDIAN__
vuzp1_f32(float32x2_t __p0,float32x2_t __p1)61322 __ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
61323   float32x2_t __ret;
61324   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61325   return __ret;
61326 }
61327 #else
vuzp1_f32(float32x2_t __p0,float32x2_t __p1)61328 __ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
61329   float32x2_t __ret;
61330   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61331   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61332   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61333   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61334   return __ret;
61335 }
61336 #endif
61337 
61338 #ifdef __LITTLE_ENDIAN__
vuzp1_s32(int32x2_t __p0,int32x2_t __p1)61339 __ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
61340   int32x2_t __ret;
61341   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61342   return __ret;
61343 }
61344 #else
vuzp1_s32(int32x2_t __p0,int32x2_t __p1)61345 __ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
61346   int32x2_t __ret;
61347   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61348   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61349   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61350   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61351   return __ret;
61352 }
61353 #endif
61354 
61355 #ifdef __LITTLE_ENDIAN__
vuzp1_s16(int16x4_t __p0,int16x4_t __p1)61356 __ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
61357   int16x4_t __ret;
61358   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
61359   return __ret;
61360 }
61361 #else
vuzp1_s16(int16x4_t __p0,int16x4_t __p1)61362 __ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
61363   int16x4_t __ret;
61364   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61365   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61366   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
61367   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61368   return __ret;
61369 }
61370 #endif
61371 
61372 #ifdef __LITTLE_ENDIAN__
vuzp2_p8(poly8x8_t __p0,poly8x8_t __p1)61373 __ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
61374   poly8x8_t __ret;
61375   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
61376   return __ret;
61377 }
61378 #else
vuzp2_p8(poly8x8_t __p0,poly8x8_t __p1)61379 __ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
61380   poly8x8_t __ret;
61381   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61382   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61383   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
61384   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61385   return __ret;
61386 }
61387 #endif
61388 
61389 #ifdef __LITTLE_ENDIAN__
vuzp2_p16(poly16x4_t __p0,poly16x4_t __p1)61390 __ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
61391   poly16x4_t __ret;
61392   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
61393   return __ret;
61394 }
61395 #else
vuzp2_p16(poly16x4_t __p0,poly16x4_t __p1)61396 __ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
61397   poly16x4_t __ret;
61398   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61399   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61400   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
61401   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61402   return __ret;
61403 }
61404 #endif
61405 
61406 #ifdef __LITTLE_ENDIAN__
vuzp2q_p8(poly8x16_t __p0,poly8x16_t __p1)61407 __ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
61408   poly8x16_t __ret;
61409   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
61410   return __ret;
61411 }
61412 #else
vuzp2q_p8(poly8x16_t __p0,poly8x16_t __p1)61413 __ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
61414   poly8x16_t __ret;
61415   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61416   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61417   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
61418   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61419   return __ret;
61420 }
61421 #endif
61422 
61423 #ifdef __LITTLE_ENDIAN__
vuzp2q_p64(poly64x2_t __p0,poly64x2_t __p1)61424 __ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
61425   poly64x2_t __ret;
61426   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
61427   return __ret;
61428 }
61429 #else
vuzp2q_p64(poly64x2_t __p0,poly64x2_t __p1)61430 __ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
61431   poly64x2_t __ret;
61432   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61433   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61434   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
61435   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61436   return __ret;
61437 }
61438 #endif
61439 
61440 #ifdef __LITTLE_ENDIAN__
vuzp2q_p16(poly16x8_t __p0,poly16x8_t __p1)61441 __ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
61442   poly16x8_t __ret;
61443   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
61444   return __ret;
61445 }
61446 #else
vuzp2q_p16(poly16x8_t __p0,poly16x8_t __p1)61447 __ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
61448   poly16x8_t __ret;
61449   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61450   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61451   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
61452   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61453   return __ret;
61454 }
61455 #endif
61456 
61457 #ifdef __LITTLE_ENDIAN__
vuzp2q_u8(uint8x16_t __p0,uint8x16_t __p1)61458 __ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
61459   uint8x16_t __ret;
61460   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
61461   return __ret;
61462 }
61463 #else
vuzp2q_u8(uint8x16_t __p0,uint8x16_t __p1)61464 __ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
61465   uint8x16_t __ret;
61466   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61467   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61468   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
61469   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61470   return __ret;
61471 }
61472 #endif
61473 
61474 #ifdef __LITTLE_ENDIAN__
vuzp2q_u32(uint32x4_t __p0,uint32x4_t __p1)61475 __ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
61476   uint32x4_t __ret;
61477   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
61478   return __ret;
61479 }
61480 #else
vuzp2q_u32(uint32x4_t __p0,uint32x4_t __p1)61481 __ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
61482   uint32x4_t __ret;
61483   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61484   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61485   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
61486   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61487   return __ret;
61488 }
61489 #endif
61490 
61491 #ifdef __LITTLE_ENDIAN__
vuzp2q_u64(uint64x2_t __p0,uint64x2_t __p1)61492 __ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
61493   uint64x2_t __ret;
61494   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
61495   return __ret;
61496 }
61497 #else
vuzp2q_u64(uint64x2_t __p0,uint64x2_t __p1)61498 __ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
61499   uint64x2_t __ret;
61500   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61501   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61502   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
61503   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61504   return __ret;
61505 }
61506 #endif
61507 
61508 #ifdef __LITTLE_ENDIAN__
vuzp2q_u16(uint16x8_t __p0,uint16x8_t __p1)61509 __ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
61510   uint16x8_t __ret;
61511   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
61512   return __ret;
61513 }
61514 #else
vuzp2q_u16(uint16x8_t __p0,uint16x8_t __p1)61515 __ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
61516   uint16x8_t __ret;
61517   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61518   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61519   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
61520   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61521   return __ret;
61522 }
61523 #endif
61524 
61525 #ifdef __LITTLE_ENDIAN__
vuzp2q_s8(int8x16_t __p0,int8x16_t __p1)61526 __ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
61527   int8x16_t __ret;
61528   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
61529   return __ret;
61530 }
61531 #else
vuzp2q_s8(int8x16_t __p0,int8x16_t __p1)61532 __ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
61533   int8x16_t __ret;
61534   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61535   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61536   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
61537   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61538   return __ret;
61539 }
61540 #endif
61541 
61542 #ifdef __LITTLE_ENDIAN__
vuzp2q_f64(float64x2_t __p0,float64x2_t __p1)61543 __ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
61544   float64x2_t __ret;
61545   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
61546   return __ret;
61547 }
61548 #else
vuzp2q_f64(float64x2_t __p0,float64x2_t __p1)61549 __ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
61550   float64x2_t __ret;
61551   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61552   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61553   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
61554   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61555   return __ret;
61556 }
61557 #endif
61558 
61559 #ifdef __LITTLE_ENDIAN__
vuzp2q_f32(float32x4_t __p0,float32x4_t __p1)61560 __ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
61561   float32x4_t __ret;
61562   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
61563   return __ret;
61564 }
61565 #else
vuzp2q_f32(float32x4_t __p0,float32x4_t __p1)61566 __ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
61567   float32x4_t __ret;
61568   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61569   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61570   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
61571   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61572   return __ret;
61573 }
61574 #endif
61575 
61576 #ifdef __LITTLE_ENDIAN__
vuzp2q_s32(int32x4_t __p0,int32x4_t __p1)61577 __ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
61578   int32x4_t __ret;
61579   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
61580   return __ret;
61581 }
61582 #else
vuzp2q_s32(int32x4_t __p0,int32x4_t __p1)61583 __ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
61584   int32x4_t __ret;
61585   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61586   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61587   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
61588   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61589   return __ret;
61590 }
61591 #endif
61592 
61593 #ifdef __LITTLE_ENDIAN__
vuzp2q_s64(int64x2_t __p0,int64x2_t __p1)61594 __ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
61595   int64x2_t __ret;
61596   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
61597   return __ret;
61598 }
61599 #else
vuzp2q_s64(int64x2_t __p0,int64x2_t __p1)61600 __ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
61601   int64x2_t __ret;
61602   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61603   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61604   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
61605   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61606   return __ret;
61607 }
61608 #endif
61609 
61610 #ifdef __LITTLE_ENDIAN__
vuzp2q_s16(int16x8_t __p0,int16x8_t __p1)61611 __ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
61612   int16x8_t __ret;
61613   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
61614   return __ret;
61615 }
61616 #else
vuzp2q_s16(int16x8_t __p0,int16x8_t __p1)61617 __ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
61618   int16x8_t __ret;
61619   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61620   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61621   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
61622   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61623   return __ret;
61624 }
61625 #endif
61626 
61627 #ifdef __LITTLE_ENDIAN__
vuzp2_u8(uint8x8_t __p0,uint8x8_t __p1)61628 __ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
61629   uint8x8_t __ret;
61630   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
61631   return __ret;
61632 }
61633 #else
vuzp2_u8(uint8x8_t __p0,uint8x8_t __p1)61634 __ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
61635   uint8x8_t __ret;
61636   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61637   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61638   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
61639   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61640   return __ret;
61641 }
61642 #endif
61643 
61644 #ifdef __LITTLE_ENDIAN__
vuzp2_u32(uint32x2_t __p0,uint32x2_t __p1)61645 __ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
61646   uint32x2_t __ret;
61647   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
61648   return __ret;
61649 }
61650 #else
vuzp2_u32(uint32x2_t __p0,uint32x2_t __p1)61651 __ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
61652   uint32x2_t __ret;
61653   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61654   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61655   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
61656   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61657   return __ret;
61658 }
61659 #endif
61660 
61661 #ifdef __LITTLE_ENDIAN__
vuzp2_u16(uint16x4_t __p0,uint16x4_t __p1)61662 __ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
61663   uint16x4_t __ret;
61664   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
61665   return __ret;
61666 }
61667 #else
vuzp2_u16(uint16x4_t __p0,uint16x4_t __p1)61668 __ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
61669   uint16x4_t __ret;
61670   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61671   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61672   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
61673   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61674   return __ret;
61675 }
61676 #endif
61677 
61678 #ifdef __LITTLE_ENDIAN__
vuzp2_s8(int8x8_t __p0,int8x8_t __p1)61679 __ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
61680   int8x8_t __ret;
61681   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
61682   return __ret;
61683 }
61684 #else
vuzp2_s8(int8x8_t __p0,int8x8_t __p1)61685 __ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
61686   int8x8_t __ret;
61687   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61688   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61689   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
61690   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61691   return __ret;
61692 }
61693 #endif
61694 
61695 #ifdef __LITTLE_ENDIAN__
vuzp2_f32(float32x2_t __p0,float32x2_t __p1)61696 __ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
61697   float32x2_t __ret;
61698   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
61699   return __ret;
61700 }
61701 #else
vuzp2_f32(float32x2_t __p0,float32x2_t __p1)61702 __ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
61703   float32x2_t __ret;
61704   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61705   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61706   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
61707   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61708   return __ret;
61709 }
61710 #endif
61711 
61712 #ifdef __LITTLE_ENDIAN__
vuzp2_s32(int32x2_t __p0,int32x2_t __p1)61713 __ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
61714   int32x2_t __ret;
61715   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
61716   return __ret;
61717 }
61718 #else
vuzp2_s32(int32x2_t __p0,int32x2_t __p1)61719 __ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
61720   int32x2_t __ret;
61721   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61722   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61723   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
61724   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61725   return __ret;
61726 }
61727 #endif
61728 
61729 #ifdef __LITTLE_ENDIAN__
vuzp2_s16(int16x4_t __p0,int16x4_t __p1)61730 __ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
61731   int16x4_t __ret;
61732   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
61733   return __ret;
61734 }
61735 #else
vuzp2_s16(int16x4_t __p0,int16x4_t __p1)61736 __ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
61737   int16x4_t __ret;
61738   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61739   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61740   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
61741   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61742   return __ret;
61743 }
61744 #endif
61745 
61746 #ifdef __LITTLE_ENDIAN__
vzip1_p8(poly8x8_t __p0,poly8x8_t __p1)61747 __ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
61748   poly8x8_t __ret;
61749   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
61750   return __ret;
61751 }
61752 #else
vzip1_p8(poly8x8_t __p0,poly8x8_t __p1)61753 __ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
61754   poly8x8_t __ret;
61755   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61756   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61757   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
61758   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61759   return __ret;
61760 }
61761 #endif
61762 
61763 #ifdef __LITTLE_ENDIAN__
vzip1_p16(poly16x4_t __p0,poly16x4_t __p1)61764 __ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
61765   poly16x4_t __ret;
61766   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
61767   return __ret;
61768 }
61769 #else
vzip1_p16(poly16x4_t __p0,poly16x4_t __p1)61770 __ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
61771   poly16x4_t __ret;
61772   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61773   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61774   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
61775   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61776   return __ret;
61777 }
61778 #endif
61779 
61780 #ifdef __LITTLE_ENDIAN__
vzip1q_p8(poly8x16_t __p0,poly8x16_t __p1)61781 __ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
61782   poly8x16_t __ret;
61783   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
61784   return __ret;
61785 }
61786 #else
vzip1q_p8(poly8x16_t __p0,poly8x16_t __p1)61787 __ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
61788   poly8x16_t __ret;
61789   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61790   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61791   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
61792   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61793   return __ret;
61794 }
61795 #endif
61796 
61797 #ifdef __LITTLE_ENDIAN__
vzip1q_p64(poly64x2_t __p0,poly64x2_t __p1)61798 __ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
61799   poly64x2_t __ret;
61800   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61801   return __ret;
61802 }
61803 #else
vzip1q_p64(poly64x2_t __p0,poly64x2_t __p1)61804 __ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
61805   poly64x2_t __ret;
61806   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61807   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61808   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61809   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61810   return __ret;
61811 }
61812 #endif
61813 
61814 #ifdef __LITTLE_ENDIAN__
vzip1q_p16(poly16x8_t __p0,poly16x8_t __p1)61815 __ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
61816   poly16x8_t __ret;
61817   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
61818   return __ret;
61819 }
61820 #else
vzip1q_p16(poly16x8_t __p0,poly16x8_t __p1)61821 __ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
61822   poly16x8_t __ret;
61823   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61824   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61825   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
61826   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61827   return __ret;
61828 }
61829 #endif
61830 
61831 #ifdef __LITTLE_ENDIAN__
vzip1q_u8(uint8x16_t __p0,uint8x16_t __p1)61832 __ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
61833   uint8x16_t __ret;
61834   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
61835   return __ret;
61836 }
61837 #else
vzip1q_u8(uint8x16_t __p0,uint8x16_t __p1)61838 __ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
61839   uint8x16_t __ret;
61840   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61841   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61842   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
61843   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61844   return __ret;
61845 }
61846 #endif
61847 
61848 #ifdef __LITTLE_ENDIAN__
vzip1q_u32(uint32x4_t __p0,uint32x4_t __p1)61849 __ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
61850   uint32x4_t __ret;
61851   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
61852   return __ret;
61853 }
61854 #else
vzip1q_u32(uint32x4_t __p0,uint32x4_t __p1)61855 __ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
61856   uint32x4_t __ret;
61857   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61858   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61859   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
61860   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61861   return __ret;
61862 }
61863 #endif
61864 
61865 #ifdef __LITTLE_ENDIAN__
vzip1q_u64(uint64x2_t __p0,uint64x2_t __p1)61866 __ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
61867   uint64x2_t __ret;
61868   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61869   return __ret;
61870 }
61871 #else
vzip1q_u64(uint64x2_t __p0,uint64x2_t __p1)61872 __ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
61873   uint64x2_t __ret;
61874   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61875   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61876   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61877   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61878   return __ret;
61879 }
61880 #endif
61881 
61882 #ifdef __LITTLE_ENDIAN__
vzip1q_u16(uint16x8_t __p0,uint16x8_t __p1)61883 __ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
61884   uint16x8_t __ret;
61885   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
61886   return __ret;
61887 }
61888 #else
vzip1q_u16(uint16x8_t __p0,uint16x8_t __p1)61889 __ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
61890   uint16x8_t __ret;
61891   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61892   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61893   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
61894   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61895   return __ret;
61896 }
61897 #endif
61898 
61899 #ifdef __LITTLE_ENDIAN__
vzip1q_s8(int8x16_t __p0,int8x16_t __p1)61900 __ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
61901   int8x16_t __ret;
61902   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
61903   return __ret;
61904 }
61905 #else
vzip1q_s8(int8x16_t __p0,int8x16_t __p1)61906 __ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
61907   int8x16_t __ret;
61908   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61909   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61910   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
61911   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61912   return __ret;
61913 }
61914 #endif
61915 
61916 #ifdef __LITTLE_ENDIAN__
vzip1q_f64(float64x2_t __p0,float64x2_t __p1)61917 __ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
61918   float64x2_t __ret;
61919   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61920   return __ret;
61921 }
61922 #else
vzip1q_f64(float64x2_t __p0,float64x2_t __p1)61923 __ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
61924   float64x2_t __ret;
61925   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61926   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61927   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61928   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61929   return __ret;
61930 }
61931 #endif
61932 
61933 #ifdef __LITTLE_ENDIAN__
vzip1q_f32(float32x4_t __p0,float32x4_t __p1)61934 __ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
61935   float32x4_t __ret;
61936   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
61937   return __ret;
61938 }
61939 #else
vzip1q_f32(float32x4_t __p0,float32x4_t __p1)61940 __ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
61941   float32x4_t __ret;
61942   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61943   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61944   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
61945   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61946   return __ret;
61947 }
61948 #endif
61949 
61950 #ifdef __LITTLE_ENDIAN__
vzip1q_s32(int32x4_t __p0,int32x4_t __p1)61951 __ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
61952   int32x4_t __ret;
61953   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
61954   return __ret;
61955 }
61956 #else
vzip1q_s32(int32x4_t __p0,int32x4_t __p1)61957 __ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
61958   int32x4_t __ret;
61959   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61960   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61961   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
61962   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61963   return __ret;
61964 }
61965 #endif
61966 
61967 #ifdef __LITTLE_ENDIAN__
vzip1q_s64(int64x2_t __p0,int64x2_t __p1)61968 __ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
61969   int64x2_t __ret;
61970   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61971   return __ret;
61972 }
61973 #else
vzip1q_s64(int64x2_t __p0,int64x2_t __p1)61974 __ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
61975   int64x2_t __ret;
61976   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61977   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61978   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61979   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61980   return __ret;
61981 }
61982 #endif
61983 
61984 #ifdef __LITTLE_ENDIAN__
vzip1q_s16(int16x8_t __p0,int16x8_t __p1)61985 __ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
61986   int16x8_t __ret;
61987   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
61988   return __ret;
61989 }
61990 #else
vzip1q_s16(int16x8_t __p0,int16x8_t __p1)61991 __ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
61992   int16x8_t __ret;
61993   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61994   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61995   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
61996   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61997   return __ret;
61998 }
61999 #endif
62000 
62001 #ifdef __LITTLE_ENDIAN__
vzip1_u8(uint8x8_t __p0,uint8x8_t __p1)62002 __ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
62003   uint8x8_t __ret;
62004   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
62005   return __ret;
62006 }
62007 #else
vzip1_u8(uint8x8_t __p0,uint8x8_t __p1)62008 __ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
62009   uint8x8_t __ret;
62010   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62011   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62012   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
62013   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62014   return __ret;
62015 }
62016 #endif
62017 
62018 #ifdef __LITTLE_ENDIAN__
vzip1_u32(uint32x2_t __p0,uint32x2_t __p1)62019 __ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
62020   uint32x2_t __ret;
62021   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62022   return __ret;
62023 }
62024 #else
vzip1_u32(uint32x2_t __p0,uint32x2_t __p1)62025 __ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
62026   uint32x2_t __ret;
62027   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62028   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62029   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62030   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62031   return __ret;
62032 }
62033 #endif
62034 
62035 #ifdef __LITTLE_ENDIAN__
vzip1_u16(uint16x4_t __p0,uint16x4_t __p1)62036 __ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
62037   uint16x4_t __ret;
62038   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
62039   return __ret;
62040 }
62041 #else
vzip1_u16(uint16x4_t __p0,uint16x4_t __p1)62042 __ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
62043   uint16x4_t __ret;
62044   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62045   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62046   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
62047   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62048   return __ret;
62049 }
62050 #endif
62051 
62052 #ifdef __LITTLE_ENDIAN__
vzip1_s8(int8x8_t __p0,int8x8_t __p1)62053 __ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
62054   int8x8_t __ret;
62055   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
62056   return __ret;
62057 }
62058 #else
vzip1_s8(int8x8_t __p0,int8x8_t __p1)62059 __ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
62060   int8x8_t __ret;
62061   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62062   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62063   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
62064   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62065   return __ret;
62066 }
62067 #endif
62068 
62069 #ifdef __LITTLE_ENDIAN__
vzip1_f32(float32x2_t __p0,float32x2_t __p1)62070 __ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
62071   float32x2_t __ret;
62072   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62073   return __ret;
62074 }
62075 #else
vzip1_f32(float32x2_t __p0,float32x2_t __p1)62076 __ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
62077   float32x2_t __ret;
62078   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62079   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62080   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62081   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62082   return __ret;
62083 }
62084 #endif
62085 
62086 #ifdef __LITTLE_ENDIAN__
vzip1_s32(int32x2_t __p0,int32x2_t __p1)62087 __ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
62088   int32x2_t __ret;
62089   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62090   return __ret;
62091 }
62092 #else
vzip1_s32(int32x2_t __p0,int32x2_t __p1)62093 __ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
62094   int32x2_t __ret;
62095   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62096   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62097   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62098   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62099   return __ret;
62100 }
62101 #endif
62102 
62103 #ifdef __LITTLE_ENDIAN__
vzip1_s16(int16x4_t __p0,int16x4_t __p1)62104 __ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
62105   int16x4_t __ret;
62106   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
62107   return __ret;
62108 }
62109 #else
vzip1_s16(int16x4_t __p0,int16x4_t __p1)62110 __ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
62111   int16x4_t __ret;
62112   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62113   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62114   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
62115   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62116   return __ret;
62117 }
62118 #endif
62119 
62120 #ifdef __LITTLE_ENDIAN__
vzip2_p8(poly8x8_t __p0,poly8x8_t __p1)62121 __ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
62122   poly8x8_t __ret;
62123   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
62124   return __ret;
62125 }
62126 #else
vzip2_p8(poly8x8_t __p0,poly8x8_t __p1)62127 __ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
62128   poly8x8_t __ret;
62129   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62130   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62131   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
62132   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62133   return __ret;
62134 }
62135 #endif
62136 
62137 #ifdef __LITTLE_ENDIAN__
vzip2_p16(poly16x4_t __p0,poly16x4_t __p1)62138 __ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
62139   poly16x4_t __ret;
62140   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
62141   return __ret;
62142 }
62143 #else
vzip2_p16(poly16x4_t __p0,poly16x4_t __p1)62144 __ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
62145   poly16x4_t __ret;
62146   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62147   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62148   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
62149   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62150   return __ret;
62151 }
62152 #endif
62153 
62154 #ifdef __LITTLE_ENDIAN__
vzip2q_p8(poly8x16_t __p0,poly8x16_t __p1)62155 __ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
62156   poly8x16_t __ret;
62157   __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
62158   return __ret;
62159 }
62160 #else
vzip2q_p8(poly8x16_t __p0,poly8x16_t __p1)62161 __ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
62162   poly8x16_t __ret;
62163   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62164   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62165   __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
62166   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62167   return __ret;
62168 }
62169 #endif
62170 
62171 #ifdef __LITTLE_ENDIAN__
vzip2q_p64(poly64x2_t __p0,poly64x2_t __p1)62172 __ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
62173   poly64x2_t __ret;
62174   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62175   return __ret;
62176 }
62177 #else
vzip2q_p64(poly64x2_t __p0,poly64x2_t __p1)62178 __ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
62179   poly64x2_t __ret;
62180   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62181   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62182   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62183   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62184   return __ret;
62185 }
62186 #endif
62187 
62188 #ifdef __LITTLE_ENDIAN__
vzip2q_p16(poly16x8_t __p0,poly16x8_t __p1)62189 __ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
62190   poly16x8_t __ret;
62191   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
62192   return __ret;
62193 }
62194 #else
vzip2q_p16(poly16x8_t __p0,poly16x8_t __p1)62195 __ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
62196   poly16x8_t __ret;
62197   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62198   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62199   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
62200   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62201   return __ret;
62202 }
62203 #endif
62204 
62205 #ifdef __LITTLE_ENDIAN__
vzip2q_u8(uint8x16_t __p0,uint8x16_t __p1)62206 __ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
62207   uint8x16_t __ret;
62208   __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
62209   return __ret;
62210 }
62211 #else
vzip2q_u8(uint8x16_t __p0,uint8x16_t __p1)62212 __ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
62213   uint8x16_t __ret;
62214   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62215   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62216   __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
62217   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62218   return __ret;
62219 }
62220 #endif
62221 
62222 #ifdef __LITTLE_ENDIAN__
vzip2q_u32(uint32x4_t __p0,uint32x4_t __p1)62223 __ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
62224   uint32x4_t __ret;
62225   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
62226   return __ret;
62227 }
62228 #else
vzip2q_u32(uint32x4_t __p0,uint32x4_t __p1)62229 __ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
62230   uint32x4_t __ret;
62231   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62232   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62233   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
62234   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62235   return __ret;
62236 }
62237 #endif
62238 
62239 #ifdef __LITTLE_ENDIAN__
vzip2q_u64(uint64x2_t __p0,uint64x2_t __p1)62240 __ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
62241   uint64x2_t __ret;
62242   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62243   return __ret;
62244 }
62245 #else
vzip2q_u64(uint64x2_t __p0,uint64x2_t __p1)62246 __ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
62247   uint64x2_t __ret;
62248   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62249   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62250   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62251   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62252   return __ret;
62253 }
62254 #endif
62255 
62256 #ifdef __LITTLE_ENDIAN__
vzip2q_u16(uint16x8_t __p0,uint16x8_t __p1)62257 __ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
62258   uint16x8_t __ret;
62259   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
62260   return __ret;
62261 }
62262 #else
vzip2q_u16(uint16x8_t __p0,uint16x8_t __p1)62263 __ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
62264   uint16x8_t __ret;
62265   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62266   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62267   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
62268   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62269   return __ret;
62270 }
62271 #endif
62272 
62273 #ifdef __LITTLE_ENDIAN__
vzip2q_s8(int8x16_t __p0,int8x16_t __p1)62274 __ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
62275   int8x16_t __ret;
62276   __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
62277   return __ret;
62278 }
62279 #else
vzip2q_s8(int8x16_t __p0,int8x16_t __p1)62280 __ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
62281   int8x16_t __ret;
62282   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62283   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62284   __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
62285   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62286   return __ret;
62287 }
62288 #endif
62289 
62290 #ifdef __LITTLE_ENDIAN__
vzip2q_f64(float64x2_t __p0,float64x2_t __p1)62291 __ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
62292   float64x2_t __ret;
62293   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62294   return __ret;
62295 }
62296 #else
vzip2q_f64(float64x2_t __p0,float64x2_t __p1)62297 __ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
62298   float64x2_t __ret;
62299   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62300   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62301   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62302   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62303   return __ret;
62304 }
62305 #endif
62306 
62307 #ifdef __LITTLE_ENDIAN__
vzip2q_f32(float32x4_t __p0,float32x4_t __p1)62308 __ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
62309   float32x4_t __ret;
62310   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
62311   return __ret;
62312 }
62313 #else
vzip2q_f32(float32x4_t __p0,float32x4_t __p1)62314 __ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
62315   float32x4_t __ret;
62316   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62317   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62318   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
62319   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62320   return __ret;
62321 }
62322 #endif
62323 
62324 #ifdef __LITTLE_ENDIAN__
vzip2q_s32(int32x4_t __p0,int32x4_t __p1)62325 __ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
62326   int32x4_t __ret;
62327   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
62328   return __ret;
62329 }
62330 #else
vzip2q_s32(int32x4_t __p0,int32x4_t __p1)62331 __ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
62332   int32x4_t __ret;
62333   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62334   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62335   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
62336   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62337   return __ret;
62338 }
62339 #endif
62340 
62341 #ifdef __LITTLE_ENDIAN__
vzip2q_s64(int64x2_t __p0,int64x2_t __p1)62342 __ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
62343   int64x2_t __ret;
62344   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62345   return __ret;
62346 }
62347 #else
vzip2q_s64(int64x2_t __p0,int64x2_t __p1)62348 __ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
62349   int64x2_t __ret;
62350   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62351   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62352   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62353   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62354   return __ret;
62355 }
62356 #endif
62357 
62358 #ifdef __LITTLE_ENDIAN__
vzip2q_s16(int16x8_t __p0,int16x8_t __p1)62359 __ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
62360   int16x8_t __ret;
62361   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
62362   return __ret;
62363 }
62364 #else
vzip2q_s16(int16x8_t __p0,int16x8_t __p1)62365 __ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
62366   int16x8_t __ret;
62367   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62368   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62369   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
62370   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62371   return __ret;
62372 }
62373 #endif
62374 
62375 #ifdef __LITTLE_ENDIAN__
vzip2_u8(uint8x8_t __p0,uint8x8_t __p1)62376 __ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
62377   uint8x8_t __ret;
62378   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
62379   return __ret;
62380 }
62381 #else
vzip2_u8(uint8x8_t __p0,uint8x8_t __p1)62382 __ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
62383   uint8x8_t __ret;
62384   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62385   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62386   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
62387   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62388   return __ret;
62389 }
62390 #endif
62391 
62392 #ifdef __LITTLE_ENDIAN__
vzip2_u32(uint32x2_t __p0,uint32x2_t __p1)62393 __ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
62394   uint32x2_t __ret;
62395   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62396   return __ret;
62397 }
62398 #else
vzip2_u32(uint32x2_t __p0,uint32x2_t __p1)62399 __ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
62400   uint32x2_t __ret;
62401   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62402   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62403   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62404   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62405   return __ret;
62406 }
62407 #endif
62408 
62409 #ifdef __LITTLE_ENDIAN__
vzip2_u16(uint16x4_t __p0,uint16x4_t __p1)62410 __ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
62411   uint16x4_t __ret;
62412   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
62413   return __ret;
62414 }
62415 #else
vzip2_u16(uint16x4_t __p0,uint16x4_t __p1)62416 __ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
62417   uint16x4_t __ret;
62418   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62419   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62420   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
62421   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62422   return __ret;
62423 }
62424 #endif
62425 
62426 #ifdef __LITTLE_ENDIAN__
vzip2_s8(int8x8_t __p0,int8x8_t __p1)62427 __ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
62428   int8x8_t __ret;
62429   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
62430   return __ret;
62431 }
62432 #else
vzip2_s8(int8x8_t __p0,int8x8_t __p1)62433 __ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
62434   int8x8_t __ret;
62435   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62436   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62437   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
62438   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62439   return __ret;
62440 }
62441 #endif
62442 
62443 #ifdef __LITTLE_ENDIAN__
vzip2_f32(float32x2_t __p0,float32x2_t __p1)62444 __ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
62445   float32x2_t __ret;
62446   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62447   return __ret;
62448 }
62449 #else
vzip2_f32(float32x2_t __p0,float32x2_t __p1)62450 __ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
62451   float32x2_t __ret;
62452   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62453   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62454   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62455   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62456   return __ret;
62457 }
62458 #endif
62459 
62460 #ifdef __LITTLE_ENDIAN__
vzip2_s32(int32x2_t __p0,int32x2_t __p1)62461 __ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
62462   int32x2_t __ret;
62463   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62464   return __ret;
62465 }
62466 #else
vzip2_s32(int32x2_t __p0,int32x2_t __p1)62467 __ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
62468   int32x2_t __ret;
62469   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62470   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62471   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62472   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62473   return __ret;
62474 }
62475 #endif
62476 
62477 #ifdef __LITTLE_ENDIAN__
vzip2_s16(int16x4_t __p0,int16x4_t __p1)62478 __ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
62479   int16x4_t __ret;
62480   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
62481   return __ret;
62482 }
62483 #else
vzip2_s16(int16x4_t __p0,int16x4_t __p1)62484 __ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
62485   int16x4_t __ret;
62486   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62487   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62488   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
62489   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62490   return __ret;
62491 }
62492 #endif
62493 
vmull_p64(poly64_t __p0,poly64_t __p1)62494 __ai __attribute__((target("aes"))) poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
62495   poly128_t __ret;
62496   __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
62497   return __ret;
62498 }
62499 #ifdef __LITTLE_ENDIAN__
vmull_high_p64(poly64x2_t __p0,poly64x2_t __p1)62500 __ai __attribute__((target("aes"))) poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
62501   poly128_t __ret;
62502   __ret = vmull_p64((poly64_t)(vget_high_p64(__p0)), (poly64_t)(vget_high_p64(__p1)));
62503   return __ret;
62504 }
62505 #else
vmull_high_p64(poly64x2_t __p0,poly64x2_t __p1)62506 __ai __attribute__((target("aes"))) poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
62507   poly128_t __ret;
62508   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62509   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62510   __ret = vmull_p64((poly64_t)(__noswap_vget_high_p64(__rev0)), (poly64_t)(__noswap_vget_high_p64(__rev1)));
62511   return __ret;
62512 }
62513 #endif
62514 
62515 #ifdef __LITTLE_ENDIAN__
__a64_vcvtq_low_bf16_f32(float32x4_t __p0)62516 __ai __attribute__((target("bf16"))) bfloat16x8_t __a64_vcvtq_low_bf16_f32(float32x4_t __p0) {
62517   bfloat16x8_t __ret;
62518   __ret = (bfloat16x8_t) __builtin_neon___a64_vcvtq_low_bf16_f32((int8x16_t)__p0, 43);
62519   return __ret;
62520 }
62521 #else
__a64_vcvtq_low_bf16_f32(float32x4_t __p0)62522 __ai __attribute__((target("bf16"))) bfloat16x8_t __a64_vcvtq_low_bf16_f32(float32x4_t __p0) {
62523   bfloat16x8_t __ret;
62524   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62525   __ret = (bfloat16x8_t) __builtin_neon___a64_vcvtq_low_bf16_f32((int8x16_t)__rev0, 43);
62526   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62527   return __ret;
62528 }
__noswap___a64_vcvtq_low_bf16_f32(float32x4_t __p0)62529 __ai __attribute__((target("bf16"))) bfloat16x8_t __noswap___a64_vcvtq_low_bf16_f32(float32x4_t __p0) {
62530   bfloat16x8_t __ret;
62531   __ret = (bfloat16x8_t) __builtin_neon___a64_vcvtq_low_bf16_f32((int8x16_t)__p0, 43);
62532   return __ret;
62533 }
62534 #endif
62535 
62536 #ifdef __LITTLE_ENDIAN__
62537 #define vcopyq_lane_bf16(__p0_739, __p1_739, __p2_739, __p3_739) __extension__ ({ \
62538   bfloat16x8_t __ret_739; \
62539   bfloat16x8_t __s0_739 = __p0_739; \
62540   bfloat16x4_t __s2_739 = __p2_739; \
62541   __ret_739 = vsetq_lane_bf16(vget_lane_bf16(__s2_739, __p3_739), __s0_739, __p1_739); \
62542   __ret_739; \
62543 })
62544 #else
62545 #define vcopyq_lane_bf16(__p0_740, __p1_740, __p2_740, __p3_740) __extension__ ({ \
62546   bfloat16x8_t __ret_740; \
62547   bfloat16x8_t __s0_740 = __p0_740; \
62548   bfloat16x4_t __s2_740 = __p2_740; \
62549   bfloat16x8_t __rev0_740;  __rev0_740 = __builtin_shufflevector(__s0_740, __s0_740, 7, 6, 5, 4, 3, 2, 1, 0); \
62550   bfloat16x4_t __rev2_740;  __rev2_740 = __builtin_shufflevector(__s2_740, __s2_740, 3, 2, 1, 0); \
62551   __ret_740 = __noswap_vsetq_lane_bf16(__noswap_vget_lane_bf16(__rev2_740, __p3_740), __rev0_740, __p1_740); \
62552   __ret_740 = __builtin_shufflevector(__ret_740, __ret_740, 7, 6, 5, 4, 3, 2, 1, 0); \
62553   __ret_740; \
62554 })
62555 #endif
62556 
62557 #ifdef __LITTLE_ENDIAN__
62558 #define vcopy_lane_bf16(__p0_741, __p1_741, __p2_741, __p3_741) __extension__ ({ \
62559   bfloat16x4_t __ret_741; \
62560   bfloat16x4_t __s0_741 = __p0_741; \
62561   bfloat16x4_t __s2_741 = __p2_741; \
62562   __ret_741 = vset_lane_bf16(vget_lane_bf16(__s2_741, __p3_741), __s0_741, __p1_741); \
62563   __ret_741; \
62564 })
62565 #else
62566 #define vcopy_lane_bf16(__p0_742, __p1_742, __p2_742, __p3_742) __extension__ ({ \
62567   bfloat16x4_t __ret_742; \
62568   bfloat16x4_t __s0_742 = __p0_742; \
62569   bfloat16x4_t __s2_742 = __p2_742; \
62570   bfloat16x4_t __rev0_742;  __rev0_742 = __builtin_shufflevector(__s0_742, __s0_742, 3, 2, 1, 0); \
62571   bfloat16x4_t __rev2_742;  __rev2_742 = __builtin_shufflevector(__s2_742, __s2_742, 3, 2, 1, 0); \
62572   __ret_742 = __noswap_vset_lane_bf16(__noswap_vget_lane_bf16(__rev2_742, __p3_742), __rev0_742, __p1_742); \
62573   __ret_742 = __builtin_shufflevector(__ret_742, __ret_742, 3, 2, 1, 0); \
62574   __ret_742; \
62575 })
62576 #endif
62577 
62578 #ifdef __LITTLE_ENDIAN__
62579 #define vcopyq_laneq_bf16(__p0_743, __p1_743, __p2_743, __p3_743) __extension__ ({ \
62580   bfloat16x8_t __ret_743; \
62581   bfloat16x8_t __s0_743 = __p0_743; \
62582   bfloat16x8_t __s2_743 = __p2_743; \
62583   __ret_743 = vsetq_lane_bf16(vgetq_lane_bf16(__s2_743, __p3_743), __s0_743, __p1_743); \
62584   __ret_743; \
62585 })
62586 #else
62587 #define vcopyq_laneq_bf16(__p0_744, __p1_744, __p2_744, __p3_744) __extension__ ({ \
62588   bfloat16x8_t __ret_744; \
62589   bfloat16x8_t __s0_744 = __p0_744; \
62590   bfloat16x8_t __s2_744 = __p2_744; \
62591   bfloat16x8_t __rev0_744;  __rev0_744 = __builtin_shufflevector(__s0_744, __s0_744, 7, 6, 5, 4, 3, 2, 1, 0); \
62592   bfloat16x8_t __rev2_744;  __rev2_744 = __builtin_shufflevector(__s2_744, __s2_744, 7, 6, 5, 4, 3, 2, 1, 0); \
62593   __ret_744 = __noswap_vsetq_lane_bf16(__noswap_vgetq_lane_bf16(__rev2_744, __p3_744), __rev0_744, __p1_744); \
62594   __ret_744 = __builtin_shufflevector(__ret_744, __ret_744, 7, 6, 5, 4, 3, 2, 1, 0); \
62595   __ret_744; \
62596 })
62597 #endif
62598 
62599 #ifdef __LITTLE_ENDIAN__
62600 #define vcopy_laneq_bf16(__p0_745, __p1_745, __p2_745, __p3_745) __extension__ ({ \
62601   bfloat16x4_t __ret_745; \
62602   bfloat16x4_t __s0_745 = __p0_745; \
62603   bfloat16x8_t __s2_745 = __p2_745; \
62604   __ret_745 = vset_lane_bf16(vgetq_lane_bf16(__s2_745, __p3_745), __s0_745, __p1_745); \
62605   __ret_745; \
62606 })
62607 #else
62608 #define vcopy_laneq_bf16(__p0_746, __p1_746, __p2_746, __p3_746) __extension__ ({ \
62609   bfloat16x4_t __ret_746; \
62610   bfloat16x4_t __s0_746 = __p0_746; \
62611   bfloat16x8_t __s2_746 = __p2_746; \
62612   bfloat16x4_t __rev0_746;  __rev0_746 = __builtin_shufflevector(__s0_746, __s0_746, 3, 2, 1, 0); \
62613   bfloat16x8_t __rev2_746;  __rev2_746 = __builtin_shufflevector(__s2_746, __s2_746, 7, 6, 5, 4, 3, 2, 1, 0); \
62614   __ret_746 = __noswap_vset_lane_bf16(__noswap_vgetq_lane_bf16(__rev2_746, __p3_746), __rev0_746, __p1_746); \
62615   __ret_746 = __builtin_shufflevector(__ret_746, __ret_746, 3, 2, 1, 0); \
62616   __ret_746; \
62617 })
62618 #endif
62619 
62620 #ifdef __LITTLE_ENDIAN__
vcvt_bf16_f32(float32x4_t __p0)62621 __ai __attribute__((target("bf16"))) bfloat16x4_t vcvt_bf16_f32(float32x4_t __p0) {
62622   bfloat16x4_t __ret;
62623   __ret = vget_low_bf16(__a64_vcvtq_low_bf16_f32(__p0));
62624   return __ret;
62625 }
62626 #else
vcvt_bf16_f32(float32x4_t __p0)62627 __ai __attribute__((target("bf16"))) bfloat16x4_t vcvt_bf16_f32(float32x4_t __p0) {
62628   bfloat16x4_t __ret;
62629   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62630   __ret = __noswap_vget_low_bf16(__noswap___a64_vcvtq_low_bf16_f32(__rev0));
62631   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62632   return __ret;
62633 }
62634 #endif
62635 
62636 #ifdef __LITTLE_ENDIAN__
vcvtq_high_bf16_f32(bfloat16x8_t __p0,float32x4_t __p1)62637 __ai __attribute__((target("bf16"))) bfloat16x8_t vcvtq_high_bf16_f32(bfloat16x8_t __p0, float32x4_t __p1) {
62638   bfloat16x8_t __ret;
62639   __ret = (bfloat16x8_t) __builtin_neon_vcvtq_high_bf16_f32((int8x16_t)__p0, (int8x16_t)__p1, 43);
62640   return __ret;
62641 }
62642 #else
vcvtq_high_bf16_f32(bfloat16x8_t __p0,float32x4_t __p1)62643 __ai __attribute__((target("bf16"))) bfloat16x8_t vcvtq_high_bf16_f32(bfloat16x8_t __p0, float32x4_t __p1) {
62644   bfloat16x8_t __ret;
62645   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62646   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62647   __ret = (bfloat16x8_t) __builtin_neon_vcvtq_high_bf16_f32((int8x16_t)__rev0, (int8x16_t)__rev1, 43);
62648   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62649   return __ret;
62650 }
62651 #endif
62652 
62653 #ifdef __LITTLE_ENDIAN__
vcvtq_low_bf16_f32(float32x4_t __p0)62654 __ai __attribute__((target("bf16"))) bfloat16x8_t vcvtq_low_bf16_f32(float32x4_t __p0) {
62655   bfloat16x8_t __ret;
62656   __ret = __a64_vcvtq_low_bf16_f32(__p0);
62657   return __ret;
62658 }
62659 #else
vcvtq_low_bf16_f32(float32x4_t __p0)62660 __ai __attribute__((target("bf16"))) bfloat16x8_t vcvtq_low_bf16_f32(float32x4_t __p0) {
62661   bfloat16x8_t __ret;
62662   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62663   __ret = __noswap___a64_vcvtq_low_bf16_f32(__rev0);
62664   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62665   return __ret;
62666 }
62667 #endif
62668 
vreinterpret_p8_bf16(bfloat16x4_t __p0)62669 __ai __attribute__((target("bf16"))) poly8x8_t vreinterpret_p8_bf16(bfloat16x4_t __p0) {
62670   poly8x8_t __ret;
62671   __ret = (poly8x8_t)(__p0);
62672   return __ret;
62673 }
vreinterpret_p64_bf16(bfloat16x4_t __p0)62674 __ai __attribute__((target("bf16"))) poly64x1_t vreinterpret_p64_bf16(bfloat16x4_t __p0) {
62675   poly64x1_t __ret;
62676   __ret = (poly64x1_t)(__p0);
62677   return __ret;
62678 }
vreinterpret_p16_bf16(bfloat16x4_t __p0)62679 __ai __attribute__((target("bf16"))) poly16x4_t vreinterpret_p16_bf16(bfloat16x4_t __p0) {
62680   poly16x4_t __ret;
62681   __ret = (poly16x4_t)(__p0);
62682   return __ret;
62683 }
vreinterpretq_p8_bf16(bfloat16x8_t __p0)62684 __ai __attribute__((target("bf16"))) poly8x16_t vreinterpretq_p8_bf16(bfloat16x8_t __p0) {
62685   poly8x16_t __ret;
62686   __ret = (poly8x16_t)(__p0);
62687   return __ret;
62688 }
vreinterpretq_p128_bf16(bfloat16x8_t __p0)62689 __ai __attribute__((target("bf16"))) poly128_t vreinterpretq_p128_bf16(bfloat16x8_t __p0) {
62690   poly128_t __ret;
62691   __ret = (poly128_t)(__p0);
62692   return __ret;
62693 }
vreinterpretq_p64_bf16(bfloat16x8_t __p0)62694 __ai __attribute__((target("bf16"))) poly64x2_t vreinterpretq_p64_bf16(bfloat16x8_t __p0) {
62695   poly64x2_t __ret;
62696   __ret = (poly64x2_t)(__p0);
62697   return __ret;
62698 }
vreinterpretq_p16_bf16(bfloat16x8_t __p0)62699 __ai __attribute__((target("bf16"))) poly16x8_t vreinterpretq_p16_bf16(bfloat16x8_t __p0) {
62700   poly16x8_t __ret;
62701   __ret = (poly16x8_t)(__p0);
62702   return __ret;
62703 }
vreinterpretq_u8_bf16(bfloat16x8_t __p0)62704 __ai __attribute__((target("bf16"))) uint8x16_t vreinterpretq_u8_bf16(bfloat16x8_t __p0) {
62705   uint8x16_t __ret;
62706   __ret = (uint8x16_t)(__p0);
62707   return __ret;
62708 }
vreinterpretq_u32_bf16(bfloat16x8_t __p0)62709 __ai __attribute__((target("bf16"))) uint32x4_t vreinterpretq_u32_bf16(bfloat16x8_t __p0) {
62710   uint32x4_t __ret;
62711   __ret = (uint32x4_t)(__p0);
62712   return __ret;
62713 }
vreinterpretq_u64_bf16(bfloat16x8_t __p0)62714 __ai __attribute__((target("bf16"))) uint64x2_t vreinterpretq_u64_bf16(bfloat16x8_t __p0) {
62715   uint64x2_t __ret;
62716   __ret = (uint64x2_t)(__p0);
62717   return __ret;
62718 }
vreinterpretq_u16_bf16(bfloat16x8_t __p0)62719 __ai __attribute__((target("bf16"))) uint16x8_t vreinterpretq_u16_bf16(bfloat16x8_t __p0) {
62720   uint16x8_t __ret;
62721   __ret = (uint16x8_t)(__p0);
62722   return __ret;
62723 }
vreinterpretq_s8_bf16(bfloat16x8_t __p0)62724 __ai __attribute__((target("bf16"))) int8x16_t vreinterpretq_s8_bf16(bfloat16x8_t __p0) {
62725   int8x16_t __ret;
62726   __ret = (int8x16_t)(__p0);
62727   return __ret;
62728 }
vreinterpretq_f64_bf16(bfloat16x8_t __p0)62729 __ai __attribute__((target("bf16"))) float64x2_t vreinterpretq_f64_bf16(bfloat16x8_t __p0) {
62730   float64x2_t __ret;
62731   __ret = (float64x2_t)(__p0);
62732   return __ret;
62733 }
vreinterpretq_f32_bf16(bfloat16x8_t __p0)62734 __ai __attribute__((target("bf16"))) float32x4_t vreinterpretq_f32_bf16(bfloat16x8_t __p0) {
62735   float32x4_t __ret;
62736   __ret = (float32x4_t)(__p0);
62737   return __ret;
62738 }
vreinterpretq_f16_bf16(bfloat16x8_t __p0)62739 __ai __attribute__((target("bf16"))) float16x8_t vreinterpretq_f16_bf16(bfloat16x8_t __p0) {
62740   float16x8_t __ret;
62741   __ret = (float16x8_t)(__p0);
62742   return __ret;
62743 }
vreinterpretq_s32_bf16(bfloat16x8_t __p0)62744 __ai __attribute__((target("bf16"))) int32x4_t vreinterpretq_s32_bf16(bfloat16x8_t __p0) {
62745   int32x4_t __ret;
62746   __ret = (int32x4_t)(__p0);
62747   return __ret;
62748 }
vreinterpretq_s64_bf16(bfloat16x8_t __p0)62749 __ai __attribute__((target("bf16"))) int64x2_t vreinterpretq_s64_bf16(bfloat16x8_t __p0) {
62750   int64x2_t __ret;
62751   __ret = (int64x2_t)(__p0);
62752   return __ret;
62753 }
vreinterpretq_s16_bf16(bfloat16x8_t __p0)62754 __ai __attribute__((target("bf16"))) int16x8_t vreinterpretq_s16_bf16(bfloat16x8_t __p0) {
62755   int16x8_t __ret;
62756   __ret = (int16x8_t)(__p0);
62757   return __ret;
62758 }
vreinterpret_u8_bf16(bfloat16x4_t __p0)62759 __ai __attribute__((target("bf16"))) uint8x8_t vreinterpret_u8_bf16(bfloat16x4_t __p0) {
62760   uint8x8_t __ret;
62761   __ret = (uint8x8_t)(__p0);
62762   return __ret;
62763 }
vreinterpret_u32_bf16(bfloat16x4_t __p0)62764 __ai __attribute__((target("bf16"))) uint32x2_t vreinterpret_u32_bf16(bfloat16x4_t __p0) {
62765   uint32x2_t __ret;
62766   __ret = (uint32x2_t)(__p0);
62767   return __ret;
62768 }
vreinterpret_u64_bf16(bfloat16x4_t __p0)62769 __ai __attribute__((target("bf16"))) uint64x1_t vreinterpret_u64_bf16(bfloat16x4_t __p0) {
62770   uint64x1_t __ret;
62771   __ret = (uint64x1_t)(__p0);
62772   return __ret;
62773 }
vreinterpret_u16_bf16(bfloat16x4_t __p0)62774 __ai __attribute__((target("bf16"))) uint16x4_t vreinterpret_u16_bf16(bfloat16x4_t __p0) {
62775   uint16x4_t __ret;
62776   __ret = (uint16x4_t)(__p0);
62777   return __ret;
62778 }
vreinterpret_s8_bf16(bfloat16x4_t __p0)62779 __ai __attribute__((target("bf16"))) int8x8_t vreinterpret_s8_bf16(bfloat16x4_t __p0) {
62780   int8x8_t __ret;
62781   __ret = (int8x8_t)(__p0);
62782   return __ret;
62783 }
vreinterpret_f64_bf16(bfloat16x4_t __p0)62784 __ai __attribute__((target("bf16"))) float64x1_t vreinterpret_f64_bf16(bfloat16x4_t __p0) {
62785   float64x1_t __ret;
62786   __ret = (float64x1_t)(__p0);
62787   return __ret;
62788 }
vreinterpret_f32_bf16(bfloat16x4_t __p0)62789 __ai __attribute__((target("bf16"))) float32x2_t vreinterpret_f32_bf16(bfloat16x4_t __p0) {
62790   float32x2_t __ret;
62791   __ret = (float32x2_t)(__p0);
62792   return __ret;
62793 }
vreinterpret_f16_bf16(bfloat16x4_t __p0)62794 __ai __attribute__((target("bf16"))) float16x4_t vreinterpret_f16_bf16(bfloat16x4_t __p0) {
62795   float16x4_t __ret;
62796   __ret = (float16x4_t)(__p0);
62797   return __ret;
62798 }
vreinterpret_s32_bf16(bfloat16x4_t __p0)62799 __ai __attribute__((target("bf16"))) int32x2_t vreinterpret_s32_bf16(bfloat16x4_t __p0) {
62800   int32x2_t __ret;
62801   __ret = (int32x2_t)(__p0);
62802   return __ret;
62803 }
vreinterpret_s64_bf16(bfloat16x4_t __p0)62804 __ai __attribute__((target("bf16"))) int64x1_t vreinterpret_s64_bf16(bfloat16x4_t __p0) {
62805   int64x1_t __ret;
62806   __ret = (int64x1_t)(__p0);
62807   return __ret;
62808 }
vreinterpret_s16_bf16(bfloat16x4_t __p0)62809 __ai __attribute__((target("bf16"))) int16x4_t vreinterpret_s16_bf16(bfloat16x4_t __p0) {
62810   int16x4_t __ret;
62811   __ret = (int16x4_t)(__p0);
62812   return __ret;
62813 }
vreinterpretq_bf16_p8(poly8x16_t __p0)62814 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_p8(poly8x16_t __p0) {
62815   bfloat16x8_t __ret;
62816   __ret = (bfloat16x8_t)(__p0);
62817   return __ret;
62818 }
vreinterpretq_bf16_p128(poly128_t __p0)62819 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_p128(poly128_t __p0) {
62820   bfloat16x8_t __ret;
62821   __ret = (bfloat16x8_t)(__p0);
62822   return __ret;
62823 }
vreinterpretq_bf16_p64(poly64x2_t __p0)62824 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_p64(poly64x2_t __p0) {
62825   bfloat16x8_t __ret;
62826   __ret = (bfloat16x8_t)(__p0);
62827   return __ret;
62828 }
vreinterpretq_bf16_p16(poly16x8_t __p0)62829 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_p16(poly16x8_t __p0) {
62830   bfloat16x8_t __ret;
62831   __ret = (bfloat16x8_t)(__p0);
62832   return __ret;
62833 }
vreinterpretq_bf16_u8(uint8x16_t __p0)62834 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_u8(uint8x16_t __p0) {
62835   bfloat16x8_t __ret;
62836   __ret = (bfloat16x8_t)(__p0);
62837   return __ret;
62838 }
vreinterpretq_bf16_u32(uint32x4_t __p0)62839 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_u32(uint32x4_t __p0) {
62840   bfloat16x8_t __ret;
62841   __ret = (bfloat16x8_t)(__p0);
62842   return __ret;
62843 }
vreinterpretq_bf16_u64(uint64x2_t __p0)62844 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_u64(uint64x2_t __p0) {
62845   bfloat16x8_t __ret;
62846   __ret = (bfloat16x8_t)(__p0);
62847   return __ret;
62848 }
vreinterpretq_bf16_u16(uint16x8_t __p0)62849 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_u16(uint16x8_t __p0) {
62850   bfloat16x8_t __ret;
62851   __ret = (bfloat16x8_t)(__p0);
62852   return __ret;
62853 }
vreinterpretq_bf16_s8(int8x16_t __p0)62854 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_s8(int8x16_t __p0) {
62855   bfloat16x8_t __ret;
62856   __ret = (bfloat16x8_t)(__p0);
62857   return __ret;
62858 }
vreinterpretq_bf16_f64(float64x2_t __p0)62859 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_f64(float64x2_t __p0) {
62860   bfloat16x8_t __ret;
62861   __ret = (bfloat16x8_t)(__p0);
62862   return __ret;
62863 }
vreinterpretq_bf16_f32(float32x4_t __p0)62864 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_f32(float32x4_t __p0) {
62865   bfloat16x8_t __ret;
62866   __ret = (bfloat16x8_t)(__p0);
62867   return __ret;
62868 }
vreinterpretq_bf16_f16(float16x8_t __p0)62869 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_f16(float16x8_t __p0) {
62870   bfloat16x8_t __ret;
62871   __ret = (bfloat16x8_t)(__p0);
62872   return __ret;
62873 }
vreinterpretq_bf16_s32(int32x4_t __p0)62874 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_s32(int32x4_t __p0) {
62875   bfloat16x8_t __ret;
62876   __ret = (bfloat16x8_t)(__p0);
62877   return __ret;
62878 }
vreinterpretq_bf16_s64(int64x2_t __p0)62879 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_s64(int64x2_t __p0) {
62880   bfloat16x8_t __ret;
62881   __ret = (bfloat16x8_t)(__p0);
62882   return __ret;
62883 }
vreinterpretq_bf16_s16(int16x8_t __p0)62884 __ai __attribute__((target("bf16"))) bfloat16x8_t vreinterpretq_bf16_s16(int16x8_t __p0) {
62885   bfloat16x8_t __ret;
62886   __ret = (bfloat16x8_t)(__p0);
62887   return __ret;
62888 }
vreinterpret_bf16_p8(poly8x8_t __p0)62889 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_p8(poly8x8_t __p0) {
62890   bfloat16x4_t __ret;
62891   __ret = (bfloat16x4_t)(__p0);
62892   return __ret;
62893 }
vreinterpret_bf16_p64(poly64x1_t __p0)62894 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_p64(poly64x1_t __p0) {
62895   bfloat16x4_t __ret;
62896   __ret = (bfloat16x4_t)(__p0);
62897   return __ret;
62898 }
vreinterpret_bf16_p16(poly16x4_t __p0)62899 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_p16(poly16x4_t __p0) {
62900   bfloat16x4_t __ret;
62901   __ret = (bfloat16x4_t)(__p0);
62902   return __ret;
62903 }
vreinterpret_bf16_u8(uint8x8_t __p0)62904 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_u8(uint8x8_t __p0) {
62905   bfloat16x4_t __ret;
62906   __ret = (bfloat16x4_t)(__p0);
62907   return __ret;
62908 }
vreinterpret_bf16_u32(uint32x2_t __p0)62909 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_u32(uint32x2_t __p0) {
62910   bfloat16x4_t __ret;
62911   __ret = (bfloat16x4_t)(__p0);
62912   return __ret;
62913 }
vreinterpret_bf16_u64(uint64x1_t __p0)62914 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_u64(uint64x1_t __p0) {
62915   bfloat16x4_t __ret;
62916   __ret = (bfloat16x4_t)(__p0);
62917   return __ret;
62918 }
vreinterpret_bf16_u16(uint16x4_t __p0)62919 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_u16(uint16x4_t __p0) {
62920   bfloat16x4_t __ret;
62921   __ret = (bfloat16x4_t)(__p0);
62922   return __ret;
62923 }
vreinterpret_bf16_s8(int8x8_t __p0)62924 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_s8(int8x8_t __p0) {
62925   bfloat16x4_t __ret;
62926   __ret = (bfloat16x4_t)(__p0);
62927   return __ret;
62928 }
vreinterpret_bf16_f64(float64x1_t __p0)62929 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_f64(float64x1_t __p0) {
62930   bfloat16x4_t __ret;
62931   __ret = (bfloat16x4_t)(__p0);
62932   return __ret;
62933 }
vreinterpret_bf16_f32(float32x2_t __p0)62934 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_f32(float32x2_t __p0) {
62935   bfloat16x4_t __ret;
62936   __ret = (bfloat16x4_t)(__p0);
62937   return __ret;
62938 }
vreinterpret_bf16_f16(float16x4_t __p0)62939 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_f16(float16x4_t __p0) {
62940   bfloat16x4_t __ret;
62941   __ret = (bfloat16x4_t)(__p0);
62942   return __ret;
62943 }
vreinterpret_bf16_s32(int32x2_t __p0)62944 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_s32(int32x2_t __p0) {
62945   bfloat16x4_t __ret;
62946   __ret = (bfloat16x4_t)(__p0);
62947   return __ret;
62948 }
vreinterpret_bf16_s64(int64x1_t __p0)62949 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_s64(int64x1_t __p0) {
62950   bfloat16x4_t __ret;
62951   __ret = (bfloat16x4_t)(__p0);
62952   return __ret;
62953 }
vreinterpret_bf16_s16(int16x4_t __p0)62954 __ai __attribute__((target("bf16"))) bfloat16x4_t vreinterpret_bf16_s16(int16x4_t __p0) {
62955   bfloat16x4_t __ret;
62956   __ret = (bfloat16x4_t)(__p0);
62957   return __ret;
62958 }
62959 #ifdef __LITTLE_ENDIAN__
62960 #define vdotq_laneq_u32(__p0_747, __p1_747, __p2_747, __p3_747) __extension__ ({ \
62961   uint32x4_t __ret_747; \
62962   uint32x4_t __s0_747 = __p0_747; \
62963   uint8x16_t __s1_747 = __p1_747; \
62964   uint8x16_t __s2_747 = __p2_747; \
62965 uint8x16_t __reint_747 = __s2_747; \
62966 uint32x4_t __reint1_747 = splatq_laneq_u32(*(uint32x4_t *) &__reint_747, __p3_747); \
62967   __ret_747 = vdotq_u32(__s0_747, __s1_747, *(uint8x16_t *) &__reint1_747); \
62968   __ret_747; \
62969 })
62970 #else
62971 #define vdotq_laneq_u32(__p0_748, __p1_748, __p2_748, __p3_748) __extension__ ({ \
62972   uint32x4_t __ret_748; \
62973   uint32x4_t __s0_748 = __p0_748; \
62974   uint8x16_t __s1_748 = __p1_748; \
62975   uint8x16_t __s2_748 = __p2_748; \
62976   uint32x4_t __rev0_748;  __rev0_748 = __builtin_shufflevector(__s0_748, __s0_748, 3, 2, 1, 0); \
62977   uint8x16_t __rev1_748;  __rev1_748 = __builtin_shufflevector(__s1_748, __s1_748, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62978   uint8x16_t __rev2_748;  __rev2_748 = __builtin_shufflevector(__s2_748, __s2_748, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62979 uint8x16_t __reint_748 = __rev2_748; \
62980 uint32x4_t __reint1_748 = __noswap_splatq_laneq_u32(*(uint32x4_t *) &__reint_748, __p3_748); \
62981   __ret_748 = __noswap_vdotq_u32(__rev0_748, __rev1_748, *(uint8x16_t *) &__reint1_748); \
62982   __ret_748 = __builtin_shufflevector(__ret_748, __ret_748, 3, 2, 1, 0); \
62983   __ret_748; \
62984 })
62985 #endif
62986 
62987 #ifdef __LITTLE_ENDIAN__
62988 #define vdotq_laneq_s32(__p0_749, __p1_749, __p2_749, __p3_749) __extension__ ({ \
62989   int32x4_t __ret_749; \
62990   int32x4_t __s0_749 = __p0_749; \
62991   int8x16_t __s1_749 = __p1_749; \
62992   int8x16_t __s2_749 = __p2_749; \
62993 int8x16_t __reint_749 = __s2_749; \
62994 int32x4_t __reint1_749 = splatq_laneq_s32(*(int32x4_t *) &__reint_749, __p3_749); \
62995   __ret_749 = vdotq_s32(__s0_749, __s1_749, *(int8x16_t *) &__reint1_749); \
62996   __ret_749; \
62997 })
62998 #else
62999 #define vdotq_laneq_s32(__p0_750, __p1_750, __p2_750, __p3_750) __extension__ ({ \
63000   int32x4_t __ret_750; \
63001   int32x4_t __s0_750 = __p0_750; \
63002   int8x16_t __s1_750 = __p1_750; \
63003   int8x16_t __s2_750 = __p2_750; \
63004   int32x4_t __rev0_750;  __rev0_750 = __builtin_shufflevector(__s0_750, __s0_750, 3, 2, 1, 0); \
63005   int8x16_t __rev1_750;  __rev1_750 = __builtin_shufflevector(__s1_750, __s1_750, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63006   int8x16_t __rev2_750;  __rev2_750 = __builtin_shufflevector(__s2_750, __s2_750, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63007 int8x16_t __reint_750 = __rev2_750; \
63008 int32x4_t __reint1_750 = __noswap_splatq_laneq_s32(*(int32x4_t *) &__reint_750, __p3_750); \
63009   __ret_750 = __noswap_vdotq_s32(__rev0_750, __rev1_750, *(int8x16_t *) &__reint1_750); \
63010   __ret_750 = __builtin_shufflevector(__ret_750, __ret_750, 3, 2, 1, 0); \
63011   __ret_750; \
63012 })
63013 #endif
63014 
63015 #ifdef __LITTLE_ENDIAN__
63016 #define vdot_laneq_u32(__p0_751, __p1_751, __p2_751, __p3_751) __extension__ ({ \
63017   uint32x2_t __ret_751; \
63018   uint32x2_t __s0_751 = __p0_751; \
63019   uint8x8_t __s1_751 = __p1_751; \
63020   uint8x16_t __s2_751 = __p2_751; \
63021 uint8x16_t __reint_751 = __s2_751; \
63022 uint32x2_t __reint1_751 = splat_laneq_u32(*(uint32x4_t *) &__reint_751, __p3_751); \
63023   __ret_751 = vdot_u32(__s0_751, __s1_751, *(uint8x8_t *) &__reint1_751); \
63024   __ret_751; \
63025 })
63026 #else
63027 #define vdot_laneq_u32(__p0_752, __p1_752, __p2_752, __p3_752) __extension__ ({ \
63028   uint32x2_t __ret_752; \
63029   uint32x2_t __s0_752 = __p0_752; \
63030   uint8x8_t __s1_752 = __p1_752; \
63031   uint8x16_t __s2_752 = __p2_752; \
63032   uint32x2_t __rev0_752;  __rev0_752 = __builtin_shufflevector(__s0_752, __s0_752, 1, 0); \
63033   uint8x8_t __rev1_752;  __rev1_752 = __builtin_shufflevector(__s1_752, __s1_752, 7, 6, 5, 4, 3, 2, 1, 0); \
63034   uint8x16_t __rev2_752;  __rev2_752 = __builtin_shufflevector(__s2_752, __s2_752, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63035 uint8x16_t __reint_752 = __rev2_752; \
63036 uint32x2_t __reint1_752 = __noswap_splat_laneq_u32(*(uint32x4_t *) &__reint_752, __p3_752); \
63037   __ret_752 = __noswap_vdot_u32(__rev0_752, __rev1_752, *(uint8x8_t *) &__reint1_752); \
63038   __ret_752 = __builtin_shufflevector(__ret_752, __ret_752, 1, 0); \
63039   __ret_752; \
63040 })
63041 #endif
63042 
63043 #ifdef __LITTLE_ENDIAN__
63044 #define vdot_laneq_s32(__p0_753, __p1_753, __p2_753, __p3_753) __extension__ ({ \
63045   int32x2_t __ret_753; \
63046   int32x2_t __s0_753 = __p0_753; \
63047   int8x8_t __s1_753 = __p1_753; \
63048   int8x16_t __s2_753 = __p2_753; \
63049 int8x16_t __reint_753 = __s2_753; \
63050 int32x2_t __reint1_753 = splat_laneq_s32(*(int32x4_t *) &__reint_753, __p3_753); \
63051   __ret_753 = vdot_s32(__s0_753, __s1_753, *(int8x8_t *) &__reint1_753); \
63052   __ret_753; \
63053 })
63054 #else
63055 #define vdot_laneq_s32(__p0_754, __p1_754, __p2_754, __p3_754) __extension__ ({ \
63056   int32x2_t __ret_754; \
63057   int32x2_t __s0_754 = __p0_754; \
63058   int8x8_t __s1_754 = __p1_754; \
63059   int8x16_t __s2_754 = __p2_754; \
63060   int32x2_t __rev0_754;  __rev0_754 = __builtin_shufflevector(__s0_754, __s0_754, 1, 0); \
63061   int8x8_t __rev1_754;  __rev1_754 = __builtin_shufflevector(__s1_754, __s1_754, 7, 6, 5, 4, 3, 2, 1, 0); \
63062   int8x16_t __rev2_754;  __rev2_754 = __builtin_shufflevector(__s2_754, __s2_754, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63063 int8x16_t __reint_754 = __rev2_754; \
63064 int32x2_t __reint1_754 = __noswap_splat_laneq_s32(*(int32x4_t *) &__reint_754, __p3_754); \
63065   __ret_754 = __noswap_vdot_s32(__rev0_754, __rev1_754, *(int8x8_t *) &__reint1_754); \
63066   __ret_754 = __builtin_shufflevector(__ret_754, __ret_754, 1, 0); \
63067   __ret_754; \
63068 })
63069 #endif
63070 
63071 #ifdef __LITTLE_ENDIAN__
vfmlalq_high_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63072 __ai __attribute__((target("fp16fml"))) float32x4_t vfmlalq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63073   float32x4_t __ret;
63074   __ret = (float32x4_t) __builtin_neon_vfmlalq_high_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
63075   return __ret;
63076 }
63077 #else
vfmlalq_high_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63078 __ai __attribute__((target("fp16fml"))) float32x4_t vfmlalq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63079   float32x4_t __ret;
63080   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63081   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63082   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63083   __ret = (float32x4_t) __builtin_neon_vfmlalq_high_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
63084   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63085   return __ret;
63086 }
__noswap_vfmlalq_high_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63087 __ai __attribute__((target("fp16fml"))) float32x4_t __noswap_vfmlalq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63088   float32x4_t __ret;
63089   __ret = (float32x4_t) __builtin_neon_vfmlalq_high_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
63090   return __ret;
63091 }
63092 #endif
63093 
63094 #ifdef __LITTLE_ENDIAN__
vfmlal_high_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63095 __ai __attribute__((target("fp16fml"))) float32x2_t vfmlal_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63096   float32x2_t __ret;
63097   __ret = (float32x2_t) __builtin_neon_vfmlal_high_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
63098   return __ret;
63099 }
63100 #else
vfmlal_high_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63101 __ai __attribute__((target("fp16fml"))) float32x2_t vfmlal_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63102   float32x2_t __ret;
63103   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63104   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63105   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63106   __ret = (float32x2_t) __builtin_neon_vfmlal_high_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
63107   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63108   return __ret;
63109 }
__noswap_vfmlal_high_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63110 __ai __attribute__((target("fp16fml"))) float32x2_t __noswap_vfmlal_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63111   float32x2_t __ret;
63112   __ret = (float32x2_t) __builtin_neon_vfmlal_high_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
63113   return __ret;
63114 }
63115 #endif
63116 
63117 #ifdef __LITTLE_ENDIAN__
vfmlalq_low_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63118 __ai __attribute__((target("fp16fml"))) float32x4_t vfmlalq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63119   float32x4_t __ret;
63120   __ret = (float32x4_t) __builtin_neon_vfmlalq_low_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
63121   return __ret;
63122 }
63123 #else
vfmlalq_low_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63124 __ai __attribute__((target("fp16fml"))) float32x4_t vfmlalq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63125   float32x4_t __ret;
63126   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63127   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63128   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63129   __ret = (float32x4_t) __builtin_neon_vfmlalq_low_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
63130   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63131   return __ret;
63132 }
__noswap_vfmlalq_low_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63133 __ai __attribute__((target("fp16fml"))) float32x4_t __noswap_vfmlalq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63134   float32x4_t __ret;
63135   __ret = (float32x4_t) __builtin_neon_vfmlalq_low_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
63136   return __ret;
63137 }
63138 #endif
63139 
63140 #ifdef __LITTLE_ENDIAN__
vfmlal_low_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63141 __ai __attribute__((target("fp16fml"))) float32x2_t vfmlal_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63142   float32x2_t __ret;
63143   __ret = (float32x2_t) __builtin_neon_vfmlal_low_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
63144   return __ret;
63145 }
63146 #else
vfmlal_low_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63147 __ai __attribute__((target("fp16fml"))) float32x2_t vfmlal_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63148   float32x2_t __ret;
63149   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63150   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63151   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63152   __ret = (float32x2_t) __builtin_neon_vfmlal_low_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
63153   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63154   return __ret;
63155 }
__noswap_vfmlal_low_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63156 __ai __attribute__((target("fp16fml"))) float32x2_t __noswap_vfmlal_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63157   float32x2_t __ret;
63158   __ret = (float32x2_t) __builtin_neon_vfmlal_low_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
63159   return __ret;
63160 }
63161 #endif
63162 
63163 #ifdef __LITTLE_ENDIAN__
vfmlslq_high_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63164 __ai __attribute__((target("fp16fml"))) float32x4_t vfmlslq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63165   float32x4_t __ret;
63166   __ret = (float32x4_t) __builtin_neon_vfmlslq_high_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
63167   return __ret;
63168 }
63169 #else
vfmlslq_high_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63170 __ai __attribute__((target("fp16fml"))) float32x4_t vfmlslq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63171   float32x4_t __ret;
63172   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63173   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63174   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63175   __ret = (float32x4_t) __builtin_neon_vfmlslq_high_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
63176   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63177   return __ret;
63178 }
__noswap_vfmlslq_high_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63179 __ai __attribute__((target("fp16fml"))) float32x4_t __noswap_vfmlslq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63180   float32x4_t __ret;
63181   __ret = (float32x4_t) __builtin_neon_vfmlslq_high_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
63182   return __ret;
63183 }
63184 #endif
63185 
63186 #ifdef __LITTLE_ENDIAN__
vfmlsl_high_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63187 __ai __attribute__((target("fp16fml"))) float32x2_t vfmlsl_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63188   float32x2_t __ret;
63189   __ret = (float32x2_t) __builtin_neon_vfmlsl_high_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
63190   return __ret;
63191 }
63192 #else
vfmlsl_high_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63193 __ai __attribute__((target("fp16fml"))) float32x2_t vfmlsl_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63194   float32x2_t __ret;
63195   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63196   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63197   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63198   __ret = (float32x2_t) __builtin_neon_vfmlsl_high_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
63199   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63200   return __ret;
63201 }
__noswap_vfmlsl_high_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63202 __ai __attribute__((target("fp16fml"))) float32x2_t __noswap_vfmlsl_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63203   float32x2_t __ret;
63204   __ret = (float32x2_t) __builtin_neon_vfmlsl_high_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
63205   return __ret;
63206 }
63207 #endif
63208 
63209 #ifdef __LITTLE_ENDIAN__
vfmlslq_low_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63210 __ai __attribute__((target("fp16fml"))) float32x4_t vfmlslq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63211   float32x4_t __ret;
63212   __ret = (float32x4_t) __builtin_neon_vfmlslq_low_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
63213   return __ret;
63214 }
63215 #else
vfmlslq_low_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63216 __ai __attribute__((target("fp16fml"))) float32x4_t vfmlslq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63217   float32x4_t __ret;
63218   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63219   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63220   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63221   __ret = (float32x4_t) __builtin_neon_vfmlslq_low_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
63222   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63223   return __ret;
63224 }
__noswap_vfmlslq_low_f16(float32x4_t __p0,float16x8_t __p1,float16x8_t __p2)63225 __ai __attribute__((target("fp16fml"))) float32x4_t __noswap_vfmlslq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
63226   float32x4_t __ret;
63227   __ret = (float32x4_t) __builtin_neon_vfmlslq_low_f16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
63228   return __ret;
63229 }
63230 #endif
63231 
63232 #ifdef __LITTLE_ENDIAN__
vfmlsl_low_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63233 __ai __attribute__((target("fp16fml"))) float32x2_t vfmlsl_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63234   float32x2_t __ret;
63235   __ret = (float32x2_t) __builtin_neon_vfmlsl_low_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
63236   return __ret;
63237 }
63238 #else
vfmlsl_low_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63239 __ai __attribute__((target("fp16fml"))) float32x2_t vfmlsl_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63240   float32x2_t __ret;
63241   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63242   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63243   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63244   __ret = (float32x2_t) __builtin_neon_vfmlsl_low_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
63245   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63246   return __ret;
63247 }
__noswap_vfmlsl_low_f16(float32x2_t __p0,float16x4_t __p1,float16x4_t __p2)63248 __ai __attribute__((target("fp16fml"))) float32x2_t __noswap_vfmlsl_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
63249   float32x2_t __ret;
63250   __ret = (float32x2_t) __builtin_neon_vfmlsl_low_f16((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
63251   return __ret;
63252 }
63253 #endif
63254 
63255 #ifdef __LITTLE_ENDIAN__
vdivq_f16(float16x8_t __p0,float16x8_t __p1)63256 __ai __attribute__((target("fullfp16"))) float16x8_t vdivq_f16(float16x8_t __p0, float16x8_t __p1) {
63257   float16x8_t __ret;
63258   __ret = __p0 / __p1;
63259   return __ret;
63260 }
63261 #else
vdivq_f16(float16x8_t __p0,float16x8_t __p1)63262 __ai __attribute__((target("fullfp16"))) float16x8_t vdivq_f16(float16x8_t __p0, float16x8_t __p1) {
63263   float16x8_t __ret;
63264   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63265   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63266   __ret = __rev0 / __rev1;
63267   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63268   return __ret;
63269 }
63270 #endif
63271 
63272 #ifdef __LITTLE_ENDIAN__
vdiv_f16(float16x4_t __p0,float16x4_t __p1)63273 __ai __attribute__((target("fullfp16"))) float16x4_t vdiv_f16(float16x4_t __p0, float16x4_t __p1) {
63274   float16x4_t __ret;
63275   __ret = __p0 / __p1;
63276   return __ret;
63277 }
63278 #else
vdiv_f16(float16x4_t __p0,float16x4_t __p1)63279 __ai __attribute__((target("fullfp16"))) float16x4_t vdiv_f16(float16x4_t __p0, float16x4_t __p1) {
63280   float16x4_t __ret;
63281   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63282   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63283   __ret = __rev0 / __rev1;
63284   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63285   return __ret;
63286 }
63287 #endif
63288 
63289 #ifdef __LITTLE_ENDIAN__
63290 #define vduph_lane_f16(__p0, __p1) __extension__ ({ \
63291   float16_t __ret; \
63292   float16x4_t __s0 = __p0; \
63293   __ret = (float16_t) __builtin_neon_vduph_lane_f16((float16x4_t)__s0, __p1); \
63294   __ret; \
63295 })
63296 #else
63297 #define vduph_lane_f16(__p0, __p1) __extension__ ({ \
63298   float16_t __ret; \
63299   float16x4_t __s0 = __p0; \
63300   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
63301   __ret = (float16_t) __builtin_neon_vduph_lane_f16((float16x4_t)__rev0, __p1); \
63302   __ret; \
63303 })
63304 #endif
63305 
63306 #ifdef __LITTLE_ENDIAN__
63307 #define vduph_laneq_f16(__p0, __p1) __extension__ ({ \
63308   float16_t __ret; \
63309   float16x8_t __s0 = __p0; \
63310   __ret = (float16_t) __builtin_neon_vduph_laneq_f16((float16x8_t)__s0, __p1); \
63311   __ret; \
63312 })
63313 #else
63314 #define vduph_laneq_f16(__p0, __p1) __extension__ ({ \
63315   float16_t __ret; \
63316   float16x8_t __s0 = __p0; \
63317   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
63318   __ret = (float16_t) __builtin_neon_vduph_laneq_f16((float16x8_t)__rev0, __p1); \
63319   __ret; \
63320 })
63321 #endif
63322 
63323 #ifdef __LITTLE_ENDIAN__
63324 #define vfmah_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63325   float16_t __ret; \
63326   float16_t __s0 = __p0; \
63327   float16_t __s1 = __p1; \
63328   float16x4_t __s2 = __p2; \
63329   __ret = (float16_t) __builtin_neon_vfmah_lane_f16(__s0, __s1, (float16x4_t)__s2, __p3); \
63330   __ret; \
63331 })
63332 #else
63333 #define vfmah_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63334   float16_t __ret; \
63335   float16_t __s0 = __p0; \
63336   float16_t __s1 = __p1; \
63337   float16x4_t __s2 = __p2; \
63338   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
63339   __ret = (float16_t) __builtin_neon_vfmah_lane_f16(__s0, __s1, (float16x4_t)__rev2, __p3); \
63340   __ret; \
63341 })
63342 #define __noswap_vfmah_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63343   float16_t __ret; \
63344   float16_t __s0 = __p0; \
63345   float16_t __s1 = __p1; \
63346   float16x4_t __s2 = __p2; \
63347   __ret = (float16_t) __builtin_neon_vfmah_lane_f16(__s0, __s1, (float16x4_t)__s2, __p3); \
63348   __ret; \
63349 })
63350 #endif
63351 
63352 #ifdef __LITTLE_ENDIAN__
63353 #define vfmaq_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63354   float16x8_t __ret; \
63355   float16x8_t __s0 = __p0; \
63356   float16x8_t __s1 = __p1; \
63357   float16x4_t __s2 = __p2; \
63358   __ret = (float16x8_t) __builtin_neon_vfmaq_lane_f16((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 40); \
63359   __ret; \
63360 })
63361 #else
63362 #define vfmaq_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63363   float16x8_t __ret; \
63364   float16x8_t __s0 = __p0; \
63365   float16x8_t __s1 = __p1; \
63366   float16x4_t __s2 = __p2; \
63367   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
63368   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
63369   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
63370   __ret = (float16x8_t) __builtin_neon_vfmaq_lane_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, __p3, 40); \
63371   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
63372   __ret; \
63373 })
63374 #define __noswap_vfmaq_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63375   float16x8_t __ret; \
63376   float16x8_t __s0 = __p0; \
63377   float16x8_t __s1 = __p1; \
63378   float16x4_t __s2 = __p2; \
63379   __ret = (float16x8_t) __builtin_neon_vfmaq_lane_f16((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 40); \
63380   __ret; \
63381 })
63382 #endif
63383 
63384 #ifdef __LITTLE_ENDIAN__
63385 #define vfma_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63386   float16x4_t __ret; \
63387   float16x4_t __s0 = __p0; \
63388   float16x4_t __s1 = __p1; \
63389   float16x4_t __s2 = __p2; \
63390   __ret = (float16x4_t) __builtin_neon_vfma_lane_f16((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 8); \
63391   __ret; \
63392 })
63393 #else
63394 #define vfma_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63395   float16x4_t __ret; \
63396   float16x4_t __s0 = __p0; \
63397   float16x4_t __s1 = __p1; \
63398   float16x4_t __s2 = __p2; \
63399   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
63400   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
63401   float16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
63402   __ret = (float16x4_t) __builtin_neon_vfma_lane_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, __p3, 8); \
63403   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
63404   __ret; \
63405 })
63406 #define __noswap_vfma_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63407   float16x4_t __ret; \
63408   float16x4_t __s0 = __p0; \
63409   float16x4_t __s1 = __p1; \
63410   float16x4_t __s2 = __p2; \
63411   __ret = (float16x4_t) __builtin_neon_vfma_lane_f16((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 8); \
63412   __ret; \
63413 })
63414 #endif
63415 
63416 #ifdef __LITTLE_ENDIAN__
63417 #define vfmah_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63418   float16_t __ret; \
63419   float16_t __s0 = __p0; \
63420   float16_t __s1 = __p1; \
63421   float16x8_t __s2 = __p2; \
63422   __ret = (float16_t) __builtin_neon_vfmah_laneq_f16(__s0, __s1, (float16x8_t)__s2, __p3); \
63423   __ret; \
63424 })
63425 #else
63426 #define vfmah_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63427   float16_t __ret; \
63428   float16_t __s0 = __p0; \
63429   float16_t __s1 = __p1; \
63430   float16x8_t __s2 = __p2; \
63431   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
63432   __ret = (float16_t) __builtin_neon_vfmah_laneq_f16(__s0, __s1, (float16x8_t)__rev2, __p3); \
63433   __ret; \
63434 })
63435 #define __noswap_vfmah_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63436   float16_t __ret; \
63437   float16_t __s0 = __p0; \
63438   float16_t __s1 = __p1; \
63439   float16x8_t __s2 = __p2; \
63440   __ret = (float16_t) __builtin_neon_vfmah_laneq_f16(__s0, __s1, (float16x8_t)__s2, __p3); \
63441   __ret; \
63442 })
63443 #endif
63444 
63445 #ifdef __LITTLE_ENDIAN__
63446 #define vfmaq_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63447   float16x8_t __ret; \
63448   float16x8_t __s0 = __p0; \
63449   float16x8_t __s1 = __p1; \
63450   float16x8_t __s2 = __p2; \
63451   __ret = (float16x8_t) __builtin_neon_vfmaq_laneq_f16((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 40); \
63452   __ret; \
63453 })
63454 #else
63455 #define vfmaq_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63456   float16x8_t __ret; \
63457   float16x8_t __s0 = __p0; \
63458   float16x8_t __s1 = __p1; \
63459   float16x8_t __s2 = __p2; \
63460   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
63461   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
63462   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
63463   __ret = (float16x8_t) __builtin_neon_vfmaq_laneq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 40); \
63464   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
63465   __ret; \
63466 })
63467 #define __noswap_vfmaq_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63468   float16x8_t __ret; \
63469   float16x8_t __s0 = __p0; \
63470   float16x8_t __s1 = __p1; \
63471   float16x8_t __s2 = __p2; \
63472   __ret = (float16x8_t) __builtin_neon_vfmaq_laneq_f16((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 40); \
63473   __ret; \
63474 })
63475 #endif
63476 
63477 #ifdef __LITTLE_ENDIAN__
63478 #define vfma_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63479   float16x4_t __ret; \
63480   float16x4_t __s0 = __p0; \
63481   float16x4_t __s1 = __p1; \
63482   float16x8_t __s2 = __p2; \
63483   __ret = (float16x4_t) __builtin_neon_vfma_laneq_f16((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 8); \
63484   __ret; \
63485 })
63486 #else
63487 #define vfma_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63488   float16x4_t __ret; \
63489   float16x4_t __s0 = __p0; \
63490   float16x4_t __s1 = __p1; \
63491   float16x8_t __s2 = __p2; \
63492   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
63493   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
63494   float16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
63495   __ret = (float16x4_t) __builtin_neon_vfma_laneq_f16((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x16_t)__rev2, __p3, 8); \
63496   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
63497   __ret; \
63498 })
63499 #define __noswap_vfma_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
63500   float16x4_t __ret; \
63501   float16x4_t __s0 = __p0; \
63502   float16x4_t __s1 = __p1; \
63503   float16x8_t __s2 = __p2; \
63504   __ret = (float16x4_t) __builtin_neon_vfma_laneq_f16((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 8); \
63505   __ret; \
63506 })
63507 #endif
63508 
63509 #ifdef __LITTLE_ENDIAN__
63510 #define vfmaq_n_f16(__p0, __p1, __p2) __extension__ ({ \
63511   float16x8_t __ret; \
63512   float16x8_t __s0 = __p0; \
63513   float16x8_t __s1 = __p1; \
63514   float16_t __s2 = __p2; \
63515   __ret = vfmaq_f16(__s0, __s1, (float16x8_t) {__s2, __s2, __s2, __s2, __s2, __s2, __s2, __s2}); \
63516   __ret; \
63517 })
63518 #else
63519 #define vfmaq_n_f16(__p0, __p1, __p2) __extension__ ({ \
63520   float16x8_t __ret; \
63521   float16x8_t __s0 = __p0; \
63522   float16x8_t __s1 = __p1; \
63523   float16_t __s2 = __p2; \
63524   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
63525   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
63526   __ret = __noswap_vfmaq_f16(__rev0, __rev1, (float16x8_t) {__s2, __s2, __s2, __s2, __s2, __s2, __s2, __s2}); \
63527   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
63528   __ret; \
63529 })
63530 #endif
63531 
63532 #ifdef __LITTLE_ENDIAN__
63533 #define vfma_n_f16(__p0, __p1, __p2) __extension__ ({ \
63534   float16x4_t __ret; \
63535   float16x4_t __s0 = __p0; \
63536   float16x4_t __s1 = __p1; \
63537   float16_t __s2 = __p2; \
63538   __ret = vfma_f16(__s0, __s1, (float16x4_t) {__s2, __s2, __s2, __s2}); \
63539   __ret; \
63540 })
63541 #else
63542 #define vfma_n_f16(__p0, __p1, __p2) __extension__ ({ \
63543   float16x4_t __ret; \
63544   float16x4_t __s0 = __p0; \
63545   float16x4_t __s1 = __p1; \
63546   float16_t __s2 = __p2; \
63547   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
63548   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
63549   __ret = __noswap_vfma_f16(__rev0, __rev1, (float16x4_t) {__s2, __s2, __s2, __s2}); \
63550   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
63551   __ret; \
63552 })
63553 #endif
63554 
63555 #ifdef __LITTLE_ENDIAN__
63556 #define vfmsh_lane_f16(__p0_755, __p1_755, __p2_755, __p3_755) __extension__ ({ \
63557   float16_t __ret_755; \
63558   float16_t __s0_755 = __p0_755; \
63559   float16_t __s1_755 = __p1_755; \
63560   float16x4_t __s2_755 = __p2_755; \
63561   __ret_755 = vfmah_lane_f16(__s0_755, -__s1_755, __s2_755, __p3_755); \
63562   __ret_755; \
63563 })
63564 #else
63565 #define vfmsh_lane_f16(__p0_756, __p1_756, __p2_756, __p3_756) __extension__ ({ \
63566   float16_t __ret_756; \
63567   float16_t __s0_756 = __p0_756; \
63568   float16_t __s1_756 = __p1_756; \
63569   float16x4_t __s2_756 = __p2_756; \
63570   float16x4_t __rev2_756;  __rev2_756 = __builtin_shufflevector(__s2_756, __s2_756, 3, 2, 1, 0); \
63571   __ret_756 = __noswap_vfmah_lane_f16(__s0_756, -__s1_756, __rev2_756, __p3_756); \
63572   __ret_756; \
63573 })
63574 #endif
63575 
63576 #ifdef __LITTLE_ENDIAN__
63577 #define vfmsq_lane_f16(__p0_757, __p1_757, __p2_757, __p3_757) __extension__ ({ \
63578   float16x8_t __ret_757; \
63579   float16x8_t __s0_757 = __p0_757; \
63580   float16x8_t __s1_757 = __p1_757; \
63581   float16x4_t __s2_757 = __p2_757; \
63582   __ret_757 = vfmaq_lane_f16(__s0_757, -__s1_757, __s2_757, __p3_757); \
63583   __ret_757; \
63584 })
63585 #else
63586 #define vfmsq_lane_f16(__p0_758, __p1_758, __p2_758, __p3_758) __extension__ ({ \
63587   float16x8_t __ret_758; \
63588   float16x8_t __s0_758 = __p0_758; \
63589   float16x8_t __s1_758 = __p1_758; \
63590   float16x4_t __s2_758 = __p2_758; \
63591   float16x8_t __rev0_758;  __rev0_758 = __builtin_shufflevector(__s0_758, __s0_758, 7, 6, 5, 4, 3, 2, 1, 0); \
63592   float16x8_t __rev1_758;  __rev1_758 = __builtin_shufflevector(__s1_758, __s1_758, 7, 6, 5, 4, 3, 2, 1, 0); \
63593   float16x4_t __rev2_758;  __rev2_758 = __builtin_shufflevector(__s2_758, __s2_758, 3, 2, 1, 0); \
63594   __ret_758 = __noswap_vfmaq_lane_f16(__rev0_758, -__rev1_758, __rev2_758, __p3_758); \
63595   __ret_758 = __builtin_shufflevector(__ret_758, __ret_758, 7, 6, 5, 4, 3, 2, 1, 0); \
63596   __ret_758; \
63597 })
63598 #endif
63599 
63600 #ifdef __LITTLE_ENDIAN__
63601 #define vfms_lane_f16(__p0_759, __p1_759, __p2_759, __p3_759) __extension__ ({ \
63602   float16x4_t __ret_759; \
63603   float16x4_t __s0_759 = __p0_759; \
63604   float16x4_t __s1_759 = __p1_759; \
63605   float16x4_t __s2_759 = __p2_759; \
63606   __ret_759 = vfma_lane_f16(__s0_759, -__s1_759, __s2_759, __p3_759); \
63607   __ret_759; \
63608 })
63609 #else
63610 #define vfms_lane_f16(__p0_760, __p1_760, __p2_760, __p3_760) __extension__ ({ \
63611   float16x4_t __ret_760; \
63612   float16x4_t __s0_760 = __p0_760; \
63613   float16x4_t __s1_760 = __p1_760; \
63614   float16x4_t __s2_760 = __p2_760; \
63615   float16x4_t __rev0_760;  __rev0_760 = __builtin_shufflevector(__s0_760, __s0_760, 3, 2, 1, 0); \
63616   float16x4_t __rev1_760;  __rev1_760 = __builtin_shufflevector(__s1_760, __s1_760, 3, 2, 1, 0); \
63617   float16x4_t __rev2_760;  __rev2_760 = __builtin_shufflevector(__s2_760, __s2_760, 3, 2, 1, 0); \
63618   __ret_760 = __noswap_vfma_lane_f16(__rev0_760, -__rev1_760, __rev2_760, __p3_760); \
63619   __ret_760 = __builtin_shufflevector(__ret_760, __ret_760, 3, 2, 1, 0); \
63620   __ret_760; \
63621 })
63622 #endif
63623 
63624 #ifdef __LITTLE_ENDIAN__
63625 #define vfmsh_laneq_f16(__p0_761, __p1_761, __p2_761, __p3_761) __extension__ ({ \
63626   float16_t __ret_761; \
63627   float16_t __s0_761 = __p0_761; \
63628   float16_t __s1_761 = __p1_761; \
63629   float16x8_t __s2_761 = __p2_761; \
63630   __ret_761 = vfmah_laneq_f16(__s0_761, -__s1_761, __s2_761, __p3_761); \
63631   __ret_761; \
63632 })
63633 #else
63634 #define vfmsh_laneq_f16(__p0_762, __p1_762, __p2_762, __p3_762) __extension__ ({ \
63635   float16_t __ret_762; \
63636   float16_t __s0_762 = __p0_762; \
63637   float16_t __s1_762 = __p1_762; \
63638   float16x8_t __s2_762 = __p2_762; \
63639   float16x8_t __rev2_762;  __rev2_762 = __builtin_shufflevector(__s2_762, __s2_762, 7, 6, 5, 4, 3, 2, 1, 0); \
63640   __ret_762 = __noswap_vfmah_laneq_f16(__s0_762, -__s1_762, __rev2_762, __p3_762); \
63641   __ret_762; \
63642 })
63643 #endif
63644 
63645 #ifdef __LITTLE_ENDIAN__
63646 #define vfmsq_laneq_f16(__p0_763, __p1_763, __p2_763, __p3_763) __extension__ ({ \
63647   float16x8_t __ret_763; \
63648   float16x8_t __s0_763 = __p0_763; \
63649   float16x8_t __s1_763 = __p1_763; \
63650   float16x8_t __s2_763 = __p2_763; \
63651   __ret_763 = vfmaq_laneq_f16(__s0_763, -__s1_763, __s2_763, __p3_763); \
63652   __ret_763; \
63653 })
63654 #else
63655 #define vfmsq_laneq_f16(__p0_764, __p1_764, __p2_764, __p3_764) __extension__ ({ \
63656   float16x8_t __ret_764; \
63657   float16x8_t __s0_764 = __p0_764; \
63658   float16x8_t __s1_764 = __p1_764; \
63659   float16x8_t __s2_764 = __p2_764; \
63660   float16x8_t __rev0_764;  __rev0_764 = __builtin_shufflevector(__s0_764, __s0_764, 7, 6, 5, 4, 3, 2, 1, 0); \
63661   float16x8_t __rev1_764;  __rev1_764 = __builtin_shufflevector(__s1_764, __s1_764, 7, 6, 5, 4, 3, 2, 1, 0); \
63662   float16x8_t __rev2_764;  __rev2_764 = __builtin_shufflevector(__s2_764, __s2_764, 7, 6, 5, 4, 3, 2, 1, 0); \
63663   __ret_764 = __noswap_vfmaq_laneq_f16(__rev0_764, -__rev1_764, __rev2_764, __p3_764); \
63664   __ret_764 = __builtin_shufflevector(__ret_764, __ret_764, 7, 6, 5, 4, 3, 2, 1, 0); \
63665   __ret_764; \
63666 })
63667 #endif
63668 
63669 #ifdef __LITTLE_ENDIAN__
63670 #define vfms_laneq_f16(__p0_765, __p1_765, __p2_765, __p3_765) __extension__ ({ \
63671   float16x4_t __ret_765; \
63672   float16x4_t __s0_765 = __p0_765; \
63673   float16x4_t __s1_765 = __p1_765; \
63674   float16x8_t __s2_765 = __p2_765; \
63675   __ret_765 = vfma_laneq_f16(__s0_765, -__s1_765, __s2_765, __p3_765); \
63676   __ret_765; \
63677 })
63678 #else
63679 #define vfms_laneq_f16(__p0_766, __p1_766, __p2_766, __p3_766) __extension__ ({ \
63680   float16x4_t __ret_766; \
63681   float16x4_t __s0_766 = __p0_766; \
63682   float16x4_t __s1_766 = __p1_766; \
63683   float16x8_t __s2_766 = __p2_766; \
63684   float16x4_t __rev0_766;  __rev0_766 = __builtin_shufflevector(__s0_766, __s0_766, 3, 2, 1, 0); \
63685   float16x4_t __rev1_766;  __rev1_766 = __builtin_shufflevector(__s1_766, __s1_766, 3, 2, 1, 0); \
63686   float16x8_t __rev2_766;  __rev2_766 = __builtin_shufflevector(__s2_766, __s2_766, 7, 6, 5, 4, 3, 2, 1, 0); \
63687   __ret_766 = __noswap_vfma_laneq_f16(__rev0_766, -__rev1_766, __rev2_766, __p3_766); \
63688   __ret_766 = __builtin_shufflevector(__ret_766, __ret_766, 3, 2, 1, 0); \
63689   __ret_766; \
63690 })
63691 #endif
63692 
63693 #ifdef __LITTLE_ENDIAN__
63694 #define vfmsq_n_f16(__p0, __p1, __p2) __extension__ ({ \
63695   float16x8_t __ret; \
63696   float16x8_t __s0 = __p0; \
63697   float16x8_t __s1 = __p1; \
63698   float16_t __s2 = __p2; \
63699   __ret = vfmaq_f16(__s0, -__s1, (float16x8_t) {__s2, __s2, __s2, __s2, __s2, __s2, __s2, __s2}); \
63700   __ret; \
63701 })
63702 #else
63703 #define vfmsq_n_f16(__p0, __p1, __p2) __extension__ ({ \
63704   float16x8_t __ret; \
63705   float16x8_t __s0 = __p0; \
63706   float16x8_t __s1 = __p1; \
63707   float16_t __s2 = __p2; \
63708   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
63709   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
63710   __ret = __noswap_vfmaq_f16(__rev0, -__rev1, (float16x8_t) {__s2, __s2, __s2, __s2, __s2, __s2, __s2, __s2}); \
63711   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
63712   __ret; \
63713 })
63714 #endif
63715 
63716 #ifdef __LITTLE_ENDIAN__
63717 #define vfms_n_f16(__p0, __p1, __p2) __extension__ ({ \
63718   float16x4_t __ret; \
63719   float16x4_t __s0 = __p0; \
63720   float16x4_t __s1 = __p1; \
63721   float16_t __s2 = __p2; \
63722   __ret = vfma_f16(__s0, -__s1, (float16x4_t) {__s2, __s2, __s2, __s2}); \
63723   __ret; \
63724 })
63725 #else
63726 #define vfms_n_f16(__p0, __p1, __p2) __extension__ ({ \
63727   float16x4_t __ret; \
63728   float16x4_t __s0 = __p0; \
63729   float16x4_t __s1 = __p1; \
63730   float16_t __s2 = __p2; \
63731   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
63732   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
63733   __ret = __noswap_vfma_f16(__rev0, -__rev1, (float16x4_t) {__s2, __s2, __s2, __s2}); \
63734   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
63735   __ret; \
63736 })
63737 #endif
63738 
63739 #ifdef __LITTLE_ENDIAN__
63740 #define vmaxnmvq_f16(__p0) __extension__ ({ \
63741   float16_t __ret; \
63742   float16x8_t __s0 = __p0; \
63743   __ret = (float16_t) __builtin_neon_vmaxnmvq_f16((int8x16_t)__s0); \
63744   __ret; \
63745 })
63746 #else
63747 #define vmaxnmvq_f16(__p0) __extension__ ({ \
63748   float16_t __ret; \
63749   float16x8_t __s0 = __p0; \
63750   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
63751   __ret = (float16_t) __builtin_neon_vmaxnmvq_f16((int8x16_t)__rev0); \
63752   __ret; \
63753 })
63754 #endif
63755 
63756 #ifdef __LITTLE_ENDIAN__
63757 #define vmaxnmv_f16(__p0) __extension__ ({ \
63758   float16_t __ret; \
63759   float16x4_t __s0 = __p0; \
63760   __ret = (float16_t) __builtin_neon_vmaxnmv_f16((int8x8_t)__s0); \
63761   __ret; \
63762 })
63763 #else
63764 #define vmaxnmv_f16(__p0) __extension__ ({ \
63765   float16_t __ret; \
63766   float16x4_t __s0 = __p0; \
63767   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
63768   __ret = (float16_t) __builtin_neon_vmaxnmv_f16((int8x8_t)__rev0); \
63769   __ret; \
63770 })
63771 #endif
63772 
63773 #ifdef __LITTLE_ENDIAN__
63774 #define vmaxvq_f16(__p0) __extension__ ({ \
63775   float16_t __ret; \
63776   float16x8_t __s0 = __p0; \
63777   __ret = (float16_t) __builtin_neon_vmaxvq_f16((int8x16_t)__s0); \
63778   __ret; \
63779 })
63780 #else
63781 #define vmaxvq_f16(__p0) __extension__ ({ \
63782   float16_t __ret; \
63783   float16x8_t __s0 = __p0; \
63784   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
63785   __ret = (float16_t) __builtin_neon_vmaxvq_f16((int8x16_t)__rev0); \
63786   __ret; \
63787 })
63788 #endif
63789 
63790 #ifdef __LITTLE_ENDIAN__
63791 #define vmaxv_f16(__p0) __extension__ ({ \
63792   float16_t __ret; \
63793   float16x4_t __s0 = __p0; \
63794   __ret = (float16_t) __builtin_neon_vmaxv_f16((int8x8_t)__s0); \
63795   __ret; \
63796 })
63797 #else
63798 #define vmaxv_f16(__p0) __extension__ ({ \
63799   float16_t __ret; \
63800   float16x4_t __s0 = __p0; \
63801   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
63802   __ret = (float16_t) __builtin_neon_vmaxv_f16((int8x8_t)__rev0); \
63803   __ret; \
63804 })
63805 #endif
63806 
63807 #ifdef __LITTLE_ENDIAN__
63808 #define vminnmvq_f16(__p0) __extension__ ({ \
63809   float16_t __ret; \
63810   float16x8_t __s0 = __p0; \
63811   __ret = (float16_t) __builtin_neon_vminnmvq_f16((int8x16_t)__s0); \
63812   __ret; \
63813 })
63814 #else
63815 #define vminnmvq_f16(__p0) __extension__ ({ \
63816   float16_t __ret; \
63817   float16x8_t __s0 = __p0; \
63818   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
63819   __ret = (float16_t) __builtin_neon_vminnmvq_f16((int8x16_t)__rev0); \
63820   __ret; \
63821 })
63822 #endif
63823 
63824 #ifdef __LITTLE_ENDIAN__
63825 #define vminnmv_f16(__p0) __extension__ ({ \
63826   float16_t __ret; \
63827   float16x4_t __s0 = __p0; \
63828   __ret = (float16_t) __builtin_neon_vminnmv_f16((int8x8_t)__s0); \
63829   __ret; \
63830 })
63831 #else
63832 #define vminnmv_f16(__p0) __extension__ ({ \
63833   float16_t __ret; \
63834   float16x4_t __s0 = __p0; \
63835   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
63836   __ret = (float16_t) __builtin_neon_vminnmv_f16((int8x8_t)__rev0); \
63837   __ret; \
63838 })
63839 #endif
63840 
63841 #ifdef __LITTLE_ENDIAN__
63842 #define vminvq_f16(__p0) __extension__ ({ \
63843   float16_t __ret; \
63844   float16x8_t __s0 = __p0; \
63845   __ret = (float16_t) __builtin_neon_vminvq_f16((int8x16_t)__s0); \
63846   __ret; \
63847 })
63848 #else
63849 #define vminvq_f16(__p0) __extension__ ({ \
63850   float16_t __ret; \
63851   float16x8_t __s0 = __p0; \
63852   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
63853   __ret = (float16_t) __builtin_neon_vminvq_f16((int8x16_t)__rev0); \
63854   __ret; \
63855 })
63856 #endif
63857 
63858 #ifdef __LITTLE_ENDIAN__
63859 #define vminv_f16(__p0) __extension__ ({ \
63860   float16_t __ret; \
63861   float16x4_t __s0 = __p0; \
63862   __ret = (float16_t) __builtin_neon_vminv_f16((int8x8_t)__s0); \
63863   __ret; \
63864 })
63865 #else
63866 #define vminv_f16(__p0) __extension__ ({ \
63867   float16_t __ret; \
63868   float16x4_t __s0 = __p0; \
63869   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
63870   __ret = (float16_t) __builtin_neon_vminv_f16((int8x8_t)__rev0); \
63871   __ret; \
63872 })
63873 #endif
63874 
63875 #ifdef __LITTLE_ENDIAN__
63876 #define vmulq_laneq_f16(__p0_767, __p1_767, __p2_767) __extension__ ({ \
63877   float16x8_t __ret_767; \
63878   float16x8_t __s0_767 = __p0_767; \
63879   float16x8_t __s1_767 = __p1_767; \
63880   __ret_767 = __s0_767 * splatq_laneq_f16(__s1_767, __p2_767); \
63881   __ret_767; \
63882 })
63883 #else
63884 #define vmulq_laneq_f16(__p0_768, __p1_768, __p2_768) __extension__ ({ \
63885   float16x8_t __ret_768; \
63886   float16x8_t __s0_768 = __p0_768; \
63887   float16x8_t __s1_768 = __p1_768; \
63888   float16x8_t __rev0_768;  __rev0_768 = __builtin_shufflevector(__s0_768, __s0_768, 7, 6, 5, 4, 3, 2, 1, 0); \
63889   float16x8_t __rev1_768;  __rev1_768 = __builtin_shufflevector(__s1_768, __s1_768, 7, 6, 5, 4, 3, 2, 1, 0); \
63890   __ret_768 = __rev0_768 * __noswap_splatq_laneq_f16(__rev1_768, __p2_768); \
63891   __ret_768 = __builtin_shufflevector(__ret_768, __ret_768, 7, 6, 5, 4, 3, 2, 1, 0); \
63892   __ret_768; \
63893 })
63894 #endif
63895 
63896 #ifdef __LITTLE_ENDIAN__
63897 #define vmul_laneq_f16(__p0_769, __p1_769, __p2_769) __extension__ ({ \
63898   float16x4_t __ret_769; \
63899   float16x4_t __s0_769 = __p0_769; \
63900   float16x8_t __s1_769 = __p1_769; \
63901   __ret_769 = __s0_769 * splat_laneq_f16(__s1_769, __p2_769); \
63902   __ret_769; \
63903 })
63904 #else
63905 #define vmul_laneq_f16(__p0_770, __p1_770, __p2_770) __extension__ ({ \
63906   float16x4_t __ret_770; \
63907   float16x4_t __s0_770 = __p0_770; \
63908   float16x8_t __s1_770 = __p1_770; \
63909   float16x4_t __rev0_770;  __rev0_770 = __builtin_shufflevector(__s0_770, __s0_770, 3, 2, 1, 0); \
63910   float16x8_t __rev1_770;  __rev1_770 = __builtin_shufflevector(__s1_770, __s1_770, 7, 6, 5, 4, 3, 2, 1, 0); \
63911   __ret_770 = __rev0_770 * __noswap_splat_laneq_f16(__rev1_770, __p2_770); \
63912   __ret_770 = __builtin_shufflevector(__ret_770, __ret_770, 3, 2, 1, 0); \
63913   __ret_770; \
63914 })
63915 #endif
63916 
63917 #ifdef __LITTLE_ENDIAN__
vmulxq_f16(float16x8_t __p0,float16x8_t __p1)63918 __ai __attribute__((target("fullfp16"))) float16x8_t vmulxq_f16(float16x8_t __p0, float16x8_t __p1) {
63919   float16x8_t __ret;
63920   __ret = (float16x8_t) __builtin_neon_vmulxq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
63921   return __ret;
63922 }
63923 #else
vmulxq_f16(float16x8_t __p0,float16x8_t __p1)63924 __ai __attribute__((target("fullfp16"))) float16x8_t vmulxq_f16(float16x8_t __p0, float16x8_t __p1) {
63925   float16x8_t __ret;
63926   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63927   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63928   __ret = (float16x8_t) __builtin_neon_vmulxq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
63929   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63930   return __ret;
63931 }
__noswap_vmulxq_f16(float16x8_t __p0,float16x8_t __p1)63932 __ai __attribute__((target("fullfp16"))) float16x8_t __noswap_vmulxq_f16(float16x8_t __p0, float16x8_t __p1) {
63933   float16x8_t __ret;
63934   __ret = (float16x8_t) __builtin_neon_vmulxq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
63935   return __ret;
63936 }
63937 #endif
63938 
63939 #ifdef __LITTLE_ENDIAN__
vmulx_f16(float16x4_t __p0,float16x4_t __p1)63940 __ai __attribute__((target("fullfp16"))) float16x4_t vmulx_f16(float16x4_t __p0, float16x4_t __p1) {
63941   float16x4_t __ret;
63942   __ret = (float16x4_t) __builtin_neon_vmulx_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
63943   return __ret;
63944 }
63945 #else
vmulx_f16(float16x4_t __p0,float16x4_t __p1)63946 __ai __attribute__((target("fullfp16"))) float16x4_t vmulx_f16(float16x4_t __p0, float16x4_t __p1) {
63947   float16x4_t __ret;
63948   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63949   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63950   __ret = (float16x4_t) __builtin_neon_vmulx_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
63951   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63952   return __ret;
63953 }
__noswap_vmulx_f16(float16x4_t __p0,float16x4_t __p1)63954 __ai __attribute__((target("fullfp16"))) float16x4_t __noswap_vmulx_f16(float16x4_t __p0, float16x4_t __p1) {
63955   float16x4_t __ret;
63956   __ret = (float16x4_t) __builtin_neon_vmulx_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
63957   return __ret;
63958 }
63959 #endif
63960 
63961 #ifdef __LITTLE_ENDIAN__
63962 #define vmulxh_lane_f16(__p0, __p1, __p2) __extension__ ({ \
63963   float16_t __ret; \
63964   float16_t __s0 = __p0; \
63965   float16x4_t __s1 = __p1; \
63966   __ret = (float16_t) __builtin_neon_vmulxh_lane_f16(__s0, (float16x4_t)__s1, __p2); \
63967   __ret; \
63968 })
63969 #else
63970 #define vmulxh_lane_f16(__p0, __p1, __p2) __extension__ ({ \
63971   float16_t __ret; \
63972   float16_t __s0 = __p0; \
63973   float16x4_t __s1 = __p1; \
63974   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
63975   __ret = (float16_t) __builtin_neon_vmulxh_lane_f16(__s0, (float16x4_t)__rev1, __p2); \
63976   __ret; \
63977 })
63978 #endif
63979 
63980 #ifdef __LITTLE_ENDIAN__
63981 #define vmulxq_lane_f16(__p0_771, __p1_771, __p2_771) __extension__ ({ \
63982   float16x8_t __ret_771; \
63983   float16x8_t __s0_771 = __p0_771; \
63984   float16x4_t __s1_771 = __p1_771; \
63985   __ret_771 = vmulxq_f16(__s0_771, splatq_lane_f16(__s1_771, __p2_771)); \
63986   __ret_771; \
63987 })
63988 #else
63989 #define vmulxq_lane_f16(__p0_772, __p1_772, __p2_772) __extension__ ({ \
63990   float16x8_t __ret_772; \
63991   float16x8_t __s0_772 = __p0_772; \
63992   float16x4_t __s1_772 = __p1_772; \
63993   float16x8_t __rev0_772;  __rev0_772 = __builtin_shufflevector(__s0_772, __s0_772, 7, 6, 5, 4, 3, 2, 1, 0); \
63994   float16x4_t __rev1_772;  __rev1_772 = __builtin_shufflevector(__s1_772, __s1_772, 3, 2, 1, 0); \
63995   __ret_772 = __noswap_vmulxq_f16(__rev0_772, __noswap_splatq_lane_f16(__rev1_772, __p2_772)); \
63996   __ret_772 = __builtin_shufflevector(__ret_772, __ret_772, 7, 6, 5, 4, 3, 2, 1, 0); \
63997   __ret_772; \
63998 })
63999 #endif
64000 
64001 #ifdef __LITTLE_ENDIAN__
64002 #define vmulx_lane_f16(__p0_773, __p1_773, __p2_773) __extension__ ({ \
64003   float16x4_t __ret_773; \
64004   float16x4_t __s0_773 = __p0_773; \
64005   float16x4_t __s1_773 = __p1_773; \
64006   __ret_773 = vmulx_f16(__s0_773, splat_lane_f16(__s1_773, __p2_773)); \
64007   __ret_773; \
64008 })
64009 #else
64010 #define vmulx_lane_f16(__p0_774, __p1_774, __p2_774) __extension__ ({ \
64011   float16x4_t __ret_774; \
64012   float16x4_t __s0_774 = __p0_774; \
64013   float16x4_t __s1_774 = __p1_774; \
64014   float16x4_t __rev0_774;  __rev0_774 = __builtin_shufflevector(__s0_774, __s0_774, 3, 2, 1, 0); \
64015   float16x4_t __rev1_774;  __rev1_774 = __builtin_shufflevector(__s1_774, __s1_774, 3, 2, 1, 0); \
64016   __ret_774 = __noswap_vmulx_f16(__rev0_774, __noswap_splat_lane_f16(__rev1_774, __p2_774)); \
64017   __ret_774 = __builtin_shufflevector(__ret_774, __ret_774, 3, 2, 1, 0); \
64018   __ret_774; \
64019 })
64020 #endif
64021 
64022 #ifdef __LITTLE_ENDIAN__
64023 #define vmulxh_laneq_f16(__p0, __p1, __p2) __extension__ ({ \
64024   float16_t __ret; \
64025   float16_t __s0 = __p0; \
64026   float16x8_t __s1 = __p1; \
64027   __ret = (float16_t) __builtin_neon_vmulxh_laneq_f16(__s0, (float16x8_t)__s1, __p2); \
64028   __ret; \
64029 })
64030 #else
64031 #define vmulxh_laneq_f16(__p0, __p1, __p2) __extension__ ({ \
64032   float16_t __ret; \
64033   float16_t __s0 = __p0; \
64034   float16x8_t __s1 = __p1; \
64035   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
64036   __ret = (float16_t) __builtin_neon_vmulxh_laneq_f16(__s0, (float16x8_t)__rev1, __p2); \
64037   __ret; \
64038 })
64039 #endif
64040 
64041 #ifdef __LITTLE_ENDIAN__
64042 #define vmulxq_laneq_f16(__p0_775, __p1_775, __p2_775) __extension__ ({ \
64043   float16x8_t __ret_775; \
64044   float16x8_t __s0_775 = __p0_775; \
64045   float16x8_t __s1_775 = __p1_775; \
64046   __ret_775 = vmulxq_f16(__s0_775, splatq_laneq_f16(__s1_775, __p2_775)); \
64047   __ret_775; \
64048 })
64049 #else
64050 #define vmulxq_laneq_f16(__p0_776, __p1_776, __p2_776) __extension__ ({ \
64051   float16x8_t __ret_776; \
64052   float16x8_t __s0_776 = __p0_776; \
64053   float16x8_t __s1_776 = __p1_776; \
64054   float16x8_t __rev0_776;  __rev0_776 = __builtin_shufflevector(__s0_776, __s0_776, 7, 6, 5, 4, 3, 2, 1, 0); \
64055   float16x8_t __rev1_776;  __rev1_776 = __builtin_shufflevector(__s1_776, __s1_776, 7, 6, 5, 4, 3, 2, 1, 0); \
64056   __ret_776 = __noswap_vmulxq_f16(__rev0_776, __noswap_splatq_laneq_f16(__rev1_776, __p2_776)); \
64057   __ret_776 = __builtin_shufflevector(__ret_776, __ret_776, 7, 6, 5, 4, 3, 2, 1, 0); \
64058   __ret_776; \
64059 })
64060 #endif
64061 
64062 #ifdef __LITTLE_ENDIAN__
64063 #define vmulx_laneq_f16(__p0_777, __p1_777, __p2_777) __extension__ ({ \
64064   float16x4_t __ret_777; \
64065   float16x4_t __s0_777 = __p0_777; \
64066   float16x8_t __s1_777 = __p1_777; \
64067   __ret_777 = vmulx_f16(__s0_777, splat_laneq_f16(__s1_777, __p2_777)); \
64068   __ret_777; \
64069 })
64070 #else
64071 #define vmulx_laneq_f16(__p0_778, __p1_778, __p2_778) __extension__ ({ \
64072   float16x4_t __ret_778; \
64073   float16x4_t __s0_778 = __p0_778; \
64074   float16x8_t __s1_778 = __p1_778; \
64075   float16x4_t __rev0_778;  __rev0_778 = __builtin_shufflevector(__s0_778, __s0_778, 3, 2, 1, 0); \
64076   float16x8_t __rev1_778;  __rev1_778 = __builtin_shufflevector(__s1_778, __s1_778, 7, 6, 5, 4, 3, 2, 1, 0); \
64077   __ret_778 = __noswap_vmulx_f16(__rev0_778, __noswap_splat_laneq_f16(__rev1_778, __p2_778)); \
64078   __ret_778 = __builtin_shufflevector(__ret_778, __ret_778, 3, 2, 1, 0); \
64079   __ret_778; \
64080 })
64081 #endif
64082 
64083 #ifdef __LITTLE_ENDIAN__
64084 #define vmulxq_n_f16(__p0, __p1) __extension__ ({ \
64085   float16x8_t __ret; \
64086   float16x8_t __s0 = __p0; \
64087   float16_t __s1 = __p1; \
64088   __ret = vmulxq_f16(__s0, (float16x8_t) {__s1, __s1, __s1, __s1, __s1, __s1, __s1, __s1}); \
64089   __ret; \
64090 })
64091 #else
64092 #define vmulxq_n_f16(__p0, __p1) __extension__ ({ \
64093   float16x8_t __ret; \
64094   float16x8_t __s0 = __p0; \
64095   float16_t __s1 = __p1; \
64096   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
64097   __ret = __noswap_vmulxq_f16(__rev0, (float16x8_t) {__s1, __s1, __s1, __s1, __s1, __s1, __s1, __s1}); \
64098   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
64099   __ret; \
64100 })
64101 #endif
64102 
64103 #ifdef __LITTLE_ENDIAN__
64104 #define vmulx_n_f16(__p0, __p1) __extension__ ({ \
64105   float16x4_t __ret; \
64106   float16x4_t __s0 = __p0; \
64107   float16_t __s1 = __p1; \
64108   __ret = vmulx_f16(__s0, (float16x4_t) {__s1, __s1, __s1, __s1}); \
64109   __ret; \
64110 })
64111 #else
64112 #define vmulx_n_f16(__p0, __p1) __extension__ ({ \
64113   float16x4_t __ret; \
64114   float16x4_t __s0 = __p0; \
64115   float16_t __s1 = __p1; \
64116   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
64117   __ret = __noswap_vmulx_f16(__rev0, (float16x4_t) {__s1, __s1, __s1, __s1}); \
64118   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
64119   __ret; \
64120 })
64121 #endif
64122 
64123 #ifdef __LITTLE_ENDIAN__
vpaddq_f16(float16x8_t __p0,float16x8_t __p1)64124 __ai __attribute__((target("fullfp16"))) float16x8_t vpaddq_f16(float16x8_t __p0, float16x8_t __p1) {
64125   float16x8_t __ret;
64126   __ret = (float16x8_t) __builtin_neon_vpaddq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
64127   return __ret;
64128 }
64129 #else
vpaddq_f16(float16x8_t __p0,float16x8_t __p1)64130 __ai __attribute__((target("fullfp16"))) float16x8_t vpaddq_f16(float16x8_t __p0, float16x8_t __p1) {
64131   float16x8_t __ret;
64132   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64133   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64134   __ret = (float16x8_t) __builtin_neon_vpaddq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
64135   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64136   return __ret;
64137 }
64138 #endif
64139 
64140 #ifdef __LITTLE_ENDIAN__
vpmaxq_f16(float16x8_t __p0,float16x8_t __p1)64141 __ai __attribute__((target("fullfp16"))) float16x8_t vpmaxq_f16(float16x8_t __p0, float16x8_t __p1) {
64142   float16x8_t __ret;
64143   __ret = (float16x8_t) __builtin_neon_vpmaxq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
64144   return __ret;
64145 }
64146 #else
vpmaxq_f16(float16x8_t __p0,float16x8_t __p1)64147 __ai __attribute__((target("fullfp16"))) float16x8_t vpmaxq_f16(float16x8_t __p0, float16x8_t __p1) {
64148   float16x8_t __ret;
64149   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64150   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64151   __ret = (float16x8_t) __builtin_neon_vpmaxq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
64152   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64153   return __ret;
64154 }
64155 #endif
64156 
64157 #ifdef __LITTLE_ENDIAN__
vpmaxnmq_f16(float16x8_t __p0,float16x8_t __p1)64158 __ai __attribute__((target("fullfp16"))) float16x8_t vpmaxnmq_f16(float16x8_t __p0, float16x8_t __p1) {
64159   float16x8_t __ret;
64160   __ret = (float16x8_t) __builtin_neon_vpmaxnmq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
64161   return __ret;
64162 }
64163 #else
vpmaxnmq_f16(float16x8_t __p0,float16x8_t __p1)64164 __ai __attribute__((target("fullfp16"))) float16x8_t vpmaxnmq_f16(float16x8_t __p0, float16x8_t __p1) {
64165   float16x8_t __ret;
64166   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64167   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64168   __ret = (float16x8_t) __builtin_neon_vpmaxnmq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
64169   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64170   return __ret;
64171 }
64172 #endif
64173 
64174 #ifdef __LITTLE_ENDIAN__
vpmaxnm_f16(float16x4_t __p0,float16x4_t __p1)64175 __ai __attribute__((target("fullfp16"))) float16x4_t vpmaxnm_f16(float16x4_t __p0, float16x4_t __p1) {
64176   float16x4_t __ret;
64177   __ret = (float16x4_t) __builtin_neon_vpmaxnm_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
64178   return __ret;
64179 }
64180 #else
vpmaxnm_f16(float16x4_t __p0,float16x4_t __p1)64181 __ai __attribute__((target("fullfp16"))) float16x4_t vpmaxnm_f16(float16x4_t __p0, float16x4_t __p1) {
64182   float16x4_t __ret;
64183   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64184   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64185   __ret = (float16x4_t) __builtin_neon_vpmaxnm_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
64186   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64187   return __ret;
64188 }
64189 #endif
64190 
64191 #ifdef __LITTLE_ENDIAN__
vpminq_f16(float16x8_t __p0,float16x8_t __p1)64192 __ai __attribute__((target("fullfp16"))) float16x8_t vpminq_f16(float16x8_t __p0, float16x8_t __p1) {
64193   float16x8_t __ret;
64194   __ret = (float16x8_t) __builtin_neon_vpminq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
64195   return __ret;
64196 }
64197 #else
vpminq_f16(float16x8_t __p0,float16x8_t __p1)64198 __ai __attribute__((target("fullfp16"))) float16x8_t vpminq_f16(float16x8_t __p0, float16x8_t __p1) {
64199   float16x8_t __ret;
64200   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64201   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64202   __ret = (float16x8_t) __builtin_neon_vpminq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
64203   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64204   return __ret;
64205 }
64206 #endif
64207 
64208 #ifdef __LITTLE_ENDIAN__
vpminnmq_f16(float16x8_t __p0,float16x8_t __p1)64209 __ai __attribute__((target("fullfp16"))) float16x8_t vpminnmq_f16(float16x8_t __p0, float16x8_t __p1) {
64210   float16x8_t __ret;
64211   __ret = (float16x8_t) __builtin_neon_vpminnmq_f16((int8x16_t)__p0, (int8x16_t)__p1, 40);
64212   return __ret;
64213 }
64214 #else
vpminnmq_f16(float16x8_t __p0,float16x8_t __p1)64215 __ai __attribute__((target("fullfp16"))) float16x8_t vpminnmq_f16(float16x8_t __p0, float16x8_t __p1) {
64216   float16x8_t __ret;
64217   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64218   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64219   __ret = (float16x8_t) __builtin_neon_vpminnmq_f16((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
64220   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64221   return __ret;
64222 }
64223 #endif
64224 
64225 #ifdef __LITTLE_ENDIAN__
vpminnm_f16(float16x4_t __p0,float16x4_t __p1)64226 __ai __attribute__((target("fullfp16"))) float16x4_t vpminnm_f16(float16x4_t __p0, float16x4_t __p1) {
64227   float16x4_t __ret;
64228   __ret = (float16x4_t) __builtin_neon_vpminnm_f16((int8x8_t)__p0, (int8x8_t)__p1, 8);
64229   return __ret;
64230 }
64231 #else
vpminnm_f16(float16x4_t __p0,float16x4_t __p1)64232 __ai __attribute__((target("fullfp16"))) float16x4_t vpminnm_f16(float16x4_t __p0, float16x4_t __p1) {
64233   float16x4_t __ret;
64234   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64235   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64236   __ret = (float16x4_t) __builtin_neon_vpminnm_f16((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
64237   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64238   return __ret;
64239 }
64240 #endif
64241 
64242 #ifdef __LITTLE_ENDIAN__
vrndiq_f16(float16x8_t __p0)64243 __ai __attribute__((target("fullfp16"))) float16x8_t vrndiq_f16(float16x8_t __p0) {
64244   float16x8_t __ret;
64245   __ret = (float16x8_t) __builtin_neon_vrndiq_f16((int8x16_t)__p0, 40);
64246   return __ret;
64247 }
64248 #else
vrndiq_f16(float16x8_t __p0)64249 __ai __attribute__((target("fullfp16"))) float16x8_t vrndiq_f16(float16x8_t __p0) {
64250   float16x8_t __ret;
64251   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64252   __ret = (float16x8_t) __builtin_neon_vrndiq_f16((int8x16_t)__rev0, 40);
64253   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64254   return __ret;
64255 }
64256 #endif
64257 
64258 #ifdef __LITTLE_ENDIAN__
vrndi_f16(float16x4_t __p0)64259 __ai __attribute__((target("fullfp16"))) float16x4_t vrndi_f16(float16x4_t __p0) {
64260   float16x4_t __ret;
64261   __ret = (float16x4_t) __builtin_neon_vrndi_f16((int8x8_t)__p0, 8);
64262   return __ret;
64263 }
64264 #else
vrndi_f16(float16x4_t __p0)64265 __ai __attribute__((target("fullfp16"))) float16x4_t vrndi_f16(float16x4_t __p0) {
64266   float16x4_t __ret;
64267   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64268   __ret = (float16x4_t) __builtin_neon_vrndi_f16((int8x8_t)__rev0, 8);
64269   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64270   return __ret;
64271 }
64272 #endif
64273 
64274 #ifdef __LITTLE_ENDIAN__
vsqrtq_f16(float16x8_t __p0)64275 __ai __attribute__((target("fullfp16"))) float16x8_t vsqrtq_f16(float16x8_t __p0) {
64276   float16x8_t __ret;
64277   __ret = (float16x8_t) __builtin_neon_vsqrtq_f16((int8x16_t)__p0, 40);
64278   return __ret;
64279 }
64280 #else
vsqrtq_f16(float16x8_t __p0)64281 __ai __attribute__((target("fullfp16"))) float16x8_t vsqrtq_f16(float16x8_t __p0) {
64282   float16x8_t __ret;
64283   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64284   __ret = (float16x8_t) __builtin_neon_vsqrtq_f16((int8x16_t)__rev0, 40);
64285   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64286   return __ret;
64287 }
64288 #endif
64289 
64290 #ifdef __LITTLE_ENDIAN__
vsqrt_f16(float16x4_t __p0)64291 __ai __attribute__((target("fullfp16"))) float16x4_t vsqrt_f16(float16x4_t __p0) {
64292   float16x4_t __ret;
64293   __ret = (float16x4_t) __builtin_neon_vsqrt_f16((int8x8_t)__p0, 8);
64294   return __ret;
64295 }
64296 #else
vsqrt_f16(float16x4_t __p0)64297 __ai __attribute__((target("fullfp16"))) float16x4_t vsqrt_f16(float16x4_t __p0) {
64298   float16x4_t __ret;
64299   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64300   __ret = (float16x4_t) __builtin_neon_vsqrt_f16((int8x8_t)__rev0, 8);
64301   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64302   return __ret;
64303 }
64304 #endif
64305 
64306 #ifdef __LITTLE_ENDIAN__
vtrn1q_f16(float16x8_t __p0,float16x8_t __p1)64307 __ai __attribute__((target("fullfp16"))) float16x8_t vtrn1q_f16(float16x8_t __p0, float16x8_t __p1) {
64308   float16x8_t __ret;
64309   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64310   return __ret;
64311 }
64312 #else
vtrn1q_f16(float16x8_t __p0,float16x8_t __p1)64313 __ai __attribute__((target("fullfp16"))) float16x8_t vtrn1q_f16(float16x8_t __p0, float16x8_t __p1) {
64314   float16x8_t __ret;
64315   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64316   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64317   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64318   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64319   return __ret;
64320 }
64321 #endif
64322 
64323 #ifdef __LITTLE_ENDIAN__
vtrn1_f16(float16x4_t __p0,float16x4_t __p1)64324 __ai __attribute__((target("fullfp16"))) float16x4_t vtrn1_f16(float16x4_t __p0, float16x4_t __p1) {
64325   float16x4_t __ret;
64326   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64327   return __ret;
64328 }
64329 #else
vtrn1_f16(float16x4_t __p0,float16x4_t __p1)64330 __ai __attribute__((target("fullfp16"))) float16x4_t vtrn1_f16(float16x4_t __p0, float16x4_t __p1) {
64331   float16x4_t __ret;
64332   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64333   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64334   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64335   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64336   return __ret;
64337 }
64338 #endif
64339 
64340 #ifdef __LITTLE_ENDIAN__
vtrn2q_f16(float16x8_t __p0,float16x8_t __p1)64341 __ai __attribute__((target("fullfp16"))) float16x8_t vtrn2q_f16(float16x8_t __p0, float16x8_t __p1) {
64342   float16x8_t __ret;
64343   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64344   return __ret;
64345 }
64346 #else
vtrn2q_f16(float16x8_t __p0,float16x8_t __p1)64347 __ai __attribute__((target("fullfp16"))) float16x8_t vtrn2q_f16(float16x8_t __p0, float16x8_t __p1) {
64348   float16x8_t __ret;
64349   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64350   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64351   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64352   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64353   return __ret;
64354 }
64355 #endif
64356 
64357 #ifdef __LITTLE_ENDIAN__
vtrn2_f16(float16x4_t __p0,float16x4_t __p1)64358 __ai __attribute__((target("fullfp16"))) float16x4_t vtrn2_f16(float16x4_t __p0, float16x4_t __p1) {
64359   float16x4_t __ret;
64360   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64361   return __ret;
64362 }
64363 #else
vtrn2_f16(float16x4_t __p0,float16x4_t __p1)64364 __ai __attribute__((target("fullfp16"))) float16x4_t vtrn2_f16(float16x4_t __p0, float16x4_t __p1) {
64365   float16x4_t __ret;
64366   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64367   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64368   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64369   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64370   return __ret;
64371 }
64372 #endif
64373 
64374 #ifdef __LITTLE_ENDIAN__
vuzp1q_f16(float16x8_t __p0,float16x8_t __p1)64375 __ai __attribute__((target("fullfp16"))) float16x8_t vuzp1q_f16(float16x8_t __p0, float16x8_t __p1) {
64376   float16x8_t __ret;
64377   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
64378   return __ret;
64379 }
64380 #else
vuzp1q_f16(float16x8_t __p0,float16x8_t __p1)64381 __ai __attribute__((target("fullfp16"))) float16x8_t vuzp1q_f16(float16x8_t __p0, float16x8_t __p1) {
64382   float16x8_t __ret;
64383   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64384   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64385   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
64386   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64387   return __ret;
64388 }
64389 #endif
64390 
64391 #ifdef __LITTLE_ENDIAN__
vuzp1_f16(float16x4_t __p0,float16x4_t __p1)64392 __ai __attribute__((target("fullfp16"))) float16x4_t vuzp1_f16(float16x4_t __p0, float16x4_t __p1) {
64393   float16x4_t __ret;
64394   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
64395   return __ret;
64396 }
64397 #else
vuzp1_f16(float16x4_t __p0,float16x4_t __p1)64398 __ai __attribute__((target("fullfp16"))) float16x4_t vuzp1_f16(float16x4_t __p0, float16x4_t __p1) {
64399   float16x4_t __ret;
64400   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64401   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64402   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
64403   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64404   return __ret;
64405 }
64406 #endif
64407 
64408 #ifdef __LITTLE_ENDIAN__
vuzp2q_f16(float16x8_t __p0,float16x8_t __p1)64409 __ai __attribute__((target("fullfp16"))) float16x8_t vuzp2q_f16(float16x8_t __p0, float16x8_t __p1) {
64410   float16x8_t __ret;
64411   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
64412   return __ret;
64413 }
64414 #else
vuzp2q_f16(float16x8_t __p0,float16x8_t __p1)64415 __ai __attribute__((target("fullfp16"))) float16x8_t vuzp2q_f16(float16x8_t __p0, float16x8_t __p1) {
64416   float16x8_t __ret;
64417   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64418   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64419   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
64420   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64421   return __ret;
64422 }
64423 #endif
64424 
64425 #ifdef __LITTLE_ENDIAN__
vuzp2_f16(float16x4_t __p0,float16x4_t __p1)64426 __ai __attribute__((target("fullfp16"))) float16x4_t vuzp2_f16(float16x4_t __p0, float16x4_t __p1) {
64427   float16x4_t __ret;
64428   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
64429   return __ret;
64430 }
64431 #else
vuzp2_f16(float16x4_t __p0,float16x4_t __p1)64432 __ai __attribute__((target("fullfp16"))) float16x4_t vuzp2_f16(float16x4_t __p0, float16x4_t __p1) {
64433   float16x4_t __ret;
64434   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64435   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64436   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
64437   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64438   return __ret;
64439 }
64440 #endif
64441 
64442 #ifdef __LITTLE_ENDIAN__
vzip1q_f16(float16x8_t __p0,float16x8_t __p1)64443 __ai __attribute__((target("fullfp16"))) float16x8_t vzip1q_f16(float16x8_t __p0, float16x8_t __p1) {
64444   float16x8_t __ret;
64445   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
64446   return __ret;
64447 }
64448 #else
vzip1q_f16(float16x8_t __p0,float16x8_t __p1)64449 __ai __attribute__((target("fullfp16"))) float16x8_t vzip1q_f16(float16x8_t __p0, float16x8_t __p1) {
64450   float16x8_t __ret;
64451   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64452   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64453   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
64454   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64455   return __ret;
64456 }
64457 #endif
64458 
64459 #ifdef __LITTLE_ENDIAN__
vzip1_f16(float16x4_t __p0,float16x4_t __p1)64460 __ai __attribute__((target("fullfp16"))) float16x4_t vzip1_f16(float16x4_t __p0, float16x4_t __p1) {
64461   float16x4_t __ret;
64462   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
64463   return __ret;
64464 }
64465 #else
vzip1_f16(float16x4_t __p0,float16x4_t __p1)64466 __ai __attribute__((target("fullfp16"))) float16x4_t vzip1_f16(float16x4_t __p0, float16x4_t __p1) {
64467   float16x4_t __ret;
64468   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64469   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64470   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
64471   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64472   return __ret;
64473 }
64474 #endif
64475 
64476 #ifdef __LITTLE_ENDIAN__
vzip2q_f16(float16x8_t __p0,float16x8_t __p1)64477 __ai __attribute__((target("fullfp16"))) float16x8_t vzip2q_f16(float16x8_t __p0, float16x8_t __p1) {
64478   float16x8_t __ret;
64479   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
64480   return __ret;
64481 }
64482 #else
vzip2q_f16(float16x8_t __p0,float16x8_t __p1)64483 __ai __attribute__((target("fullfp16"))) float16x8_t vzip2q_f16(float16x8_t __p0, float16x8_t __p1) {
64484   float16x8_t __ret;
64485   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64486   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64487   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
64488   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64489   return __ret;
64490 }
64491 #endif
64492 
64493 #ifdef __LITTLE_ENDIAN__
vzip2_f16(float16x4_t __p0,float16x4_t __p1)64494 __ai __attribute__((target("fullfp16"))) float16x4_t vzip2_f16(float16x4_t __p0, float16x4_t __p1) {
64495   float16x4_t __ret;
64496   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
64497   return __ret;
64498 }
64499 #else
vzip2_f16(float16x4_t __p0,float16x4_t __p1)64500 __ai __attribute__((target("fullfp16"))) float16x4_t vzip2_f16(float16x4_t __p0, float16x4_t __p1) {
64501   float16x4_t __ret;
64502   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64503   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64504   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
64505   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64506   return __ret;
64507 }
64508 #endif
64509 
64510 #ifdef __LITTLE_ENDIAN__
64511 #define vsudotq_laneq_s32(__p0_779, __p1_779, __p2_779, __p3_779) __extension__ ({ \
64512   int32x4_t __ret_779; \
64513   int32x4_t __s0_779 = __p0_779; \
64514   int8x16_t __s1_779 = __p1_779; \
64515   uint8x16_t __s2_779 = __p2_779; \
64516 uint8x16_t __reint_779 = __s2_779; \
64517   __ret_779 = vusdotq_s32(__s0_779, (uint8x16_t)(splatq_laneq_s32(*(int32x4_t *) &__reint_779, __p3_779)), __s1_779); \
64518   __ret_779; \
64519 })
64520 #else
64521 #define vsudotq_laneq_s32(__p0_780, __p1_780, __p2_780, __p3_780) __extension__ ({ \
64522   int32x4_t __ret_780; \
64523   int32x4_t __s0_780 = __p0_780; \
64524   int8x16_t __s1_780 = __p1_780; \
64525   uint8x16_t __s2_780 = __p2_780; \
64526   int32x4_t __rev0_780;  __rev0_780 = __builtin_shufflevector(__s0_780, __s0_780, 3, 2, 1, 0); \
64527   int8x16_t __rev1_780;  __rev1_780 = __builtin_shufflevector(__s1_780, __s1_780, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
64528   uint8x16_t __rev2_780;  __rev2_780 = __builtin_shufflevector(__s2_780, __s2_780, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
64529 uint8x16_t __reint_780 = __rev2_780; \
64530   __ret_780 = __noswap_vusdotq_s32(__rev0_780, (uint8x16_t)(__noswap_splatq_laneq_s32(*(int32x4_t *) &__reint_780, __p3_780)), __rev1_780); \
64531   __ret_780 = __builtin_shufflevector(__ret_780, __ret_780, 3, 2, 1, 0); \
64532   __ret_780; \
64533 })
64534 #endif
64535 
64536 #ifdef __LITTLE_ENDIAN__
64537 #define vsudot_laneq_s32(__p0_781, __p1_781, __p2_781, __p3_781) __extension__ ({ \
64538   int32x2_t __ret_781; \
64539   int32x2_t __s0_781 = __p0_781; \
64540   int8x8_t __s1_781 = __p1_781; \
64541   uint8x16_t __s2_781 = __p2_781; \
64542 uint8x16_t __reint_781 = __s2_781; \
64543   __ret_781 = vusdot_s32(__s0_781, (uint8x8_t)(splat_laneq_s32(*(int32x4_t *) &__reint_781, __p3_781)), __s1_781); \
64544   __ret_781; \
64545 })
64546 #else
64547 #define vsudot_laneq_s32(__p0_782, __p1_782, __p2_782, __p3_782) __extension__ ({ \
64548   int32x2_t __ret_782; \
64549   int32x2_t __s0_782 = __p0_782; \
64550   int8x8_t __s1_782 = __p1_782; \
64551   uint8x16_t __s2_782 = __p2_782; \
64552   int32x2_t __rev0_782;  __rev0_782 = __builtin_shufflevector(__s0_782, __s0_782, 1, 0); \
64553   int8x8_t __rev1_782;  __rev1_782 = __builtin_shufflevector(__s1_782, __s1_782, 7, 6, 5, 4, 3, 2, 1, 0); \
64554   uint8x16_t __rev2_782;  __rev2_782 = __builtin_shufflevector(__s2_782, __s2_782, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
64555 uint8x16_t __reint_782 = __rev2_782; \
64556   __ret_782 = __noswap_vusdot_s32(__rev0_782, (uint8x8_t)(__noswap_splat_laneq_s32(*(int32x4_t *) &__reint_782, __p3_782)), __rev1_782); \
64557   __ret_782 = __builtin_shufflevector(__ret_782, __ret_782, 1, 0); \
64558   __ret_782; \
64559 })
64560 #endif
64561 
64562 #ifdef __LITTLE_ENDIAN__
64563 #define vusdotq_laneq_s32(__p0_783, __p1_783, __p2_783, __p3_783) __extension__ ({ \
64564   int32x4_t __ret_783; \
64565   int32x4_t __s0_783 = __p0_783; \
64566   uint8x16_t __s1_783 = __p1_783; \
64567   int8x16_t __s2_783 = __p2_783; \
64568 int8x16_t __reint_783 = __s2_783; \
64569   __ret_783 = vusdotq_s32(__s0_783, __s1_783, (int8x16_t)(splatq_laneq_s32(*(int32x4_t *) &__reint_783, __p3_783))); \
64570   __ret_783; \
64571 })
64572 #else
64573 #define vusdotq_laneq_s32(__p0_784, __p1_784, __p2_784, __p3_784) __extension__ ({ \
64574   int32x4_t __ret_784; \
64575   int32x4_t __s0_784 = __p0_784; \
64576   uint8x16_t __s1_784 = __p1_784; \
64577   int8x16_t __s2_784 = __p2_784; \
64578   int32x4_t __rev0_784;  __rev0_784 = __builtin_shufflevector(__s0_784, __s0_784, 3, 2, 1, 0); \
64579   uint8x16_t __rev1_784;  __rev1_784 = __builtin_shufflevector(__s1_784, __s1_784, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
64580   int8x16_t __rev2_784;  __rev2_784 = __builtin_shufflevector(__s2_784, __s2_784, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
64581 int8x16_t __reint_784 = __rev2_784; \
64582   __ret_784 = __noswap_vusdotq_s32(__rev0_784, __rev1_784, (int8x16_t)(__noswap_splatq_laneq_s32(*(int32x4_t *) &__reint_784, __p3_784))); \
64583   __ret_784 = __builtin_shufflevector(__ret_784, __ret_784, 3, 2, 1, 0); \
64584   __ret_784; \
64585 })
64586 #endif
64587 
64588 #ifdef __LITTLE_ENDIAN__
64589 #define vusdot_laneq_s32(__p0_785, __p1_785, __p2_785, __p3_785) __extension__ ({ \
64590   int32x2_t __ret_785; \
64591   int32x2_t __s0_785 = __p0_785; \
64592   uint8x8_t __s1_785 = __p1_785; \
64593   int8x16_t __s2_785 = __p2_785; \
64594 int8x16_t __reint_785 = __s2_785; \
64595   __ret_785 = vusdot_s32(__s0_785, __s1_785, (int8x8_t)(splat_laneq_s32(*(int32x4_t *) &__reint_785, __p3_785))); \
64596   __ret_785; \
64597 })
64598 #else
64599 #define vusdot_laneq_s32(__p0_786, __p1_786, __p2_786, __p3_786) __extension__ ({ \
64600   int32x2_t __ret_786; \
64601   int32x2_t __s0_786 = __p0_786; \
64602   uint8x8_t __s1_786 = __p1_786; \
64603   int8x16_t __s2_786 = __p2_786; \
64604   int32x2_t __rev0_786;  __rev0_786 = __builtin_shufflevector(__s0_786, __s0_786, 1, 0); \
64605   uint8x8_t __rev1_786;  __rev1_786 = __builtin_shufflevector(__s1_786, __s1_786, 7, 6, 5, 4, 3, 2, 1, 0); \
64606   int8x16_t __rev2_786;  __rev2_786 = __builtin_shufflevector(__s2_786, __s2_786, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
64607 int8x16_t __reint_786 = __rev2_786; \
64608   __ret_786 = __noswap_vusdot_s32(__rev0_786, __rev1_786, (int8x8_t)(__noswap_splat_laneq_s32(*(int32x4_t *) &__reint_786, __p3_786))); \
64609   __ret_786 = __builtin_shufflevector(__ret_786, __ret_786, 1, 0); \
64610   __ret_786; \
64611 })
64612 #endif
64613 
64614 #define vldap1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
64615   poly64x1_t __ret; \
64616   poly64x1_t __s1 = __p1; \
64617   __ret = (poly64x1_t) __builtin_neon_vldap1_lane_p64(__p0, (int8x8_t)__s1, __p2, 6); \
64618   __ret; \
64619 })
64620 #ifdef __LITTLE_ENDIAN__
64621 #define vldap1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
64622   poly64x2_t __ret; \
64623   poly64x2_t __s1 = __p1; \
64624   __ret = (poly64x2_t) __builtin_neon_vldap1q_lane_p64(__p0, (int8x16_t)__s1, __p2, 38); \
64625   __ret; \
64626 })
64627 #else
64628 #define vldap1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
64629   poly64x2_t __ret; \
64630   poly64x2_t __s1 = __p1; \
64631   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
64632   __ret = (poly64x2_t) __builtin_neon_vldap1q_lane_p64(__p0, (int8x16_t)__rev1, __p2, 38); \
64633   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
64634   __ret; \
64635 })
64636 #endif
64637 
64638 #ifdef __LITTLE_ENDIAN__
64639 #define vldap1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
64640   uint64x2_t __ret; \
64641   uint64x2_t __s1 = __p1; \
64642   __ret = (uint64x2_t) __builtin_neon_vldap1q_lane_u64(__p0, (int8x16_t)__s1, __p2, 51); \
64643   __ret; \
64644 })
64645 #else
64646 #define vldap1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
64647   uint64x2_t __ret; \
64648   uint64x2_t __s1 = __p1; \
64649   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
64650   __ret = (uint64x2_t) __builtin_neon_vldap1q_lane_u64(__p0, (int8x16_t)__rev1, __p2, 51); \
64651   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
64652   __ret; \
64653 })
64654 #endif
64655 
64656 #ifdef __LITTLE_ENDIAN__
64657 #define vldap1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
64658   float64x2_t __ret; \
64659   float64x2_t __s1 = __p1; \
64660   __ret = (float64x2_t) __builtin_neon_vldap1q_lane_f64(__p0, (int8x16_t)__s1, __p2, 42); \
64661   __ret; \
64662 })
64663 #else
64664 #define vldap1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
64665   float64x2_t __ret; \
64666   float64x2_t __s1 = __p1; \
64667   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
64668   __ret = (float64x2_t) __builtin_neon_vldap1q_lane_f64(__p0, (int8x16_t)__rev1, __p2, 42); \
64669   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
64670   __ret; \
64671 })
64672 #endif
64673 
64674 #ifdef __LITTLE_ENDIAN__
64675 #define vldap1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
64676   int64x2_t __ret; \
64677   int64x2_t __s1 = __p1; \
64678   __ret = (int64x2_t) __builtin_neon_vldap1q_lane_s64(__p0, (int8x16_t)__s1, __p2, 35); \
64679   __ret; \
64680 })
64681 #else
64682 #define vldap1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
64683   int64x2_t __ret; \
64684   int64x2_t __s1 = __p1; \
64685   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
64686   __ret = (int64x2_t) __builtin_neon_vldap1q_lane_s64(__p0, (int8x16_t)__rev1, __p2, 35); \
64687   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
64688   __ret; \
64689 })
64690 #endif
64691 
64692 #define vldap1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
64693   uint64x1_t __ret; \
64694   uint64x1_t __s1 = __p1; \
64695   __ret = (uint64x1_t) __builtin_neon_vldap1_lane_u64(__p0, (int8x8_t)__s1, __p2, 19); \
64696   __ret; \
64697 })
64698 #define vldap1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
64699   float64x1_t __ret; \
64700   float64x1_t __s1 = __p1; \
64701   __ret = (float64x1_t) __builtin_neon_vldap1_lane_f64(__p0, (int8x8_t)__s1, __p2, 10); \
64702   __ret; \
64703 })
64704 #define vldap1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
64705   int64x1_t __ret; \
64706   int64x1_t __s1 = __p1; \
64707   __ret = (int64x1_t) __builtin_neon_vldap1_lane_s64(__p0, (int8x8_t)__s1, __p2, 3); \
64708   __ret; \
64709 })
64710 #define vstl1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
64711   poly64x1_t __s1 = __p1; \
64712   __builtin_neon_vstl1_lane_p64(__p0, (int8x8_t)__s1, __p2, 6); \
64713 })
64714 #ifdef __LITTLE_ENDIAN__
64715 #define vstl1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
64716   poly64x2_t __s1 = __p1; \
64717   __builtin_neon_vstl1q_lane_p64(__p0, (int8x16_t)__s1, __p2, 38); \
64718 })
64719 #else
64720 #define vstl1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
64721   poly64x2_t __s1 = __p1; \
64722   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
64723   __builtin_neon_vstl1q_lane_p64(__p0, (int8x16_t)__rev1, __p2, 38); \
64724 })
64725 #endif
64726 
64727 #ifdef __LITTLE_ENDIAN__
64728 #define vstl1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
64729   uint64x2_t __s1 = __p1; \
64730   __builtin_neon_vstl1q_lane_u64(__p0, (int8x16_t)__s1, __p2, 51); \
64731 })
64732 #else
64733 #define vstl1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
64734   uint64x2_t __s1 = __p1; \
64735   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
64736   __builtin_neon_vstl1q_lane_u64(__p0, (int8x16_t)__rev1, __p2, 51); \
64737 })
64738 #endif
64739 
64740 #ifdef __LITTLE_ENDIAN__
64741 #define vstl1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
64742   float64x2_t __s1 = __p1; \
64743   __builtin_neon_vstl1q_lane_f64(__p0, (int8x16_t)__s1, __p2, 42); \
64744 })
64745 #else
64746 #define vstl1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
64747   float64x2_t __s1 = __p1; \
64748   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
64749   __builtin_neon_vstl1q_lane_f64(__p0, (int8x16_t)__rev1, __p2, 42); \
64750 })
64751 #endif
64752 
64753 #ifdef __LITTLE_ENDIAN__
64754 #define vstl1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
64755   int64x2_t __s1 = __p1; \
64756   __builtin_neon_vstl1q_lane_s64(__p0, (int8x16_t)__s1, __p2, 35); \
64757 })
64758 #else
64759 #define vstl1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
64760   int64x2_t __s1 = __p1; \
64761   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
64762   __builtin_neon_vstl1q_lane_s64(__p0, (int8x16_t)__rev1, __p2, 35); \
64763 })
64764 #endif
64765 
64766 #define vstl1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
64767   uint64x1_t __s1 = __p1; \
64768   __builtin_neon_vstl1_lane_u64(__p0, (int8x8_t)__s1, __p2, 19); \
64769 })
64770 #define vstl1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
64771   float64x1_t __s1 = __p1; \
64772   __builtin_neon_vstl1_lane_f64(__p0, (int8x8_t)__s1, __p2, 10); \
64773 })
64774 #define vstl1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
64775   int64x1_t __s1 = __p1; \
64776   __builtin_neon_vstl1_lane_s64(__p0, (int8x8_t)__s1, __p2, 3); \
64777 })
64778 #ifdef __LITTLE_ENDIAN__
vbcaxq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)64779 __ai __attribute__((target("sha3"))) uint8x16_t vbcaxq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
64780   uint8x16_t __ret;
64781   __ret = (uint8x16_t) __builtin_neon_vbcaxq_u8((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
64782   return __ret;
64783 }
64784 #else
vbcaxq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)64785 __ai __attribute__((target("sha3"))) uint8x16_t vbcaxq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
64786   uint8x16_t __ret;
64787   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64788   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64789   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64790   __ret = (uint8x16_t) __builtin_neon_vbcaxq_u8((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
64791   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64792   return __ret;
64793 }
64794 #endif
64795 
64796 #ifdef __LITTLE_ENDIAN__
vbcaxq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)64797 __ai __attribute__((target("sha3"))) uint32x4_t vbcaxq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
64798   uint32x4_t __ret;
64799   __ret = (uint32x4_t) __builtin_neon_vbcaxq_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
64800   return __ret;
64801 }
64802 #else
vbcaxq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)64803 __ai __attribute__((target("sha3"))) uint32x4_t vbcaxq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
64804   uint32x4_t __ret;
64805   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64806   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64807   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
64808   __ret = (uint32x4_t) __builtin_neon_vbcaxq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
64809   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64810   return __ret;
64811 }
64812 #endif
64813 
64814 #ifdef __LITTLE_ENDIAN__
vbcaxq_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)64815 __ai __attribute__((target("sha3"))) uint64x2_t vbcaxq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
64816   uint64x2_t __ret;
64817   __ret = (uint64x2_t) __builtin_neon_vbcaxq_u64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
64818   return __ret;
64819 }
64820 #else
vbcaxq_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)64821 __ai __attribute__((target("sha3"))) uint64x2_t vbcaxq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
64822   uint64x2_t __ret;
64823   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64824   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64825   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
64826   __ret = (uint64x2_t) __builtin_neon_vbcaxq_u64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
64827   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64828   return __ret;
64829 }
64830 #endif
64831 
64832 #ifdef __LITTLE_ENDIAN__
vbcaxq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)64833 __ai __attribute__((target("sha3"))) uint16x8_t vbcaxq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
64834   uint16x8_t __ret;
64835   __ret = (uint16x8_t) __builtin_neon_vbcaxq_u16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49);
64836   return __ret;
64837 }
64838 #else
vbcaxq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)64839 __ai __attribute__((target("sha3"))) uint16x8_t vbcaxq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
64840   uint16x8_t __ret;
64841   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64842   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64843   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
64844   __ret = (uint16x8_t) __builtin_neon_vbcaxq_u16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49);
64845   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64846   return __ret;
64847 }
64848 #endif
64849 
64850 #ifdef __LITTLE_ENDIAN__
vbcaxq_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)64851 __ai __attribute__((target("sha3"))) int8x16_t vbcaxq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
64852   int8x16_t __ret;
64853   __ret = (int8x16_t) __builtin_neon_vbcaxq_s8((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
64854   return __ret;
64855 }
64856 #else
vbcaxq_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)64857 __ai __attribute__((target("sha3"))) int8x16_t vbcaxq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
64858   int8x16_t __ret;
64859   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64860   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64861   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64862   __ret = (int8x16_t) __builtin_neon_vbcaxq_s8((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
64863   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64864   return __ret;
64865 }
64866 #endif
64867 
64868 #ifdef __LITTLE_ENDIAN__
vbcaxq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)64869 __ai __attribute__((target("sha3"))) int32x4_t vbcaxq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
64870   int32x4_t __ret;
64871   __ret = (int32x4_t) __builtin_neon_vbcaxq_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
64872   return __ret;
64873 }
64874 #else
vbcaxq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)64875 __ai __attribute__((target("sha3"))) int32x4_t vbcaxq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
64876   int32x4_t __ret;
64877   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64878   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64879   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
64880   __ret = (int32x4_t) __builtin_neon_vbcaxq_s32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
64881   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64882   return __ret;
64883 }
64884 #endif
64885 
64886 #ifdef __LITTLE_ENDIAN__
vbcaxq_s64(int64x2_t __p0,int64x2_t __p1,int64x2_t __p2)64887 __ai __attribute__((target("sha3"))) int64x2_t vbcaxq_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
64888   int64x2_t __ret;
64889   __ret = (int64x2_t) __builtin_neon_vbcaxq_s64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35);
64890   return __ret;
64891 }
64892 #else
vbcaxq_s64(int64x2_t __p0,int64x2_t __p1,int64x2_t __p2)64893 __ai __attribute__((target("sha3"))) int64x2_t vbcaxq_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
64894   int64x2_t __ret;
64895   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64896   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64897   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
64898   __ret = (int64x2_t) __builtin_neon_vbcaxq_s64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35);
64899   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64900   return __ret;
64901 }
64902 #endif
64903 
64904 #ifdef __LITTLE_ENDIAN__
vbcaxq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)64905 __ai __attribute__((target("sha3"))) int16x8_t vbcaxq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
64906   int16x8_t __ret;
64907   __ret = (int16x8_t) __builtin_neon_vbcaxq_s16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
64908   return __ret;
64909 }
64910 #else
vbcaxq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)64911 __ai __attribute__((target("sha3"))) int16x8_t vbcaxq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
64912   int16x8_t __ret;
64913   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64914   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64915   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
64916   __ret = (int16x8_t) __builtin_neon_vbcaxq_s16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
64917   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64918   return __ret;
64919 }
64920 #endif
64921 
64922 #ifdef __LITTLE_ENDIAN__
veor3q_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)64923 __ai __attribute__((target("sha3"))) uint8x16_t veor3q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
64924   uint8x16_t __ret;
64925   __ret = (uint8x16_t) __builtin_neon_veor3q_u8((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
64926   return __ret;
64927 }
64928 #else
veor3q_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)64929 __ai __attribute__((target("sha3"))) uint8x16_t veor3q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
64930   uint8x16_t __ret;
64931   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64932   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64933   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64934   __ret = (uint8x16_t) __builtin_neon_veor3q_u8((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
64935   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64936   return __ret;
64937 }
64938 #endif
64939 
64940 #ifdef __LITTLE_ENDIAN__
veor3q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)64941 __ai __attribute__((target("sha3"))) uint32x4_t veor3q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
64942   uint32x4_t __ret;
64943   __ret = (uint32x4_t) __builtin_neon_veor3q_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
64944   return __ret;
64945 }
64946 #else
veor3q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)64947 __ai __attribute__((target("sha3"))) uint32x4_t veor3q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
64948   uint32x4_t __ret;
64949   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64950   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64951   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
64952   __ret = (uint32x4_t) __builtin_neon_veor3q_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
64953   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64954   return __ret;
64955 }
64956 #endif
64957 
64958 #ifdef __LITTLE_ENDIAN__
veor3q_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)64959 __ai __attribute__((target("sha3"))) uint64x2_t veor3q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
64960   uint64x2_t __ret;
64961   __ret = (uint64x2_t) __builtin_neon_veor3q_u64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
64962   return __ret;
64963 }
64964 #else
veor3q_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)64965 __ai __attribute__((target("sha3"))) uint64x2_t veor3q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
64966   uint64x2_t __ret;
64967   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64968   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64969   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
64970   __ret = (uint64x2_t) __builtin_neon_veor3q_u64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
64971   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64972   return __ret;
64973 }
64974 #endif
64975 
64976 #ifdef __LITTLE_ENDIAN__
veor3q_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)64977 __ai __attribute__((target("sha3"))) uint16x8_t veor3q_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
64978   uint16x8_t __ret;
64979   __ret = (uint16x8_t) __builtin_neon_veor3q_u16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49);
64980   return __ret;
64981 }
64982 #else
veor3q_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)64983 __ai __attribute__((target("sha3"))) uint16x8_t veor3q_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
64984   uint16x8_t __ret;
64985   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64986   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64987   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
64988   __ret = (uint16x8_t) __builtin_neon_veor3q_u16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49);
64989   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64990   return __ret;
64991 }
64992 #endif
64993 
64994 #ifdef __LITTLE_ENDIAN__
veor3q_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)64995 __ai __attribute__((target("sha3"))) int8x16_t veor3q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
64996   int8x16_t __ret;
64997   __ret = (int8x16_t) __builtin_neon_veor3q_s8((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
64998   return __ret;
64999 }
65000 #else
veor3q_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)65001 __ai __attribute__((target("sha3"))) int8x16_t veor3q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
65002   int8x16_t __ret;
65003   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65004   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65005   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65006   __ret = (int8x16_t) __builtin_neon_veor3q_s8((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
65007   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65008   return __ret;
65009 }
65010 #endif
65011 
65012 #ifdef __LITTLE_ENDIAN__
veor3q_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)65013 __ai __attribute__((target("sha3"))) int32x4_t veor3q_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
65014   int32x4_t __ret;
65015   __ret = (int32x4_t) __builtin_neon_veor3q_s32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
65016   return __ret;
65017 }
65018 #else
veor3q_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)65019 __ai __attribute__((target("sha3"))) int32x4_t veor3q_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
65020   int32x4_t __ret;
65021   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65022   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65023   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
65024   __ret = (int32x4_t) __builtin_neon_veor3q_s32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
65025   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65026   return __ret;
65027 }
65028 #endif
65029 
65030 #ifdef __LITTLE_ENDIAN__
veor3q_s64(int64x2_t __p0,int64x2_t __p1,int64x2_t __p2)65031 __ai __attribute__((target("sha3"))) int64x2_t veor3q_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
65032   int64x2_t __ret;
65033   __ret = (int64x2_t) __builtin_neon_veor3q_s64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35);
65034   return __ret;
65035 }
65036 #else
veor3q_s64(int64x2_t __p0,int64x2_t __p1,int64x2_t __p2)65037 __ai __attribute__((target("sha3"))) int64x2_t veor3q_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
65038   int64x2_t __ret;
65039   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65040   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65041   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
65042   __ret = (int64x2_t) __builtin_neon_veor3q_s64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35);
65043   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65044   return __ret;
65045 }
65046 #endif
65047 
65048 #ifdef __LITTLE_ENDIAN__
veor3q_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)65049 __ai __attribute__((target("sha3"))) int16x8_t veor3q_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
65050   int16x8_t __ret;
65051   __ret = (int16x8_t) __builtin_neon_veor3q_s16((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
65052   return __ret;
65053 }
65054 #else
veor3q_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)65055 __ai __attribute__((target("sha3"))) int16x8_t veor3q_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
65056   int16x8_t __ret;
65057   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65058   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65059   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
65060   __ret = (int16x8_t) __builtin_neon_veor3q_s16((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
65061   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65062   return __ret;
65063 }
65064 #endif
65065 
65066 #ifdef __LITTLE_ENDIAN__
vrax1q_u64(uint64x2_t __p0,uint64x2_t __p1)65067 __ai __attribute__((target("sha3"))) uint64x2_t vrax1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65068   uint64x2_t __ret;
65069   __ret = (uint64x2_t) __builtin_neon_vrax1q_u64((int8x16_t)__p0, (int8x16_t)__p1, 51);
65070   return __ret;
65071 }
65072 #else
vrax1q_u64(uint64x2_t __p0,uint64x2_t __p1)65073 __ai __attribute__((target("sha3"))) uint64x2_t vrax1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65074   uint64x2_t __ret;
65075   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65076   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65077   __ret = (uint64x2_t) __builtin_neon_vrax1q_u64((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
65078   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65079   return __ret;
65080 }
65081 #endif
65082 
65083 #ifdef __LITTLE_ENDIAN__
vsha512hq_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)65084 __ai __attribute__((target("sha3"))) uint64x2_t vsha512hq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
65085   uint64x2_t __ret;
65086   __ret = (uint64x2_t) __builtin_neon_vsha512hq_u64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
65087   return __ret;
65088 }
65089 #else
vsha512hq_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)65090 __ai __attribute__((target("sha3"))) uint64x2_t vsha512hq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
65091   uint64x2_t __ret;
65092   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65093   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65094   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
65095   __ret = (uint64x2_t) __builtin_neon_vsha512hq_u64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
65096   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65097   return __ret;
65098 }
65099 #endif
65100 
65101 #ifdef __LITTLE_ENDIAN__
vsha512h2q_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)65102 __ai __attribute__((target("sha3"))) uint64x2_t vsha512h2q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
65103   uint64x2_t __ret;
65104   __ret = (uint64x2_t) __builtin_neon_vsha512h2q_u64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
65105   return __ret;
65106 }
65107 #else
vsha512h2q_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)65108 __ai __attribute__((target("sha3"))) uint64x2_t vsha512h2q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
65109   uint64x2_t __ret;
65110   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65111   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65112   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
65113   __ret = (uint64x2_t) __builtin_neon_vsha512h2q_u64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
65114   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65115   return __ret;
65116 }
65117 #endif
65118 
65119 #ifdef __LITTLE_ENDIAN__
vsha512su0q_u64(uint64x2_t __p0,uint64x2_t __p1)65120 __ai __attribute__((target("sha3"))) uint64x2_t vsha512su0q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65121   uint64x2_t __ret;
65122   __ret = (uint64x2_t) __builtin_neon_vsha512su0q_u64((int8x16_t)__p0, (int8x16_t)__p1, 51);
65123   return __ret;
65124 }
65125 #else
vsha512su0q_u64(uint64x2_t __p0,uint64x2_t __p1)65126 __ai __attribute__((target("sha3"))) uint64x2_t vsha512su0q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65127   uint64x2_t __ret;
65128   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65129   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65130   __ret = (uint64x2_t) __builtin_neon_vsha512su0q_u64((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
65131   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65132   return __ret;
65133 }
65134 #endif
65135 
65136 #ifdef __LITTLE_ENDIAN__
vsha512su1q_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)65137 __ai __attribute__((target("sha3"))) uint64x2_t vsha512su1q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
65138   uint64x2_t __ret;
65139   __ret = (uint64x2_t) __builtin_neon_vsha512su1q_u64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
65140   return __ret;
65141 }
65142 #else
vsha512su1q_u64(uint64x2_t __p0,uint64x2_t __p1,uint64x2_t __p2)65143 __ai __attribute__((target("sha3"))) uint64x2_t vsha512su1q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
65144   uint64x2_t __ret;
65145   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65146   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65147   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
65148   __ret = (uint64x2_t) __builtin_neon_vsha512su1q_u64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
65149   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65150   return __ret;
65151 }
65152 #endif
65153 
65154 #ifdef __LITTLE_ENDIAN__
65155 #define vxarq_u64(__p0, __p1, __p2) __extension__ ({ \
65156   uint64x2_t __ret; \
65157   uint64x2_t __s0 = __p0; \
65158   uint64x2_t __s1 = __p1; \
65159   __ret = (uint64x2_t) __builtin_neon_vxarq_u64((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
65160   __ret; \
65161 })
65162 #else
65163 #define vxarq_u64(__p0, __p1, __p2) __extension__ ({ \
65164   uint64x2_t __ret; \
65165   uint64x2_t __s0 = __p0; \
65166   uint64x2_t __s1 = __p1; \
65167   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
65168   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
65169   __ret = (uint64x2_t) __builtin_neon_vxarq_u64((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
65170   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
65171   __ret; \
65172 })
65173 #endif
65174 
65175 #ifdef __LITTLE_ENDIAN__
vsm3partw1q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)65176 __ai __attribute__((target("sm4"))) uint32x4_t vsm3partw1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
65177   uint32x4_t __ret;
65178   __ret = (uint32x4_t) __builtin_neon_vsm3partw1q_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
65179   return __ret;
65180 }
65181 #else
vsm3partw1q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)65182 __ai __attribute__((target("sm4"))) uint32x4_t vsm3partw1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
65183   uint32x4_t __ret;
65184   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65185   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65186   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
65187   __ret = (uint32x4_t) __builtin_neon_vsm3partw1q_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
65188   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65189   return __ret;
65190 }
65191 #endif
65192 
65193 #ifdef __LITTLE_ENDIAN__
vsm3partw2q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)65194 __ai __attribute__((target("sm4"))) uint32x4_t vsm3partw2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
65195   uint32x4_t __ret;
65196   __ret = (uint32x4_t) __builtin_neon_vsm3partw2q_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
65197   return __ret;
65198 }
65199 #else
vsm3partw2q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)65200 __ai __attribute__((target("sm4"))) uint32x4_t vsm3partw2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
65201   uint32x4_t __ret;
65202   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65203   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65204   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
65205   __ret = (uint32x4_t) __builtin_neon_vsm3partw2q_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
65206   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65207   return __ret;
65208 }
65209 #endif
65210 
65211 #ifdef __LITTLE_ENDIAN__
vsm3ss1q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)65212 __ai __attribute__((target("sm4"))) uint32x4_t vsm3ss1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
65213   uint32x4_t __ret;
65214   __ret = (uint32x4_t) __builtin_neon_vsm3ss1q_u32((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
65215   return __ret;
65216 }
65217 #else
vsm3ss1q_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)65218 __ai __attribute__((target("sm4"))) uint32x4_t vsm3ss1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
65219   uint32x4_t __ret;
65220   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65221   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65222   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
65223   __ret = (uint32x4_t) __builtin_neon_vsm3ss1q_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
65224   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65225   return __ret;
65226 }
65227 #endif
65228 
65229 #ifdef __LITTLE_ENDIAN__
65230 #define vsm3tt1aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
65231   uint32x4_t __ret; \
65232   uint32x4_t __s0 = __p0; \
65233   uint32x4_t __s1 = __p1; \
65234   uint32x4_t __s2 = __p2; \
65235   __ret = (uint32x4_t) __builtin_neon_vsm3tt1aq_u32((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \
65236   __ret; \
65237 })
65238 #else
65239 #define vsm3tt1aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
65240   uint32x4_t __ret; \
65241   uint32x4_t __s0 = __p0; \
65242   uint32x4_t __s1 = __p1; \
65243   uint32x4_t __s2 = __p2; \
65244   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
65245   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
65246   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
65247   __ret = (uint32x4_t) __builtin_neon_vsm3tt1aq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \
65248   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
65249   __ret; \
65250 })
65251 #endif
65252 
65253 #ifdef __LITTLE_ENDIAN__
65254 #define vsm3tt1bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
65255   uint32x4_t __ret; \
65256   uint32x4_t __s0 = __p0; \
65257   uint32x4_t __s1 = __p1; \
65258   uint32x4_t __s2 = __p2; \
65259   __ret = (uint32x4_t) __builtin_neon_vsm3tt1bq_u32((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \
65260   __ret; \
65261 })
65262 #else
65263 #define vsm3tt1bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
65264   uint32x4_t __ret; \
65265   uint32x4_t __s0 = __p0; \
65266   uint32x4_t __s1 = __p1; \
65267   uint32x4_t __s2 = __p2; \
65268   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
65269   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
65270   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
65271   __ret = (uint32x4_t) __builtin_neon_vsm3tt1bq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \
65272   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
65273   __ret; \
65274 })
65275 #endif
65276 
65277 #ifdef __LITTLE_ENDIAN__
65278 #define vsm3tt2aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
65279   uint32x4_t __ret; \
65280   uint32x4_t __s0 = __p0; \
65281   uint32x4_t __s1 = __p1; \
65282   uint32x4_t __s2 = __p2; \
65283   __ret = (uint32x4_t) __builtin_neon_vsm3tt2aq_u32((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \
65284   __ret; \
65285 })
65286 #else
65287 #define vsm3tt2aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
65288   uint32x4_t __ret; \
65289   uint32x4_t __s0 = __p0; \
65290   uint32x4_t __s1 = __p1; \
65291   uint32x4_t __s2 = __p2; \
65292   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
65293   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
65294   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
65295   __ret = (uint32x4_t) __builtin_neon_vsm3tt2aq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \
65296   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
65297   __ret; \
65298 })
65299 #endif
65300 
65301 #ifdef __LITTLE_ENDIAN__
65302 #define vsm3tt2bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
65303   uint32x4_t __ret; \
65304   uint32x4_t __s0 = __p0; \
65305   uint32x4_t __s1 = __p1; \
65306   uint32x4_t __s2 = __p2; \
65307   __ret = (uint32x4_t) __builtin_neon_vsm3tt2bq_u32((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \
65308   __ret; \
65309 })
65310 #else
65311 #define vsm3tt2bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
65312   uint32x4_t __ret; \
65313   uint32x4_t __s0 = __p0; \
65314   uint32x4_t __s1 = __p1; \
65315   uint32x4_t __s2 = __p2; \
65316   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
65317   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
65318   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
65319   __ret = (uint32x4_t) __builtin_neon_vsm3tt2bq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \
65320   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
65321   __ret; \
65322 })
65323 #endif
65324 
65325 #ifdef __LITTLE_ENDIAN__
vsm4eq_u32(uint32x4_t __p0,uint32x4_t __p1)65326 __ai __attribute__((target("sm4"))) uint32x4_t vsm4eq_u32(uint32x4_t __p0, uint32x4_t __p1) {
65327   uint32x4_t __ret;
65328   __ret = (uint32x4_t) __builtin_neon_vsm4eq_u32((int8x16_t)__p0, (int8x16_t)__p1, 50);
65329   return __ret;
65330 }
65331 #else
vsm4eq_u32(uint32x4_t __p0,uint32x4_t __p1)65332 __ai __attribute__((target("sm4"))) uint32x4_t vsm4eq_u32(uint32x4_t __p0, uint32x4_t __p1) {
65333   uint32x4_t __ret;
65334   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65335   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65336   __ret = (uint32x4_t) __builtin_neon_vsm4eq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
65337   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65338   return __ret;
65339 }
65340 #endif
65341 
65342 #ifdef __LITTLE_ENDIAN__
vsm4ekeyq_u32(uint32x4_t __p0,uint32x4_t __p1)65343 __ai __attribute__((target("sm4"))) uint32x4_t vsm4ekeyq_u32(uint32x4_t __p0, uint32x4_t __p1) {
65344   uint32x4_t __ret;
65345   __ret = (uint32x4_t) __builtin_neon_vsm4ekeyq_u32((int8x16_t)__p0, (int8x16_t)__p1, 50);
65346   return __ret;
65347 }
65348 #else
vsm4ekeyq_u32(uint32x4_t __p0,uint32x4_t __p1)65349 __ai __attribute__((target("sm4"))) uint32x4_t vsm4ekeyq_u32(uint32x4_t __p0, uint32x4_t __p1) {
65350   uint32x4_t __ret;
65351   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65352   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65353   __ret = (uint32x4_t) __builtin_neon_vsm4ekeyq_u32((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
65354   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65355   return __ret;
65356 }
65357 #endif
65358 
vqrdmlahs_s32(int32_t __p0,int32_t __p1,int32_t __p2)65359 __ai __attribute__((target("v8.1a"))) int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
65360   int32_t __ret;
65361   __ret = (int32_t) __builtin_neon_vqrdmlahs_s32(__p0, __p1, __p2);
65362   return __ret;
65363 }
vqrdmlahh_s16(int16_t __p0,int16_t __p1,int16_t __p2)65364 __ai __attribute__((target("v8.1a"))) int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
65365   int16_t __ret;
65366   __ret = (int16_t) __builtin_neon_vqrdmlahh_s16(__p0, __p1, __p2);
65367   return __ret;
65368 }
65369 #ifdef __LITTLE_ENDIAN__
65370 #define vqrdmlahs_lane_s32(__p0_787, __p1_787, __p2_787, __p3_787) __extension__ ({ \
65371   int32_t __ret_787; \
65372   int32_t __s0_787 = __p0_787; \
65373   int32_t __s1_787 = __p1_787; \
65374   int32x2_t __s2_787 = __p2_787; \
65375   __ret_787 = vqrdmlahs_s32(__s0_787, __s1_787, vget_lane_s32(__s2_787, __p3_787)); \
65376   __ret_787; \
65377 })
65378 #else
65379 #define vqrdmlahs_lane_s32(__p0_788, __p1_788, __p2_788, __p3_788) __extension__ ({ \
65380   int32_t __ret_788; \
65381   int32_t __s0_788 = __p0_788; \
65382   int32_t __s1_788 = __p1_788; \
65383   int32x2_t __s2_788 = __p2_788; \
65384   int32x2_t __rev2_788;  __rev2_788 = __builtin_shufflevector(__s2_788, __s2_788, 1, 0); \
65385   __ret_788 = vqrdmlahs_s32(__s0_788, __s1_788, __noswap_vget_lane_s32(__rev2_788, __p3_788)); \
65386   __ret_788; \
65387 })
65388 #endif
65389 
65390 #ifdef __LITTLE_ENDIAN__
65391 #define vqrdmlahh_lane_s16(__p0_789, __p1_789, __p2_789, __p3_789) __extension__ ({ \
65392   int16_t __ret_789; \
65393   int16_t __s0_789 = __p0_789; \
65394   int16_t __s1_789 = __p1_789; \
65395   int16x4_t __s2_789 = __p2_789; \
65396   __ret_789 = vqrdmlahh_s16(__s0_789, __s1_789, vget_lane_s16(__s2_789, __p3_789)); \
65397   __ret_789; \
65398 })
65399 #else
65400 #define vqrdmlahh_lane_s16(__p0_790, __p1_790, __p2_790, __p3_790) __extension__ ({ \
65401   int16_t __ret_790; \
65402   int16_t __s0_790 = __p0_790; \
65403   int16_t __s1_790 = __p1_790; \
65404   int16x4_t __s2_790 = __p2_790; \
65405   int16x4_t __rev2_790;  __rev2_790 = __builtin_shufflevector(__s2_790, __s2_790, 3, 2, 1, 0); \
65406   __ret_790 = vqrdmlahh_s16(__s0_790, __s1_790, __noswap_vget_lane_s16(__rev2_790, __p3_790)); \
65407   __ret_790; \
65408 })
65409 #endif
65410 
65411 #ifdef __LITTLE_ENDIAN__
65412 #define vqrdmlahs_laneq_s32(__p0_791, __p1_791, __p2_791, __p3_791) __extension__ ({ \
65413   int32_t __ret_791; \
65414   int32_t __s0_791 = __p0_791; \
65415   int32_t __s1_791 = __p1_791; \
65416   int32x4_t __s2_791 = __p2_791; \
65417   __ret_791 = vqrdmlahs_s32(__s0_791, __s1_791, vgetq_lane_s32(__s2_791, __p3_791)); \
65418   __ret_791; \
65419 })
65420 #else
65421 #define vqrdmlahs_laneq_s32(__p0_792, __p1_792, __p2_792, __p3_792) __extension__ ({ \
65422   int32_t __ret_792; \
65423   int32_t __s0_792 = __p0_792; \
65424   int32_t __s1_792 = __p1_792; \
65425   int32x4_t __s2_792 = __p2_792; \
65426   int32x4_t __rev2_792;  __rev2_792 = __builtin_shufflevector(__s2_792, __s2_792, 3, 2, 1, 0); \
65427   __ret_792 = vqrdmlahs_s32(__s0_792, __s1_792, __noswap_vgetq_lane_s32(__rev2_792, __p3_792)); \
65428   __ret_792; \
65429 })
65430 #endif
65431 
65432 #ifdef __LITTLE_ENDIAN__
65433 #define vqrdmlahh_laneq_s16(__p0_793, __p1_793, __p2_793, __p3_793) __extension__ ({ \
65434   int16_t __ret_793; \
65435   int16_t __s0_793 = __p0_793; \
65436   int16_t __s1_793 = __p1_793; \
65437   int16x8_t __s2_793 = __p2_793; \
65438   __ret_793 = vqrdmlahh_s16(__s0_793, __s1_793, vgetq_lane_s16(__s2_793, __p3_793)); \
65439   __ret_793; \
65440 })
65441 #else
65442 #define vqrdmlahh_laneq_s16(__p0_794, __p1_794, __p2_794, __p3_794) __extension__ ({ \
65443   int16_t __ret_794; \
65444   int16_t __s0_794 = __p0_794; \
65445   int16_t __s1_794 = __p1_794; \
65446   int16x8_t __s2_794 = __p2_794; \
65447   int16x8_t __rev2_794;  __rev2_794 = __builtin_shufflevector(__s2_794, __s2_794, 7, 6, 5, 4, 3, 2, 1, 0); \
65448   __ret_794 = vqrdmlahh_s16(__s0_794, __s1_794, __noswap_vgetq_lane_s16(__rev2_794, __p3_794)); \
65449   __ret_794; \
65450 })
65451 #endif
65452 
65453 #ifdef __LITTLE_ENDIAN__
65454 #define vqrdmlahq_laneq_s32(__p0_795, __p1_795, __p2_795, __p3_795) __extension__ ({ \
65455   int32x4_t __ret_795; \
65456   int32x4_t __s0_795 = __p0_795; \
65457   int32x4_t __s1_795 = __p1_795; \
65458   int32x4_t __s2_795 = __p2_795; \
65459   __ret_795 = vqrdmlahq_s32(__s0_795, __s1_795, splatq_laneq_s32(__s2_795, __p3_795)); \
65460   __ret_795; \
65461 })
65462 #else
65463 #define vqrdmlahq_laneq_s32(__p0_796, __p1_796, __p2_796, __p3_796) __extension__ ({ \
65464   int32x4_t __ret_796; \
65465   int32x4_t __s0_796 = __p0_796; \
65466   int32x4_t __s1_796 = __p1_796; \
65467   int32x4_t __s2_796 = __p2_796; \
65468   int32x4_t __rev0_796;  __rev0_796 = __builtin_shufflevector(__s0_796, __s0_796, 3, 2, 1, 0); \
65469   int32x4_t __rev1_796;  __rev1_796 = __builtin_shufflevector(__s1_796, __s1_796, 3, 2, 1, 0); \
65470   int32x4_t __rev2_796;  __rev2_796 = __builtin_shufflevector(__s2_796, __s2_796, 3, 2, 1, 0); \
65471   __ret_796 = __noswap_vqrdmlahq_s32(__rev0_796, __rev1_796, __noswap_splatq_laneq_s32(__rev2_796, __p3_796)); \
65472   __ret_796 = __builtin_shufflevector(__ret_796, __ret_796, 3, 2, 1, 0); \
65473   __ret_796; \
65474 })
65475 #endif
65476 
65477 #ifdef __LITTLE_ENDIAN__
65478 #define vqrdmlahq_laneq_s16(__p0_797, __p1_797, __p2_797, __p3_797) __extension__ ({ \
65479   int16x8_t __ret_797; \
65480   int16x8_t __s0_797 = __p0_797; \
65481   int16x8_t __s1_797 = __p1_797; \
65482   int16x8_t __s2_797 = __p2_797; \
65483   __ret_797 = vqrdmlahq_s16(__s0_797, __s1_797, splatq_laneq_s16(__s2_797, __p3_797)); \
65484   __ret_797; \
65485 })
65486 #else
65487 #define vqrdmlahq_laneq_s16(__p0_798, __p1_798, __p2_798, __p3_798) __extension__ ({ \
65488   int16x8_t __ret_798; \
65489   int16x8_t __s0_798 = __p0_798; \
65490   int16x8_t __s1_798 = __p1_798; \
65491   int16x8_t __s2_798 = __p2_798; \
65492   int16x8_t __rev0_798;  __rev0_798 = __builtin_shufflevector(__s0_798, __s0_798, 7, 6, 5, 4, 3, 2, 1, 0); \
65493   int16x8_t __rev1_798;  __rev1_798 = __builtin_shufflevector(__s1_798, __s1_798, 7, 6, 5, 4, 3, 2, 1, 0); \
65494   int16x8_t __rev2_798;  __rev2_798 = __builtin_shufflevector(__s2_798, __s2_798, 7, 6, 5, 4, 3, 2, 1, 0); \
65495   __ret_798 = __noswap_vqrdmlahq_s16(__rev0_798, __rev1_798, __noswap_splatq_laneq_s16(__rev2_798, __p3_798)); \
65496   __ret_798 = __builtin_shufflevector(__ret_798, __ret_798, 7, 6, 5, 4, 3, 2, 1, 0); \
65497   __ret_798; \
65498 })
65499 #endif
65500 
65501 #ifdef __LITTLE_ENDIAN__
65502 #define vqrdmlah_laneq_s32(__p0_799, __p1_799, __p2_799, __p3_799) __extension__ ({ \
65503   int32x2_t __ret_799; \
65504   int32x2_t __s0_799 = __p0_799; \
65505   int32x2_t __s1_799 = __p1_799; \
65506   int32x4_t __s2_799 = __p2_799; \
65507   __ret_799 = vqrdmlah_s32(__s0_799, __s1_799, splat_laneq_s32(__s2_799, __p3_799)); \
65508   __ret_799; \
65509 })
65510 #else
65511 #define vqrdmlah_laneq_s32(__p0_800, __p1_800, __p2_800, __p3_800) __extension__ ({ \
65512   int32x2_t __ret_800; \
65513   int32x2_t __s0_800 = __p0_800; \
65514   int32x2_t __s1_800 = __p1_800; \
65515   int32x4_t __s2_800 = __p2_800; \
65516   int32x2_t __rev0_800;  __rev0_800 = __builtin_shufflevector(__s0_800, __s0_800, 1, 0); \
65517   int32x2_t __rev1_800;  __rev1_800 = __builtin_shufflevector(__s1_800, __s1_800, 1, 0); \
65518   int32x4_t __rev2_800;  __rev2_800 = __builtin_shufflevector(__s2_800, __s2_800, 3, 2, 1, 0); \
65519   __ret_800 = __noswap_vqrdmlah_s32(__rev0_800, __rev1_800, __noswap_splat_laneq_s32(__rev2_800, __p3_800)); \
65520   __ret_800 = __builtin_shufflevector(__ret_800, __ret_800, 1, 0); \
65521   __ret_800; \
65522 })
65523 #endif
65524 
65525 #ifdef __LITTLE_ENDIAN__
65526 #define vqrdmlah_laneq_s16(__p0_801, __p1_801, __p2_801, __p3_801) __extension__ ({ \
65527   int16x4_t __ret_801; \
65528   int16x4_t __s0_801 = __p0_801; \
65529   int16x4_t __s1_801 = __p1_801; \
65530   int16x8_t __s2_801 = __p2_801; \
65531   __ret_801 = vqrdmlah_s16(__s0_801, __s1_801, splat_laneq_s16(__s2_801, __p3_801)); \
65532   __ret_801; \
65533 })
65534 #else
65535 #define vqrdmlah_laneq_s16(__p0_802, __p1_802, __p2_802, __p3_802) __extension__ ({ \
65536   int16x4_t __ret_802; \
65537   int16x4_t __s0_802 = __p0_802; \
65538   int16x4_t __s1_802 = __p1_802; \
65539   int16x8_t __s2_802 = __p2_802; \
65540   int16x4_t __rev0_802;  __rev0_802 = __builtin_shufflevector(__s0_802, __s0_802, 3, 2, 1, 0); \
65541   int16x4_t __rev1_802;  __rev1_802 = __builtin_shufflevector(__s1_802, __s1_802, 3, 2, 1, 0); \
65542   int16x8_t __rev2_802;  __rev2_802 = __builtin_shufflevector(__s2_802, __s2_802, 7, 6, 5, 4, 3, 2, 1, 0); \
65543   __ret_802 = __noswap_vqrdmlah_s16(__rev0_802, __rev1_802, __noswap_splat_laneq_s16(__rev2_802, __p3_802)); \
65544   __ret_802 = __builtin_shufflevector(__ret_802, __ret_802, 3, 2, 1, 0); \
65545   __ret_802; \
65546 })
65547 #endif
65548 
vqrdmlshs_s32(int32_t __p0,int32_t __p1,int32_t __p2)65549 __ai __attribute__((target("v8.1a"))) int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
65550   int32_t __ret;
65551   __ret = (int32_t) __builtin_neon_vqrdmlshs_s32(__p0, __p1, __p2);
65552   return __ret;
65553 }
vqrdmlshh_s16(int16_t __p0,int16_t __p1,int16_t __p2)65554 __ai __attribute__((target("v8.1a"))) int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
65555   int16_t __ret;
65556   __ret = (int16_t) __builtin_neon_vqrdmlshh_s16(__p0, __p1, __p2);
65557   return __ret;
65558 }
65559 #ifdef __LITTLE_ENDIAN__
65560 #define vqrdmlshs_lane_s32(__p0_803, __p1_803, __p2_803, __p3_803) __extension__ ({ \
65561   int32_t __ret_803; \
65562   int32_t __s0_803 = __p0_803; \
65563   int32_t __s1_803 = __p1_803; \
65564   int32x2_t __s2_803 = __p2_803; \
65565   __ret_803 = vqrdmlshs_s32(__s0_803, __s1_803, vget_lane_s32(__s2_803, __p3_803)); \
65566   __ret_803; \
65567 })
65568 #else
65569 #define vqrdmlshs_lane_s32(__p0_804, __p1_804, __p2_804, __p3_804) __extension__ ({ \
65570   int32_t __ret_804; \
65571   int32_t __s0_804 = __p0_804; \
65572   int32_t __s1_804 = __p1_804; \
65573   int32x2_t __s2_804 = __p2_804; \
65574   int32x2_t __rev2_804;  __rev2_804 = __builtin_shufflevector(__s2_804, __s2_804, 1, 0); \
65575   __ret_804 = vqrdmlshs_s32(__s0_804, __s1_804, __noswap_vget_lane_s32(__rev2_804, __p3_804)); \
65576   __ret_804; \
65577 })
65578 #endif
65579 
65580 #ifdef __LITTLE_ENDIAN__
65581 #define vqrdmlshh_lane_s16(__p0_805, __p1_805, __p2_805, __p3_805) __extension__ ({ \
65582   int16_t __ret_805; \
65583   int16_t __s0_805 = __p0_805; \
65584   int16_t __s1_805 = __p1_805; \
65585   int16x4_t __s2_805 = __p2_805; \
65586   __ret_805 = vqrdmlshh_s16(__s0_805, __s1_805, vget_lane_s16(__s2_805, __p3_805)); \
65587   __ret_805; \
65588 })
65589 #else
65590 #define vqrdmlshh_lane_s16(__p0_806, __p1_806, __p2_806, __p3_806) __extension__ ({ \
65591   int16_t __ret_806; \
65592   int16_t __s0_806 = __p0_806; \
65593   int16_t __s1_806 = __p1_806; \
65594   int16x4_t __s2_806 = __p2_806; \
65595   int16x4_t __rev2_806;  __rev2_806 = __builtin_shufflevector(__s2_806, __s2_806, 3, 2, 1, 0); \
65596   __ret_806 = vqrdmlshh_s16(__s0_806, __s1_806, __noswap_vget_lane_s16(__rev2_806, __p3_806)); \
65597   __ret_806; \
65598 })
65599 #endif
65600 
65601 #ifdef __LITTLE_ENDIAN__
65602 #define vqrdmlshs_laneq_s32(__p0_807, __p1_807, __p2_807, __p3_807) __extension__ ({ \
65603   int32_t __ret_807; \
65604   int32_t __s0_807 = __p0_807; \
65605   int32_t __s1_807 = __p1_807; \
65606   int32x4_t __s2_807 = __p2_807; \
65607   __ret_807 = vqrdmlshs_s32(__s0_807, __s1_807, vgetq_lane_s32(__s2_807, __p3_807)); \
65608   __ret_807; \
65609 })
65610 #else
65611 #define vqrdmlshs_laneq_s32(__p0_808, __p1_808, __p2_808, __p3_808) __extension__ ({ \
65612   int32_t __ret_808; \
65613   int32_t __s0_808 = __p0_808; \
65614   int32_t __s1_808 = __p1_808; \
65615   int32x4_t __s2_808 = __p2_808; \
65616   int32x4_t __rev2_808;  __rev2_808 = __builtin_shufflevector(__s2_808, __s2_808, 3, 2, 1, 0); \
65617   __ret_808 = vqrdmlshs_s32(__s0_808, __s1_808, __noswap_vgetq_lane_s32(__rev2_808, __p3_808)); \
65618   __ret_808; \
65619 })
65620 #endif
65621 
65622 #ifdef __LITTLE_ENDIAN__
65623 #define vqrdmlshh_laneq_s16(__p0_809, __p1_809, __p2_809, __p3_809) __extension__ ({ \
65624   int16_t __ret_809; \
65625   int16_t __s0_809 = __p0_809; \
65626   int16_t __s1_809 = __p1_809; \
65627   int16x8_t __s2_809 = __p2_809; \
65628   __ret_809 = vqrdmlshh_s16(__s0_809, __s1_809, vgetq_lane_s16(__s2_809, __p3_809)); \
65629   __ret_809; \
65630 })
65631 #else
65632 #define vqrdmlshh_laneq_s16(__p0_810, __p1_810, __p2_810, __p3_810) __extension__ ({ \
65633   int16_t __ret_810; \
65634   int16_t __s0_810 = __p0_810; \
65635   int16_t __s1_810 = __p1_810; \
65636   int16x8_t __s2_810 = __p2_810; \
65637   int16x8_t __rev2_810;  __rev2_810 = __builtin_shufflevector(__s2_810, __s2_810, 7, 6, 5, 4, 3, 2, 1, 0); \
65638   __ret_810 = vqrdmlshh_s16(__s0_810, __s1_810, __noswap_vgetq_lane_s16(__rev2_810, __p3_810)); \
65639   __ret_810; \
65640 })
65641 #endif
65642 
65643 #ifdef __LITTLE_ENDIAN__
65644 #define vqrdmlshq_laneq_s32(__p0_811, __p1_811, __p2_811, __p3_811) __extension__ ({ \
65645   int32x4_t __ret_811; \
65646   int32x4_t __s0_811 = __p0_811; \
65647   int32x4_t __s1_811 = __p1_811; \
65648   int32x4_t __s2_811 = __p2_811; \
65649   __ret_811 = vqrdmlshq_s32(__s0_811, __s1_811, splatq_laneq_s32(__s2_811, __p3_811)); \
65650   __ret_811; \
65651 })
65652 #else
65653 #define vqrdmlshq_laneq_s32(__p0_812, __p1_812, __p2_812, __p3_812) __extension__ ({ \
65654   int32x4_t __ret_812; \
65655   int32x4_t __s0_812 = __p0_812; \
65656   int32x4_t __s1_812 = __p1_812; \
65657   int32x4_t __s2_812 = __p2_812; \
65658   int32x4_t __rev0_812;  __rev0_812 = __builtin_shufflevector(__s0_812, __s0_812, 3, 2, 1, 0); \
65659   int32x4_t __rev1_812;  __rev1_812 = __builtin_shufflevector(__s1_812, __s1_812, 3, 2, 1, 0); \
65660   int32x4_t __rev2_812;  __rev2_812 = __builtin_shufflevector(__s2_812, __s2_812, 3, 2, 1, 0); \
65661   __ret_812 = __noswap_vqrdmlshq_s32(__rev0_812, __rev1_812, __noswap_splatq_laneq_s32(__rev2_812, __p3_812)); \
65662   __ret_812 = __builtin_shufflevector(__ret_812, __ret_812, 3, 2, 1, 0); \
65663   __ret_812; \
65664 })
65665 #endif
65666 
65667 #ifdef __LITTLE_ENDIAN__
65668 #define vqrdmlshq_laneq_s16(__p0_813, __p1_813, __p2_813, __p3_813) __extension__ ({ \
65669   int16x8_t __ret_813; \
65670   int16x8_t __s0_813 = __p0_813; \
65671   int16x8_t __s1_813 = __p1_813; \
65672   int16x8_t __s2_813 = __p2_813; \
65673   __ret_813 = vqrdmlshq_s16(__s0_813, __s1_813, splatq_laneq_s16(__s2_813, __p3_813)); \
65674   __ret_813; \
65675 })
65676 #else
65677 #define vqrdmlshq_laneq_s16(__p0_814, __p1_814, __p2_814, __p3_814) __extension__ ({ \
65678   int16x8_t __ret_814; \
65679   int16x8_t __s0_814 = __p0_814; \
65680   int16x8_t __s1_814 = __p1_814; \
65681   int16x8_t __s2_814 = __p2_814; \
65682   int16x8_t __rev0_814;  __rev0_814 = __builtin_shufflevector(__s0_814, __s0_814, 7, 6, 5, 4, 3, 2, 1, 0); \
65683   int16x8_t __rev1_814;  __rev1_814 = __builtin_shufflevector(__s1_814, __s1_814, 7, 6, 5, 4, 3, 2, 1, 0); \
65684   int16x8_t __rev2_814;  __rev2_814 = __builtin_shufflevector(__s2_814, __s2_814, 7, 6, 5, 4, 3, 2, 1, 0); \
65685   __ret_814 = __noswap_vqrdmlshq_s16(__rev0_814, __rev1_814, __noswap_splatq_laneq_s16(__rev2_814, __p3_814)); \
65686   __ret_814 = __builtin_shufflevector(__ret_814, __ret_814, 7, 6, 5, 4, 3, 2, 1, 0); \
65687   __ret_814; \
65688 })
65689 #endif
65690 
65691 #ifdef __LITTLE_ENDIAN__
65692 #define vqrdmlsh_laneq_s32(__p0_815, __p1_815, __p2_815, __p3_815) __extension__ ({ \
65693   int32x2_t __ret_815; \
65694   int32x2_t __s0_815 = __p0_815; \
65695   int32x2_t __s1_815 = __p1_815; \
65696   int32x4_t __s2_815 = __p2_815; \
65697   __ret_815 = vqrdmlsh_s32(__s0_815, __s1_815, splat_laneq_s32(__s2_815, __p3_815)); \
65698   __ret_815; \
65699 })
65700 #else
65701 #define vqrdmlsh_laneq_s32(__p0_816, __p1_816, __p2_816, __p3_816) __extension__ ({ \
65702   int32x2_t __ret_816; \
65703   int32x2_t __s0_816 = __p0_816; \
65704   int32x2_t __s1_816 = __p1_816; \
65705   int32x4_t __s2_816 = __p2_816; \
65706   int32x2_t __rev0_816;  __rev0_816 = __builtin_shufflevector(__s0_816, __s0_816, 1, 0); \
65707   int32x2_t __rev1_816;  __rev1_816 = __builtin_shufflevector(__s1_816, __s1_816, 1, 0); \
65708   int32x4_t __rev2_816;  __rev2_816 = __builtin_shufflevector(__s2_816, __s2_816, 3, 2, 1, 0); \
65709   __ret_816 = __noswap_vqrdmlsh_s32(__rev0_816, __rev1_816, __noswap_splat_laneq_s32(__rev2_816, __p3_816)); \
65710   __ret_816 = __builtin_shufflevector(__ret_816, __ret_816, 1, 0); \
65711   __ret_816; \
65712 })
65713 #endif
65714 
65715 #ifdef __LITTLE_ENDIAN__
65716 #define vqrdmlsh_laneq_s16(__p0_817, __p1_817, __p2_817, __p3_817) __extension__ ({ \
65717   int16x4_t __ret_817; \
65718   int16x4_t __s0_817 = __p0_817; \
65719   int16x4_t __s1_817 = __p1_817; \
65720   int16x8_t __s2_817 = __p2_817; \
65721   __ret_817 = vqrdmlsh_s16(__s0_817, __s1_817, splat_laneq_s16(__s2_817, __p3_817)); \
65722   __ret_817; \
65723 })
65724 #else
65725 #define vqrdmlsh_laneq_s16(__p0_818, __p1_818, __p2_818, __p3_818) __extension__ ({ \
65726   int16x4_t __ret_818; \
65727   int16x4_t __s0_818 = __p0_818; \
65728   int16x4_t __s1_818 = __p1_818; \
65729   int16x8_t __s2_818 = __p2_818; \
65730   int16x4_t __rev0_818;  __rev0_818 = __builtin_shufflevector(__s0_818, __s0_818, 3, 2, 1, 0); \
65731   int16x4_t __rev1_818;  __rev1_818 = __builtin_shufflevector(__s1_818, __s1_818, 3, 2, 1, 0); \
65732   int16x8_t __rev2_818;  __rev2_818 = __builtin_shufflevector(__s2_818, __s2_818, 7, 6, 5, 4, 3, 2, 1, 0); \
65733   __ret_818 = __noswap_vqrdmlsh_s16(__rev0_818, __rev1_818, __noswap_splat_laneq_s16(__rev2_818, __p3_818)); \
65734   __ret_818 = __builtin_shufflevector(__ret_818, __ret_818, 3, 2, 1, 0); \
65735   __ret_818; \
65736 })
65737 #endif
65738 
65739 #ifdef __LITTLE_ENDIAN__
vcaddq_rot270_f64(float64x2_t __p0,float64x2_t __p1)65740 __ai __attribute__((target("v8.3a"))) float64x2_t vcaddq_rot270_f64(float64x2_t __p0, float64x2_t __p1) {
65741   float64x2_t __ret;
65742   __ret = (float64x2_t) __builtin_neon_vcaddq_rot270_f64((int8x16_t)__p0, (int8x16_t)__p1, 42);
65743   return __ret;
65744 }
65745 #else
vcaddq_rot270_f64(float64x2_t __p0,float64x2_t __p1)65746 __ai __attribute__((target("v8.3a"))) float64x2_t vcaddq_rot270_f64(float64x2_t __p0, float64x2_t __p1) {
65747   float64x2_t __ret;
65748   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65749   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65750   __ret = (float64x2_t) __builtin_neon_vcaddq_rot270_f64((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
65751   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65752   return __ret;
65753 }
65754 #endif
65755 
65756 #ifdef __LITTLE_ENDIAN__
vcaddq_rot90_f64(float64x2_t __p0,float64x2_t __p1)65757 __ai __attribute__((target("v8.3a"))) float64x2_t vcaddq_rot90_f64(float64x2_t __p0, float64x2_t __p1) {
65758   float64x2_t __ret;
65759   __ret = (float64x2_t) __builtin_neon_vcaddq_rot90_f64((int8x16_t)__p0, (int8x16_t)__p1, 42);
65760   return __ret;
65761 }
65762 #else
vcaddq_rot90_f64(float64x2_t __p0,float64x2_t __p1)65763 __ai __attribute__((target("v8.3a"))) float64x2_t vcaddq_rot90_f64(float64x2_t __p0, float64x2_t __p1) {
65764   float64x2_t __ret;
65765   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65766   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65767   __ret = (float64x2_t) __builtin_neon_vcaddq_rot90_f64((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
65768   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65769   return __ret;
65770 }
65771 #endif
65772 
65773 #ifdef __LITTLE_ENDIAN__
vcmlaq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)65774 __ai __attribute__((target("v8.3a"))) float64x2_t vcmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
65775   float64x2_t __ret;
65776   __ret = (float64x2_t) __builtin_neon_vcmlaq_f64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
65777   return __ret;
65778 }
65779 #else
vcmlaq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)65780 __ai __attribute__((target("v8.3a"))) float64x2_t vcmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
65781   float64x2_t __ret;
65782   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65783   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65784   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
65785   __ret = (float64x2_t) __builtin_neon_vcmlaq_f64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
65786   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65787   return __ret;
65788 }
__noswap_vcmlaq_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)65789 __ai __attribute__((target("v8.3a"))) float64x2_t __noswap_vcmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
65790   float64x2_t __ret;
65791   __ret = (float64x2_t) __builtin_neon_vcmlaq_f64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
65792   return __ret;
65793 }
65794 #endif
65795 
vcmla_f64(float64x1_t __p0,float64x1_t __p1,float64x1_t __p2)65796 __ai __attribute__((target("v8.3a"))) float64x1_t vcmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
65797   float64x1_t __ret;
65798   __ret = (float64x1_t) __builtin_neon_vcmla_f64((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
65799   return __ret;
65800 }
65801 #define vcmla_lane_f64(__p0_819, __p1_819, __p2_819, __p3_819) __extension__ ({ \
65802   float64x1_t __ret_819; \
65803   float64x1_t __s0_819 = __p0_819; \
65804   float64x1_t __s1_819 = __p1_819; \
65805   float64x1_t __s2_819 = __p2_819; \
65806 float64x1_t __reint_819 = __s2_819; \
65807 uint64x2_t __reint1_819 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_819, __p3_819), vgetq_lane_u64(*(uint64x2_t *) &__reint_819, __p3_819)}; \
65808   __ret_819 = vcmla_f64(__s0_819, __s1_819, *(float64x1_t *) &__reint1_819); \
65809   __ret_819; \
65810 })
65811 #ifdef __LITTLE_ENDIAN__
65812 #define vcmlaq_lane_f64(__p0_820, __p1_820, __p2_820, __p3_820) __extension__ ({ \
65813   float64x2_t __ret_820; \
65814   float64x2_t __s0_820 = __p0_820; \
65815   float64x2_t __s1_820 = __p1_820; \
65816   float64x1_t __s2_820 = __p2_820; \
65817 float64x1_t __reint_820 = __s2_820; \
65818 uint64x2_t __reint1_820 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_820, __p3_820), vgetq_lane_u64(*(uint64x2_t *) &__reint_820, __p3_820)}; \
65819   __ret_820 = vcmlaq_f64(__s0_820, __s1_820, *(float64x2_t *) &__reint1_820); \
65820   __ret_820; \
65821 })
65822 #else
65823 #define vcmlaq_lane_f64(__p0_821, __p1_821, __p2_821, __p3_821) __extension__ ({ \
65824   float64x2_t __ret_821; \
65825   float64x2_t __s0_821 = __p0_821; \
65826   float64x2_t __s1_821 = __p1_821; \
65827   float64x1_t __s2_821 = __p2_821; \
65828   float64x2_t __rev0_821;  __rev0_821 = __builtin_shufflevector(__s0_821, __s0_821, 1, 0); \
65829   float64x2_t __rev1_821;  __rev1_821 = __builtin_shufflevector(__s1_821, __s1_821, 1, 0); \
65830 float64x1_t __reint_821 = __s2_821; \
65831 uint64x2_t __reint1_821 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_821, __p3_821), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_821, __p3_821)}; \
65832   __ret_821 = __noswap_vcmlaq_f64(__rev0_821, __rev1_821, *(float64x2_t *) &__reint1_821); \
65833   __ret_821 = __builtin_shufflevector(__ret_821, __ret_821, 1, 0); \
65834   __ret_821; \
65835 })
65836 #endif
65837 
65838 #ifdef __LITTLE_ENDIAN__
65839 #define vcmla_laneq_f64(__p0_822, __p1_822, __p2_822, __p3_822) __extension__ ({ \
65840   float64x1_t __ret_822; \
65841   float64x1_t __s0_822 = __p0_822; \
65842   float64x1_t __s1_822 = __p1_822; \
65843   float64x2_t __s2_822 = __p2_822; \
65844 float64x2_t __reint_822 = __s2_822; \
65845 uint64x2_t __reint1_822 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_822, __p3_822), vgetq_lane_u64(*(uint64x2_t *) &__reint_822, __p3_822)}; \
65846   __ret_822 = vcmla_f64(__s0_822, __s1_822, *(float64x1_t *) &__reint1_822); \
65847   __ret_822; \
65848 })
65849 #else
65850 #define vcmla_laneq_f64(__p0_823, __p1_823, __p2_823, __p3_823) __extension__ ({ \
65851   float64x1_t __ret_823; \
65852   float64x1_t __s0_823 = __p0_823; \
65853   float64x1_t __s1_823 = __p1_823; \
65854   float64x2_t __s2_823 = __p2_823; \
65855   float64x2_t __rev2_823;  __rev2_823 = __builtin_shufflevector(__s2_823, __s2_823, 1, 0); \
65856 float64x2_t __reint_823 = __rev2_823; \
65857 uint64x2_t __reint1_823 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_823, __p3_823), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_823, __p3_823)}; \
65858   __ret_823 = vcmla_f64(__s0_823, __s1_823, *(float64x1_t *) &__reint1_823); \
65859   __ret_823; \
65860 })
65861 #endif
65862 
65863 #ifdef __LITTLE_ENDIAN__
65864 #define vcmlaq_laneq_f64(__p0_824, __p1_824, __p2_824, __p3_824) __extension__ ({ \
65865   float64x2_t __ret_824; \
65866   float64x2_t __s0_824 = __p0_824; \
65867   float64x2_t __s1_824 = __p1_824; \
65868   float64x2_t __s2_824 = __p2_824; \
65869 float64x2_t __reint_824 = __s2_824; \
65870 uint64x2_t __reint1_824 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_824, __p3_824), vgetq_lane_u64(*(uint64x2_t *) &__reint_824, __p3_824)}; \
65871   __ret_824 = vcmlaq_f64(__s0_824, __s1_824, *(float64x2_t *) &__reint1_824); \
65872   __ret_824; \
65873 })
65874 #else
65875 #define vcmlaq_laneq_f64(__p0_825, __p1_825, __p2_825, __p3_825) __extension__ ({ \
65876   float64x2_t __ret_825; \
65877   float64x2_t __s0_825 = __p0_825; \
65878   float64x2_t __s1_825 = __p1_825; \
65879   float64x2_t __s2_825 = __p2_825; \
65880   float64x2_t __rev0_825;  __rev0_825 = __builtin_shufflevector(__s0_825, __s0_825, 1, 0); \
65881   float64x2_t __rev1_825;  __rev1_825 = __builtin_shufflevector(__s1_825, __s1_825, 1, 0); \
65882   float64x2_t __rev2_825;  __rev2_825 = __builtin_shufflevector(__s2_825, __s2_825, 1, 0); \
65883 float64x2_t __reint_825 = __rev2_825; \
65884 uint64x2_t __reint1_825 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_825, __p3_825), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_825, __p3_825)}; \
65885   __ret_825 = __noswap_vcmlaq_f64(__rev0_825, __rev1_825, *(float64x2_t *) &__reint1_825); \
65886   __ret_825 = __builtin_shufflevector(__ret_825, __ret_825, 1, 0); \
65887   __ret_825; \
65888 })
65889 #endif
65890 
65891 #ifdef __LITTLE_ENDIAN__
vcmlaq_rot180_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)65892 __ai __attribute__((target("v8.3a"))) float64x2_t vcmlaq_rot180_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
65893   float64x2_t __ret;
65894   __ret = (float64x2_t) __builtin_neon_vcmlaq_rot180_f64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
65895   return __ret;
65896 }
65897 #else
vcmlaq_rot180_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)65898 __ai __attribute__((target("v8.3a"))) float64x2_t vcmlaq_rot180_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
65899   float64x2_t __ret;
65900   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65901   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65902   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
65903   __ret = (float64x2_t) __builtin_neon_vcmlaq_rot180_f64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
65904   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65905   return __ret;
65906 }
__noswap_vcmlaq_rot180_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)65907 __ai __attribute__((target("v8.3a"))) float64x2_t __noswap_vcmlaq_rot180_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
65908   float64x2_t __ret;
65909   __ret = (float64x2_t) __builtin_neon_vcmlaq_rot180_f64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
65910   return __ret;
65911 }
65912 #endif
65913 
vcmla_rot180_f64(float64x1_t __p0,float64x1_t __p1,float64x1_t __p2)65914 __ai __attribute__((target("v8.3a"))) float64x1_t vcmla_rot180_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
65915   float64x1_t __ret;
65916   __ret = (float64x1_t) __builtin_neon_vcmla_rot180_f64((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
65917   return __ret;
65918 }
65919 #define vcmla_rot180_lane_f64(__p0_826, __p1_826, __p2_826, __p3_826) __extension__ ({ \
65920   float64x1_t __ret_826; \
65921   float64x1_t __s0_826 = __p0_826; \
65922   float64x1_t __s1_826 = __p1_826; \
65923   float64x1_t __s2_826 = __p2_826; \
65924 float64x1_t __reint_826 = __s2_826; \
65925 uint64x2_t __reint1_826 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_826, __p3_826), vgetq_lane_u64(*(uint64x2_t *) &__reint_826, __p3_826)}; \
65926   __ret_826 = vcmla_rot180_f64(__s0_826, __s1_826, *(float64x1_t *) &__reint1_826); \
65927   __ret_826; \
65928 })
65929 #ifdef __LITTLE_ENDIAN__
65930 #define vcmlaq_rot180_lane_f64(__p0_827, __p1_827, __p2_827, __p3_827) __extension__ ({ \
65931   float64x2_t __ret_827; \
65932   float64x2_t __s0_827 = __p0_827; \
65933   float64x2_t __s1_827 = __p1_827; \
65934   float64x1_t __s2_827 = __p2_827; \
65935 float64x1_t __reint_827 = __s2_827; \
65936 uint64x2_t __reint1_827 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_827, __p3_827), vgetq_lane_u64(*(uint64x2_t *) &__reint_827, __p3_827)}; \
65937   __ret_827 = vcmlaq_rot180_f64(__s0_827, __s1_827, *(float64x2_t *) &__reint1_827); \
65938   __ret_827; \
65939 })
65940 #else
65941 #define vcmlaq_rot180_lane_f64(__p0_828, __p1_828, __p2_828, __p3_828) __extension__ ({ \
65942   float64x2_t __ret_828; \
65943   float64x2_t __s0_828 = __p0_828; \
65944   float64x2_t __s1_828 = __p1_828; \
65945   float64x1_t __s2_828 = __p2_828; \
65946   float64x2_t __rev0_828;  __rev0_828 = __builtin_shufflevector(__s0_828, __s0_828, 1, 0); \
65947   float64x2_t __rev1_828;  __rev1_828 = __builtin_shufflevector(__s1_828, __s1_828, 1, 0); \
65948 float64x1_t __reint_828 = __s2_828; \
65949 uint64x2_t __reint1_828 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_828, __p3_828), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_828, __p3_828)}; \
65950   __ret_828 = __noswap_vcmlaq_rot180_f64(__rev0_828, __rev1_828, *(float64x2_t *) &__reint1_828); \
65951   __ret_828 = __builtin_shufflevector(__ret_828, __ret_828, 1, 0); \
65952   __ret_828; \
65953 })
65954 #endif
65955 
65956 #ifdef __LITTLE_ENDIAN__
65957 #define vcmla_rot180_laneq_f64(__p0_829, __p1_829, __p2_829, __p3_829) __extension__ ({ \
65958   float64x1_t __ret_829; \
65959   float64x1_t __s0_829 = __p0_829; \
65960   float64x1_t __s1_829 = __p1_829; \
65961   float64x2_t __s2_829 = __p2_829; \
65962 float64x2_t __reint_829 = __s2_829; \
65963 uint64x2_t __reint1_829 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_829, __p3_829), vgetq_lane_u64(*(uint64x2_t *) &__reint_829, __p3_829)}; \
65964   __ret_829 = vcmla_rot180_f64(__s0_829, __s1_829, *(float64x1_t *) &__reint1_829); \
65965   __ret_829; \
65966 })
65967 #else
65968 #define vcmla_rot180_laneq_f64(__p0_830, __p1_830, __p2_830, __p3_830) __extension__ ({ \
65969   float64x1_t __ret_830; \
65970   float64x1_t __s0_830 = __p0_830; \
65971   float64x1_t __s1_830 = __p1_830; \
65972   float64x2_t __s2_830 = __p2_830; \
65973   float64x2_t __rev2_830;  __rev2_830 = __builtin_shufflevector(__s2_830, __s2_830, 1, 0); \
65974 float64x2_t __reint_830 = __rev2_830; \
65975 uint64x2_t __reint1_830 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_830, __p3_830), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_830, __p3_830)}; \
65976   __ret_830 = vcmla_rot180_f64(__s0_830, __s1_830, *(float64x1_t *) &__reint1_830); \
65977   __ret_830; \
65978 })
65979 #endif
65980 
65981 #ifdef __LITTLE_ENDIAN__
65982 #define vcmlaq_rot180_laneq_f64(__p0_831, __p1_831, __p2_831, __p3_831) __extension__ ({ \
65983   float64x2_t __ret_831; \
65984   float64x2_t __s0_831 = __p0_831; \
65985   float64x2_t __s1_831 = __p1_831; \
65986   float64x2_t __s2_831 = __p2_831; \
65987 float64x2_t __reint_831 = __s2_831; \
65988 uint64x2_t __reint1_831 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_831, __p3_831), vgetq_lane_u64(*(uint64x2_t *) &__reint_831, __p3_831)}; \
65989   __ret_831 = vcmlaq_rot180_f64(__s0_831, __s1_831, *(float64x2_t *) &__reint1_831); \
65990   __ret_831; \
65991 })
65992 #else
65993 #define vcmlaq_rot180_laneq_f64(__p0_832, __p1_832, __p2_832, __p3_832) __extension__ ({ \
65994   float64x2_t __ret_832; \
65995   float64x2_t __s0_832 = __p0_832; \
65996   float64x2_t __s1_832 = __p1_832; \
65997   float64x2_t __s2_832 = __p2_832; \
65998   float64x2_t __rev0_832;  __rev0_832 = __builtin_shufflevector(__s0_832, __s0_832, 1, 0); \
65999   float64x2_t __rev1_832;  __rev1_832 = __builtin_shufflevector(__s1_832, __s1_832, 1, 0); \
66000   float64x2_t __rev2_832;  __rev2_832 = __builtin_shufflevector(__s2_832, __s2_832, 1, 0); \
66001 float64x2_t __reint_832 = __rev2_832; \
66002 uint64x2_t __reint1_832 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_832, __p3_832), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_832, __p3_832)}; \
66003   __ret_832 = __noswap_vcmlaq_rot180_f64(__rev0_832, __rev1_832, *(float64x2_t *) &__reint1_832); \
66004   __ret_832 = __builtin_shufflevector(__ret_832, __ret_832, 1, 0); \
66005   __ret_832; \
66006 })
66007 #endif
66008 
66009 #ifdef __LITTLE_ENDIAN__
vcmlaq_rot270_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)66010 __ai __attribute__((target("v8.3a"))) float64x2_t vcmlaq_rot270_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
66011   float64x2_t __ret;
66012   __ret = (float64x2_t) __builtin_neon_vcmlaq_rot270_f64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
66013   return __ret;
66014 }
66015 #else
vcmlaq_rot270_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)66016 __ai __attribute__((target("v8.3a"))) float64x2_t vcmlaq_rot270_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
66017   float64x2_t __ret;
66018   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66019   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66020   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66021   __ret = (float64x2_t) __builtin_neon_vcmlaq_rot270_f64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
66022   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66023   return __ret;
66024 }
__noswap_vcmlaq_rot270_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)66025 __ai __attribute__((target("v8.3a"))) float64x2_t __noswap_vcmlaq_rot270_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
66026   float64x2_t __ret;
66027   __ret = (float64x2_t) __builtin_neon_vcmlaq_rot270_f64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
66028   return __ret;
66029 }
66030 #endif
66031 
vcmla_rot270_f64(float64x1_t __p0,float64x1_t __p1,float64x1_t __p2)66032 __ai __attribute__((target("v8.3a"))) float64x1_t vcmla_rot270_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
66033   float64x1_t __ret;
66034   __ret = (float64x1_t) __builtin_neon_vcmla_rot270_f64((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
66035   return __ret;
66036 }
66037 #define vcmla_rot270_lane_f64(__p0_833, __p1_833, __p2_833, __p3_833) __extension__ ({ \
66038   float64x1_t __ret_833; \
66039   float64x1_t __s0_833 = __p0_833; \
66040   float64x1_t __s1_833 = __p1_833; \
66041   float64x1_t __s2_833 = __p2_833; \
66042 float64x1_t __reint_833 = __s2_833; \
66043 uint64x2_t __reint1_833 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_833, __p3_833), vgetq_lane_u64(*(uint64x2_t *) &__reint_833, __p3_833)}; \
66044   __ret_833 = vcmla_rot270_f64(__s0_833, __s1_833, *(float64x1_t *) &__reint1_833); \
66045   __ret_833; \
66046 })
66047 #ifdef __LITTLE_ENDIAN__
66048 #define vcmlaq_rot270_lane_f64(__p0_834, __p1_834, __p2_834, __p3_834) __extension__ ({ \
66049   float64x2_t __ret_834; \
66050   float64x2_t __s0_834 = __p0_834; \
66051   float64x2_t __s1_834 = __p1_834; \
66052   float64x1_t __s2_834 = __p2_834; \
66053 float64x1_t __reint_834 = __s2_834; \
66054 uint64x2_t __reint1_834 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_834, __p3_834), vgetq_lane_u64(*(uint64x2_t *) &__reint_834, __p3_834)}; \
66055   __ret_834 = vcmlaq_rot270_f64(__s0_834, __s1_834, *(float64x2_t *) &__reint1_834); \
66056   __ret_834; \
66057 })
66058 #else
66059 #define vcmlaq_rot270_lane_f64(__p0_835, __p1_835, __p2_835, __p3_835) __extension__ ({ \
66060   float64x2_t __ret_835; \
66061   float64x2_t __s0_835 = __p0_835; \
66062   float64x2_t __s1_835 = __p1_835; \
66063   float64x1_t __s2_835 = __p2_835; \
66064   float64x2_t __rev0_835;  __rev0_835 = __builtin_shufflevector(__s0_835, __s0_835, 1, 0); \
66065   float64x2_t __rev1_835;  __rev1_835 = __builtin_shufflevector(__s1_835, __s1_835, 1, 0); \
66066 float64x1_t __reint_835 = __s2_835; \
66067 uint64x2_t __reint1_835 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_835, __p3_835), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_835, __p3_835)}; \
66068   __ret_835 = __noswap_vcmlaq_rot270_f64(__rev0_835, __rev1_835, *(float64x2_t *) &__reint1_835); \
66069   __ret_835 = __builtin_shufflevector(__ret_835, __ret_835, 1, 0); \
66070   __ret_835; \
66071 })
66072 #endif
66073 
66074 #ifdef __LITTLE_ENDIAN__
66075 #define vcmla_rot270_laneq_f64(__p0_836, __p1_836, __p2_836, __p3_836) __extension__ ({ \
66076   float64x1_t __ret_836; \
66077   float64x1_t __s0_836 = __p0_836; \
66078   float64x1_t __s1_836 = __p1_836; \
66079   float64x2_t __s2_836 = __p2_836; \
66080 float64x2_t __reint_836 = __s2_836; \
66081 uint64x2_t __reint1_836 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_836, __p3_836), vgetq_lane_u64(*(uint64x2_t *) &__reint_836, __p3_836)}; \
66082   __ret_836 = vcmla_rot270_f64(__s0_836, __s1_836, *(float64x1_t *) &__reint1_836); \
66083   __ret_836; \
66084 })
66085 #else
66086 #define vcmla_rot270_laneq_f64(__p0_837, __p1_837, __p2_837, __p3_837) __extension__ ({ \
66087   float64x1_t __ret_837; \
66088   float64x1_t __s0_837 = __p0_837; \
66089   float64x1_t __s1_837 = __p1_837; \
66090   float64x2_t __s2_837 = __p2_837; \
66091   float64x2_t __rev2_837;  __rev2_837 = __builtin_shufflevector(__s2_837, __s2_837, 1, 0); \
66092 float64x2_t __reint_837 = __rev2_837; \
66093 uint64x2_t __reint1_837 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_837, __p3_837), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_837, __p3_837)}; \
66094   __ret_837 = vcmla_rot270_f64(__s0_837, __s1_837, *(float64x1_t *) &__reint1_837); \
66095   __ret_837; \
66096 })
66097 #endif
66098 
66099 #ifdef __LITTLE_ENDIAN__
66100 #define vcmlaq_rot270_laneq_f64(__p0_838, __p1_838, __p2_838, __p3_838) __extension__ ({ \
66101   float64x2_t __ret_838; \
66102   float64x2_t __s0_838 = __p0_838; \
66103   float64x2_t __s1_838 = __p1_838; \
66104   float64x2_t __s2_838 = __p2_838; \
66105 float64x2_t __reint_838 = __s2_838; \
66106 uint64x2_t __reint1_838 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_838, __p3_838), vgetq_lane_u64(*(uint64x2_t *) &__reint_838, __p3_838)}; \
66107   __ret_838 = vcmlaq_rot270_f64(__s0_838, __s1_838, *(float64x2_t *) &__reint1_838); \
66108   __ret_838; \
66109 })
66110 #else
66111 #define vcmlaq_rot270_laneq_f64(__p0_839, __p1_839, __p2_839, __p3_839) __extension__ ({ \
66112   float64x2_t __ret_839; \
66113   float64x2_t __s0_839 = __p0_839; \
66114   float64x2_t __s1_839 = __p1_839; \
66115   float64x2_t __s2_839 = __p2_839; \
66116   float64x2_t __rev0_839;  __rev0_839 = __builtin_shufflevector(__s0_839, __s0_839, 1, 0); \
66117   float64x2_t __rev1_839;  __rev1_839 = __builtin_shufflevector(__s1_839, __s1_839, 1, 0); \
66118   float64x2_t __rev2_839;  __rev2_839 = __builtin_shufflevector(__s2_839, __s2_839, 1, 0); \
66119 float64x2_t __reint_839 = __rev2_839; \
66120 uint64x2_t __reint1_839 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_839, __p3_839), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_839, __p3_839)}; \
66121   __ret_839 = __noswap_vcmlaq_rot270_f64(__rev0_839, __rev1_839, *(float64x2_t *) &__reint1_839); \
66122   __ret_839 = __builtin_shufflevector(__ret_839, __ret_839, 1, 0); \
66123   __ret_839; \
66124 })
66125 #endif
66126 
66127 #ifdef __LITTLE_ENDIAN__
vcmlaq_rot90_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)66128 __ai __attribute__((target("v8.3a"))) float64x2_t vcmlaq_rot90_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
66129   float64x2_t __ret;
66130   __ret = (float64x2_t) __builtin_neon_vcmlaq_rot90_f64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
66131   return __ret;
66132 }
66133 #else
vcmlaq_rot90_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)66134 __ai __attribute__((target("v8.3a"))) float64x2_t vcmlaq_rot90_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
66135   float64x2_t __ret;
66136   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66137   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66138   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66139   __ret = (float64x2_t) __builtin_neon_vcmlaq_rot90_f64((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
66140   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66141   return __ret;
66142 }
__noswap_vcmlaq_rot90_f64(float64x2_t __p0,float64x2_t __p1,float64x2_t __p2)66143 __ai __attribute__((target("v8.3a"))) float64x2_t __noswap_vcmlaq_rot90_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
66144   float64x2_t __ret;
66145   __ret = (float64x2_t) __builtin_neon_vcmlaq_rot90_f64((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
66146   return __ret;
66147 }
66148 #endif
66149 
vcmla_rot90_f64(float64x1_t __p0,float64x1_t __p1,float64x1_t __p2)66150 __ai __attribute__((target("v8.3a"))) float64x1_t vcmla_rot90_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
66151   float64x1_t __ret;
66152   __ret = (float64x1_t) __builtin_neon_vcmla_rot90_f64((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
66153   return __ret;
66154 }
66155 #define vcmla_rot90_lane_f64(__p0_840, __p1_840, __p2_840, __p3_840) __extension__ ({ \
66156   float64x1_t __ret_840; \
66157   float64x1_t __s0_840 = __p0_840; \
66158   float64x1_t __s1_840 = __p1_840; \
66159   float64x1_t __s2_840 = __p2_840; \
66160 float64x1_t __reint_840 = __s2_840; \
66161 uint64x2_t __reint1_840 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_840, __p3_840), vgetq_lane_u64(*(uint64x2_t *) &__reint_840, __p3_840)}; \
66162   __ret_840 = vcmla_rot90_f64(__s0_840, __s1_840, *(float64x1_t *) &__reint1_840); \
66163   __ret_840; \
66164 })
66165 #ifdef __LITTLE_ENDIAN__
66166 #define vcmlaq_rot90_lane_f64(__p0_841, __p1_841, __p2_841, __p3_841) __extension__ ({ \
66167   float64x2_t __ret_841; \
66168   float64x2_t __s0_841 = __p0_841; \
66169   float64x2_t __s1_841 = __p1_841; \
66170   float64x1_t __s2_841 = __p2_841; \
66171 float64x1_t __reint_841 = __s2_841; \
66172 uint64x2_t __reint1_841 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_841, __p3_841), vgetq_lane_u64(*(uint64x2_t *) &__reint_841, __p3_841)}; \
66173   __ret_841 = vcmlaq_rot90_f64(__s0_841, __s1_841, *(float64x2_t *) &__reint1_841); \
66174   __ret_841; \
66175 })
66176 #else
66177 #define vcmlaq_rot90_lane_f64(__p0_842, __p1_842, __p2_842, __p3_842) __extension__ ({ \
66178   float64x2_t __ret_842; \
66179   float64x2_t __s0_842 = __p0_842; \
66180   float64x2_t __s1_842 = __p1_842; \
66181   float64x1_t __s2_842 = __p2_842; \
66182   float64x2_t __rev0_842;  __rev0_842 = __builtin_shufflevector(__s0_842, __s0_842, 1, 0); \
66183   float64x2_t __rev1_842;  __rev1_842 = __builtin_shufflevector(__s1_842, __s1_842, 1, 0); \
66184 float64x1_t __reint_842 = __s2_842; \
66185 uint64x2_t __reint1_842 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_842, __p3_842), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_842, __p3_842)}; \
66186   __ret_842 = __noswap_vcmlaq_rot90_f64(__rev0_842, __rev1_842, *(float64x2_t *) &__reint1_842); \
66187   __ret_842 = __builtin_shufflevector(__ret_842, __ret_842, 1, 0); \
66188   __ret_842; \
66189 })
66190 #endif
66191 
66192 #ifdef __LITTLE_ENDIAN__
66193 #define vcmla_rot90_laneq_f64(__p0_843, __p1_843, __p2_843, __p3_843) __extension__ ({ \
66194   float64x1_t __ret_843; \
66195   float64x1_t __s0_843 = __p0_843; \
66196   float64x1_t __s1_843 = __p1_843; \
66197   float64x2_t __s2_843 = __p2_843; \
66198 float64x2_t __reint_843 = __s2_843; \
66199 uint64x2_t __reint1_843 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_843, __p3_843), vgetq_lane_u64(*(uint64x2_t *) &__reint_843, __p3_843)}; \
66200   __ret_843 = vcmla_rot90_f64(__s0_843, __s1_843, *(float64x1_t *) &__reint1_843); \
66201   __ret_843; \
66202 })
66203 #else
66204 #define vcmla_rot90_laneq_f64(__p0_844, __p1_844, __p2_844, __p3_844) __extension__ ({ \
66205   float64x1_t __ret_844; \
66206   float64x1_t __s0_844 = __p0_844; \
66207   float64x1_t __s1_844 = __p1_844; \
66208   float64x2_t __s2_844 = __p2_844; \
66209   float64x2_t __rev2_844;  __rev2_844 = __builtin_shufflevector(__s2_844, __s2_844, 1, 0); \
66210 float64x2_t __reint_844 = __rev2_844; \
66211 uint64x2_t __reint1_844 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_844, __p3_844), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_844, __p3_844)}; \
66212   __ret_844 = vcmla_rot90_f64(__s0_844, __s1_844, *(float64x1_t *) &__reint1_844); \
66213   __ret_844; \
66214 })
66215 #endif
66216 
66217 #ifdef __LITTLE_ENDIAN__
66218 #define vcmlaq_rot90_laneq_f64(__p0_845, __p1_845, __p2_845, __p3_845) __extension__ ({ \
66219   float64x2_t __ret_845; \
66220   float64x2_t __s0_845 = __p0_845; \
66221   float64x2_t __s1_845 = __p1_845; \
66222   float64x2_t __s2_845 = __p2_845; \
66223 float64x2_t __reint_845 = __s2_845; \
66224 uint64x2_t __reint1_845 = (uint64x2_t) {vgetq_lane_u64(*(uint64x2_t *) &__reint_845, __p3_845), vgetq_lane_u64(*(uint64x2_t *) &__reint_845, __p3_845)}; \
66225   __ret_845 = vcmlaq_rot90_f64(__s0_845, __s1_845, *(float64x2_t *) &__reint1_845); \
66226   __ret_845; \
66227 })
66228 #else
66229 #define vcmlaq_rot90_laneq_f64(__p0_846, __p1_846, __p2_846, __p3_846) __extension__ ({ \
66230   float64x2_t __ret_846; \
66231   float64x2_t __s0_846 = __p0_846; \
66232   float64x2_t __s1_846 = __p1_846; \
66233   float64x2_t __s2_846 = __p2_846; \
66234   float64x2_t __rev0_846;  __rev0_846 = __builtin_shufflevector(__s0_846, __s0_846, 1, 0); \
66235   float64x2_t __rev1_846;  __rev1_846 = __builtin_shufflevector(__s1_846, __s1_846, 1, 0); \
66236   float64x2_t __rev2_846;  __rev2_846 = __builtin_shufflevector(__s2_846, __s2_846, 1, 0); \
66237 float64x2_t __reint_846 = __rev2_846; \
66238 uint64x2_t __reint1_846 = (uint64x2_t) {__noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_846, __p3_846), __noswap_vgetq_lane_u64(*(uint64x2_t *) &__reint_846, __p3_846)}; \
66239   __ret_846 = __noswap_vcmlaq_rot90_f64(__rev0_846, __rev1_846, *(float64x2_t *) &__reint1_846); \
66240   __ret_846 = __builtin_shufflevector(__ret_846, __ret_846, 1, 0); \
66241   __ret_846; \
66242 })
66243 #endif
66244 
66245 #ifdef __LITTLE_ENDIAN__
vrnd32xq_f32(float32x4_t __p0)66246 __ai __attribute__((target("v8.5a"))) float32x4_t vrnd32xq_f32(float32x4_t __p0) {
66247   float32x4_t __ret;
66248   __ret = (float32x4_t) __builtin_neon_vrnd32xq_f32((int8x16_t)__p0, 41);
66249   return __ret;
66250 }
66251 #else
vrnd32xq_f32(float32x4_t __p0)66252 __ai __attribute__((target("v8.5a"))) float32x4_t vrnd32xq_f32(float32x4_t __p0) {
66253   float32x4_t __ret;
66254   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66255   __ret = (float32x4_t) __builtin_neon_vrnd32xq_f32((int8x16_t)__rev0, 41);
66256   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66257   return __ret;
66258 }
66259 #endif
66260 
66261 #ifdef __LITTLE_ENDIAN__
vrnd32x_f32(float32x2_t __p0)66262 __ai __attribute__((target("v8.5a"))) float32x2_t vrnd32x_f32(float32x2_t __p0) {
66263   float32x2_t __ret;
66264   __ret = (float32x2_t) __builtin_neon_vrnd32x_f32((int8x8_t)__p0, 9);
66265   return __ret;
66266 }
66267 #else
vrnd32x_f32(float32x2_t __p0)66268 __ai __attribute__((target("v8.5a"))) float32x2_t vrnd32x_f32(float32x2_t __p0) {
66269   float32x2_t __ret;
66270   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66271   __ret = (float32x2_t) __builtin_neon_vrnd32x_f32((int8x8_t)__rev0, 9);
66272   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66273   return __ret;
66274 }
66275 #endif
66276 
66277 #ifdef __LITTLE_ENDIAN__
vrnd32xq_f64(float64x2_t __p0)66278 __ai __attribute__((target("v8.5a"))) float64x2_t vrnd32xq_f64(float64x2_t __p0) {
66279   float64x2_t __ret;
66280   __ret = (float64x2_t) __builtin_neon_vrnd32xq_f64((int8x16_t)__p0, 42);
66281   return __ret;
66282 }
66283 #else
vrnd32xq_f64(float64x2_t __p0)66284 __ai __attribute__((target("v8.5a"))) float64x2_t vrnd32xq_f64(float64x2_t __p0) {
66285   float64x2_t __ret;
66286   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66287   __ret = (float64x2_t) __builtin_neon_vrnd32xq_f64((int8x16_t)__rev0, 42);
66288   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66289   return __ret;
66290 }
66291 #endif
66292 
vrnd32x_f64(float64x1_t __p0)66293 __ai __attribute__((target("v8.5a"))) float64x1_t vrnd32x_f64(float64x1_t __p0) {
66294   float64x1_t __ret;
66295   __ret = (float64x1_t) __builtin_neon_vrnd32x_f64((int8x8_t)__p0, 10);
66296   return __ret;
66297 }
66298 #ifdef __LITTLE_ENDIAN__
vrnd32zq_f32(float32x4_t __p0)66299 __ai __attribute__((target("v8.5a"))) float32x4_t vrnd32zq_f32(float32x4_t __p0) {
66300   float32x4_t __ret;
66301   __ret = (float32x4_t) __builtin_neon_vrnd32zq_f32((int8x16_t)__p0, 41);
66302   return __ret;
66303 }
66304 #else
vrnd32zq_f32(float32x4_t __p0)66305 __ai __attribute__((target("v8.5a"))) float32x4_t vrnd32zq_f32(float32x4_t __p0) {
66306   float32x4_t __ret;
66307   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66308   __ret = (float32x4_t) __builtin_neon_vrnd32zq_f32((int8x16_t)__rev0, 41);
66309   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66310   return __ret;
66311 }
66312 #endif
66313 
66314 #ifdef __LITTLE_ENDIAN__
vrnd32z_f32(float32x2_t __p0)66315 __ai __attribute__((target("v8.5a"))) float32x2_t vrnd32z_f32(float32x2_t __p0) {
66316   float32x2_t __ret;
66317   __ret = (float32x2_t) __builtin_neon_vrnd32z_f32((int8x8_t)__p0, 9);
66318   return __ret;
66319 }
66320 #else
vrnd32z_f32(float32x2_t __p0)66321 __ai __attribute__((target("v8.5a"))) float32x2_t vrnd32z_f32(float32x2_t __p0) {
66322   float32x2_t __ret;
66323   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66324   __ret = (float32x2_t) __builtin_neon_vrnd32z_f32((int8x8_t)__rev0, 9);
66325   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66326   return __ret;
66327 }
66328 #endif
66329 
66330 #ifdef __LITTLE_ENDIAN__
vrnd32zq_f64(float64x2_t __p0)66331 __ai __attribute__((target("v8.5a"))) float64x2_t vrnd32zq_f64(float64x2_t __p0) {
66332   float64x2_t __ret;
66333   __ret = (float64x2_t) __builtin_neon_vrnd32zq_f64((int8x16_t)__p0, 42);
66334   return __ret;
66335 }
66336 #else
vrnd32zq_f64(float64x2_t __p0)66337 __ai __attribute__((target("v8.5a"))) float64x2_t vrnd32zq_f64(float64x2_t __p0) {
66338   float64x2_t __ret;
66339   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66340   __ret = (float64x2_t) __builtin_neon_vrnd32zq_f64((int8x16_t)__rev0, 42);
66341   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66342   return __ret;
66343 }
66344 #endif
66345 
vrnd32z_f64(float64x1_t __p0)66346 __ai __attribute__((target("v8.5a"))) float64x1_t vrnd32z_f64(float64x1_t __p0) {
66347   float64x1_t __ret;
66348   __ret = (float64x1_t) __builtin_neon_vrnd32z_f64((int8x8_t)__p0, 10);
66349   return __ret;
66350 }
66351 #ifdef __LITTLE_ENDIAN__
vrnd64xq_f32(float32x4_t __p0)66352 __ai __attribute__((target("v8.5a"))) float32x4_t vrnd64xq_f32(float32x4_t __p0) {
66353   float32x4_t __ret;
66354   __ret = (float32x4_t) __builtin_neon_vrnd64xq_f32((int8x16_t)__p0, 41);
66355   return __ret;
66356 }
66357 #else
vrnd64xq_f32(float32x4_t __p0)66358 __ai __attribute__((target("v8.5a"))) float32x4_t vrnd64xq_f32(float32x4_t __p0) {
66359   float32x4_t __ret;
66360   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66361   __ret = (float32x4_t) __builtin_neon_vrnd64xq_f32((int8x16_t)__rev0, 41);
66362   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66363   return __ret;
66364 }
66365 #endif
66366 
66367 #ifdef __LITTLE_ENDIAN__
vrnd64x_f32(float32x2_t __p0)66368 __ai __attribute__((target("v8.5a"))) float32x2_t vrnd64x_f32(float32x2_t __p0) {
66369   float32x2_t __ret;
66370   __ret = (float32x2_t) __builtin_neon_vrnd64x_f32((int8x8_t)__p0, 9);
66371   return __ret;
66372 }
66373 #else
vrnd64x_f32(float32x2_t __p0)66374 __ai __attribute__((target("v8.5a"))) float32x2_t vrnd64x_f32(float32x2_t __p0) {
66375   float32x2_t __ret;
66376   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66377   __ret = (float32x2_t) __builtin_neon_vrnd64x_f32((int8x8_t)__rev0, 9);
66378   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66379   return __ret;
66380 }
66381 #endif
66382 
66383 #ifdef __LITTLE_ENDIAN__
vrnd64xq_f64(float64x2_t __p0)66384 __ai __attribute__((target("v8.5a"))) float64x2_t vrnd64xq_f64(float64x2_t __p0) {
66385   float64x2_t __ret;
66386   __ret = (float64x2_t) __builtin_neon_vrnd64xq_f64((int8x16_t)__p0, 42);
66387   return __ret;
66388 }
66389 #else
vrnd64xq_f64(float64x2_t __p0)66390 __ai __attribute__((target("v8.5a"))) float64x2_t vrnd64xq_f64(float64x2_t __p0) {
66391   float64x2_t __ret;
66392   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66393   __ret = (float64x2_t) __builtin_neon_vrnd64xq_f64((int8x16_t)__rev0, 42);
66394   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66395   return __ret;
66396 }
66397 #endif
66398 
vrnd64x_f64(float64x1_t __p0)66399 __ai __attribute__((target("v8.5a"))) float64x1_t vrnd64x_f64(float64x1_t __p0) {
66400   float64x1_t __ret;
66401   __ret = (float64x1_t) __builtin_neon_vrnd64x_f64((int8x8_t)__p0, 10);
66402   return __ret;
66403 }
66404 #ifdef __LITTLE_ENDIAN__
vrnd64zq_f32(float32x4_t __p0)66405 __ai __attribute__((target("v8.5a"))) float32x4_t vrnd64zq_f32(float32x4_t __p0) {
66406   float32x4_t __ret;
66407   __ret = (float32x4_t) __builtin_neon_vrnd64zq_f32((int8x16_t)__p0, 41);
66408   return __ret;
66409 }
66410 #else
vrnd64zq_f32(float32x4_t __p0)66411 __ai __attribute__((target("v8.5a"))) float32x4_t vrnd64zq_f32(float32x4_t __p0) {
66412   float32x4_t __ret;
66413   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66414   __ret = (float32x4_t) __builtin_neon_vrnd64zq_f32((int8x16_t)__rev0, 41);
66415   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66416   return __ret;
66417 }
66418 #endif
66419 
66420 #ifdef __LITTLE_ENDIAN__
vrnd64z_f32(float32x2_t __p0)66421 __ai __attribute__((target("v8.5a"))) float32x2_t vrnd64z_f32(float32x2_t __p0) {
66422   float32x2_t __ret;
66423   __ret = (float32x2_t) __builtin_neon_vrnd64z_f32((int8x8_t)__p0, 9);
66424   return __ret;
66425 }
66426 #else
vrnd64z_f32(float32x2_t __p0)66427 __ai __attribute__((target("v8.5a"))) float32x2_t vrnd64z_f32(float32x2_t __p0) {
66428   float32x2_t __ret;
66429   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66430   __ret = (float32x2_t) __builtin_neon_vrnd64z_f32((int8x8_t)__rev0, 9);
66431   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66432   return __ret;
66433 }
66434 #endif
66435 
66436 #ifdef __LITTLE_ENDIAN__
vrnd64zq_f64(float64x2_t __p0)66437 __ai __attribute__((target("v8.5a"))) float64x2_t vrnd64zq_f64(float64x2_t __p0) {
66438   float64x2_t __ret;
66439   __ret = (float64x2_t) __builtin_neon_vrnd64zq_f64((int8x16_t)__p0, 42);
66440   return __ret;
66441 }
66442 #else
vrnd64zq_f64(float64x2_t __p0)66443 __ai __attribute__((target("v8.5a"))) float64x2_t vrnd64zq_f64(float64x2_t __p0) {
66444   float64x2_t __ret;
66445   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66446   __ret = (float64x2_t) __builtin_neon_vrnd64zq_f64((int8x16_t)__rev0, 42);
66447   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66448   return __ret;
66449 }
66450 #endif
66451 
vrnd64z_f64(float64x1_t __p0)66452 __ai __attribute__((target("v8.5a"))) float64x1_t vrnd64z_f64(float64x1_t __p0) {
66453   float64x1_t __ret;
66454   __ret = (float64x1_t) __builtin_neon_vrnd64z_f64((int8x8_t)__p0, 10);
66455   return __ret;
66456 }
66457 #endif
66458 #if defined(__aarch64__) && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
66459 #ifdef __LITTLE_ENDIAN__
vrndq_f64(float64x2_t __p0)66460 __ai float64x2_t vrndq_f64(float64x2_t __p0) {
66461   float64x2_t __ret;
66462   __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 42);
66463   return __ret;
66464 }
66465 #else
vrndq_f64(float64x2_t __p0)66466 __ai float64x2_t vrndq_f64(float64x2_t __p0) {
66467   float64x2_t __ret;
66468   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66469   __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 42);
66470   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66471   return __ret;
66472 }
66473 #endif
66474 
vrnd_f64(float64x1_t __p0)66475 __ai float64x1_t vrnd_f64(float64x1_t __p0) {
66476   float64x1_t __ret;
66477   __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
66478   return __ret;
66479 }
66480 #ifdef __LITTLE_ENDIAN__
vrndaq_f64(float64x2_t __p0)66481 __ai float64x2_t vrndaq_f64(float64x2_t __p0) {
66482   float64x2_t __ret;
66483   __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 42);
66484   return __ret;
66485 }
66486 #else
vrndaq_f64(float64x2_t __p0)66487 __ai float64x2_t vrndaq_f64(float64x2_t __p0) {
66488   float64x2_t __ret;
66489   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66490   __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 42);
66491   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66492   return __ret;
66493 }
66494 #endif
66495 
vrnda_f64(float64x1_t __p0)66496 __ai float64x1_t vrnda_f64(float64x1_t __p0) {
66497   float64x1_t __ret;
66498   __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
66499   return __ret;
66500 }
66501 #ifdef __LITTLE_ENDIAN__
vrndiq_f64(float64x2_t __p0)66502 __ai float64x2_t vrndiq_f64(float64x2_t __p0) {
66503   float64x2_t __ret;
66504   __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 42);
66505   return __ret;
66506 }
66507 #else
vrndiq_f64(float64x2_t __p0)66508 __ai float64x2_t vrndiq_f64(float64x2_t __p0) {
66509   float64x2_t __ret;
66510   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66511   __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 42);
66512   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66513   return __ret;
66514 }
66515 #endif
66516 
vrndi_f64(float64x1_t __p0)66517 __ai float64x1_t vrndi_f64(float64x1_t __p0) {
66518   float64x1_t __ret;
66519   __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
66520   return __ret;
66521 }
66522 #ifdef __LITTLE_ENDIAN__
vrndmq_f64(float64x2_t __p0)66523 __ai float64x2_t vrndmq_f64(float64x2_t __p0) {
66524   float64x2_t __ret;
66525   __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 42);
66526   return __ret;
66527 }
66528 #else
vrndmq_f64(float64x2_t __p0)66529 __ai float64x2_t vrndmq_f64(float64x2_t __p0) {
66530   float64x2_t __ret;
66531   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66532   __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 42);
66533   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66534   return __ret;
66535 }
66536 #endif
66537 
vrndm_f64(float64x1_t __p0)66538 __ai float64x1_t vrndm_f64(float64x1_t __p0) {
66539   float64x1_t __ret;
66540   __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
66541   return __ret;
66542 }
66543 #ifdef __LITTLE_ENDIAN__
vrndnq_f64(float64x2_t __p0)66544 __ai float64x2_t vrndnq_f64(float64x2_t __p0) {
66545   float64x2_t __ret;
66546   __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 42);
66547   return __ret;
66548 }
66549 #else
vrndnq_f64(float64x2_t __p0)66550 __ai float64x2_t vrndnq_f64(float64x2_t __p0) {
66551   float64x2_t __ret;
66552   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66553   __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 42);
66554   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66555   return __ret;
66556 }
66557 #endif
66558 
vrndn_f64(float64x1_t __p0)66559 __ai float64x1_t vrndn_f64(float64x1_t __p0) {
66560   float64x1_t __ret;
66561   __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
66562   return __ret;
66563 }
66564 #ifdef __LITTLE_ENDIAN__
vrndpq_f64(float64x2_t __p0)66565 __ai float64x2_t vrndpq_f64(float64x2_t __p0) {
66566   float64x2_t __ret;
66567   __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 42);
66568   return __ret;
66569 }
66570 #else
vrndpq_f64(float64x2_t __p0)66571 __ai float64x2_t vrndpq_f64(float64x2_t __p0) {
66572   float64x2_t __ret;
66573   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66574   __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 42);
66575   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66576   return __ret;
66577 }
66578 #endif
66579 
vrndp_f64(float64x1_t __p0)66580 __ai float64x1_t vrndp_f64(float64x1_t __p0) {
66581   float64x1_t __ret;
66582   __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
66583   return __ret;
66584 }
66585 #ifdef __LITTLE_ENDIAN__
vrndxq_f64(float64x2_t __p0)66586 __ai float64x2_t vrndxq_f64(float64x2_t __p0) {
66587   float64x2_t __ret;
66588   __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 42);
66589   return __ret;
66590 }
66591 #else
vrndxq_f64(float64x2_t __p0)66592 __ai float64x2_t vrndxq_f64(float64x2_t __p0) {
66593   float64x2_t __ret;
66594   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66595   __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 42);
66596   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66597   return __ret;
66598 }
66599 #endif
66600 
vrndx_f64(float64x1_t __p0)66601 __ai float64x1_t vrndx_f64(float64x1_t __p0) {
66602   float64x1_t __ret;
66603   __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
66604   return __ret;
66605 }
66606 #endif
66607 #if defined(__aarch64__) && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
66608 #ifdef __LITTLE_ENDIAN__
vmaxnmq_f64(float64x2_t __p0,float64x2_t __p1)66609 __ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
66610   float64x2_t __ret;
66611   __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
66612   return __ret;
66613 }
66614 #else
vmaxnmq_f64(float64x2_t __p0,float64x2_t __p1)66615 __ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
66616   float64x2_t __ret;
66617   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66618   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66619   __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
66620   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66621   return __ret;
66622 }
66623 #endif
66624 
vmaxnm_f64(float64x1_t __p0,float64x1_t __p1)66625 __ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
66626   float64x1_t __ret;
66627   __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
66628   return __ret;
66629 }
66630 #ifdef __LITTLE_ENDIAN__
vminnmq_f64(float64x2_t __p0,float64x2_t __p1)66631 __ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
66632   float64x2_t __ret;
66633   __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
66634   return __ret;
66635 }
66636 #else
vminnmq_f64(float64x2_t __p0,float64x2_t __p1)66637 __ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
66638   float64x2_t __ret;
66639   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66640   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66641   __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
66642   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66643   return __ret;
66644 }
66645 #endif
66646 
vminnm_f64(float64x1_t __p0,float64x1_t __p1)66647 __ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
66648   float64x1_t __ret;
66649   __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
66650   return __ret;
66651 }
66652 #endif
66653 #ifdef __LITTLE_ENDIAN__
vabaq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)66654 __ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66655   uint8x16_t __ret;
66656   __ret = __p0 + vabdq_u8(__p1, __p2);
66657   return __ret;
66658 }
66659 #else
vabaq_u8(uint8x16_t __p0,uint8x16_t __p1,uint8x16_t __p2)66660 __ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66661   uint8x16_t __ret;
66662   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66663   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66664   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66665   __ret = __rev0 + __noswap_vabdq_u8(__rev1, __rev2);
66666   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66667   return __ret;
66668 }
66669 #endif
66670 
66671 #ifdef __LITTLE_ENDIAN__
vabaq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)66672 __ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66673   uint32x4_t __ret;
66674   __ret = __p0 + vabdq_u32(__p1, __p2);
66675   return __ret;
66676 }
66677 #else
vabaq_u32(uint32x4_t __p0,uint32x4_t __p1,uint32x4_t __p2)66678 __ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66679   uint32x4_t __ret;
66680   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66681   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66682   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66683   __ret = __rev0 + __noswap_vabdq_u32(__rev1, __rev2);
66684   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66685   return __ret;
66686 }
66687 #endif
66688 
66689 #ifdef __LITTLE_ENDIAN__
vabaq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)66690 __ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66691   uint16x8_t __ret;
66692   __ret = __p0 + vabdq_u16(__p1, __p2);
66693   return __ret;
66694 }
66695 #else
vabaq_u16(uint16x8_t __p0,uint16x8_t __p1,uint16x8_t __p2)66696 __ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66697   uint16x8_t __ret;
66698   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66699   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66700   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66701   __ret = __rev0 + __noswap_vabdq_u16(__rev1, __rev2);
66702   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66703   return __ret;
66704 }
66705 #endif
66706 
66707 #ifdef __LITTLE_ENDIAN__
vabaq_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)66708 __ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66709   int8x16_t __ret;
66710   __ret = __p0 + vabdq_s8(__p1, __p2);
66711   return __ret;
66712 }
66713 #else
vabaq_s8(int8x16_t __p0,int8x16_t __p1,int8x16_t __p2)66714 __ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66715   int8x16_t __ret;
66716   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66717   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66718   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66719   __ret = __rev0 + __noswap_vabdq_s8(__rev1, __rev2);
66720   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66721   return __ret;
66722 }
66723 #endif
66724 
66725 #ifdef __LITTLE_ENDIAN__
vabaq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)66726 __ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66727   int32x4_t __ret;
66728   __ret = __p0 + vabdq_s32(__p1, __p2);
66729   return __ret;
66730 }
66731 #else
vabaq_s32(int32x4_t __p0,int32x4_t __p1,int32x4_t __p2)66732 __ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66733   int32x4_t __ret;
66734   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66735   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66736   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66737   __ret = __rev0 + __noswap_vabdq_s32(__rev1, __rev2);
66738   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66739   return __ret;
66740 }
66741 #endif
66742 
66743 #ifdef __LITTLE_ENDIAN__
vabaq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)66744 __ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66745   int16x8_t __ret;
66746   __ret = __p0 + vabdq_s16(__p1, __p2);
66747   return __ret;
66748 }
66749 #else
vabaq_s16(int16x8_t __p0,int16x8_t __p1,int16x8_t __p2)66750 __ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66751   int16x8_t __ret;
66752   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66753   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66754   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66755   __ret = __rev0 + __noswap_vabdq_s16(__rev1, __rev2);
66756   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66757   return __ret;
66758 }
66759 #endif
66760 
66761 #ifdef __LITTLE_ENDIAN__
vaba_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)66762 __ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66763   uint8x8_t __ret;
66764   __ret = __p0 + vabd_u8(__p1, __p2);
66765   return __ret;
66766 }
66767 #else
vaba_u8(uint8x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)66768 __ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66769   uint8x8_t __ret;
66770   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66771   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66772   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66773   __ret = __rev0 + __noswap_vabd_u8(__rev1, __rev2);
66774   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66775   return __ret;
66776 }
66777 #endif
66778 
66779 #ifdef __LITTLE_ENDIAN__
vaba_u32(uint32x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)66780 __ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66781   uint32x2_t __ret;
66782   __ret = __p0 + vabd_u32(__p1, __p2);
66783   return __ret;
66784 }
66785 #else
vaba_u32(uint32x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)66786 __ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66787   uint32x2_t __ret;
66788   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66789   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66790   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66791   __ret = __rev0 + __noswap_vabd_u32(__rev1, __rev2);
66792   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66793   return __ret;
66794 }
66795 #endif
66796 
66797 #ifdef __LITTLE_ENDIAN__
vaba_u16(uint16x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)66798 __ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66799   uint16x4_t __ret;
66800   __ret = __p0 + vabd_u16(__p1, __p2);
66801   return __ret;
66802 }
66803 #else
vaba_u16(uint16x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)66804 __ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66805   uint16x4_t __ret;
66806   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66807   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66808   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66809   __ret = __rev0 + __noswap_vabd_u16(__rev1, __rev2);
66810   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66811   return __ret;
66812 }
66813 #endif
66814 
66815 #ifdef __LITTLE_ENDIAN__
vaba_s8(int8x8_t __p0,int8x8_t __p1,int8x8_t __p2)66816 __ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66817   int8x8_t __ret;
66818   __ret = __p0 + vabd_s8(__p1, __p2);
66819   return __ret;
66820 }
66821 #else
vaba_s8(int8x8_t __p0,int8x8_t __p1,int8x8_t __p2)66822 __ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66823   int8x8_t __ret;
66824   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66825   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66826   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66827   __ret = __rev0 + __noswap_vabd_s8(__rev1, __rev2);
66828   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66829   return __ret;
66830 }
66831 #endif
66832 
66833 #ifdef __LITTLE_ENDIAN__
vaba_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)66834 __ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66835   int32x2_t __ret;
66836   __ret = __p0 + vabd_s32(__p1, __p2);
66837   return __ret;
66838 }
66839 #else
vaba_s32(int32x2_t __p0,int32x2_t __p1,int32x2_t __p2)66840 __ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66841   int32x2_t __ret;
66842   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66843   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66844   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66845   __ret = __rev0 + __noswap_vabd_s32(__rev1, __rev2);
66846   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66847   return __ret;
66848 }
66849 #endif
66850 
66851 #ifdef __LITTLE_ENDIAN__
vaba_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)66852 __ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66853   int16x4_t __ret;
66854   __ret = __p0 + vabd_s16(__p1, __p2);
66855   return __ret;
66856 }
66857 #else
vaba_s16(int16x4_t __p0,int16x4_t __p1,int16x4_t __p2)66858 __ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66859   int16x4_t __ret;
66860   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66861   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66862   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66863   __ret = __rev0 + __noswap_vabd_s16(__rev1, __rev2);
66864   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66865   return __ret;
66866 }
66867 #endif
66868 
66869 #ifdef __LITTLE_ENDIAN__
vabdl_u8(uint8x8_t __p0,uint8x8_t __p1)66870 __ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66871   uint16x8_t __ret;
66872   __ret = (uint16x8_t)(vmovl_u8((uint8x8_t)(vabd_u8(__p0, __p1))));
66873   return __ret;
66874 }
66875 #else
vabdl_u8(uint8x8_t __p0,uint8x8_t __p1)66876 __ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66877   uint16x8_t __ret;
66878   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66879   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66880   __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__rev0, __rev1))));
66881   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66882   return __ret;
66883 }
__noswap_vabdl_u8(uint8x8_t __p0,uint8x8_t __p1)66884 __ai uint16x8_t __noswap_vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66885   uint16x8_t __ret;
66886   __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__p0, __p1))));
66887   return __ret;
66888 }
66889 #endif
66890 
66891 #ifdef __LITTLE_ENDIAN__
vabdl_u32(uint32x2_t __p0,uint32x2_t __p1)66892 __ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66893   uint64x2_t __ret;
66894   __ret = (uint64x2_t)(vmovl_u32((uint32x2_t)(vabd_u32(__p0, __p1))));
66895   return __ret;
66896 }
66897 #else
vabdl_u32(uint32x2_t __p0,uint32x2_t __p1)66898 __ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66899   uint64x2_t __ret;
66900   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66901   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66902   __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__rev0, __rev1))));
66903   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66904   return __ret;
66905 }
__noswap_vabdl_u32(uint32x2_t __p0,uint32x2_t __p1)66906 __ai uint64x2_t __noswap_vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66907   uint64x2_t __ret;
66908   __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__p0, __p1))));
66909   return __ret;
66910 }
66911 #endif
66912 
66913 #ifdef __LITTLE_ENDIAN__
vabdl_u16(uint16x4_t __p0,uint16x4_t __p1)66914 __ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66915   uint32x4_t __ret;
66916   __ret = (uint32x4_t)(vmovl_u16((uint16x4_t)(vabd_u16(__p0, __p1))));
66917   return __ret;
66918 }
66919 #else
vabdl_u16(uint16x4_t __p0,uint16x4_t __p1)66920 __ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66921   uint32x4_t __ret;
66922   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66923   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66924   __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__rev0, __rev1))));
66925   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66926   return __ret;
66927 }
__noswap_vabdl_u16(uint16x4_t __p0,uint16x4_t __p1)66928 __ai uint32x4_t __noswap_vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66929   uint32x4_t __ret;
66930   __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__p0, __p1))));
66931   return __ret;
66932 }
66933 #endif
66934 
66935 #ifdef __LITTLE_ENDIAN__
vabdl_s8(int8x8_t __p0,int8x8_t __p1)66936 __ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66937   int16x8_t __ret;
66938   __ret = (int16x8_t)(vmovl_u8((uint8x8_t)(vabd_s8(__p0, __p1))));
66939   return __ret;
66940 }
66941 #else
vabdl_s8(int8x8_t __p0,int8x8_t __p1)66942 __ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66943   int16x8_t __ret;
66944   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66945   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66946   __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__rev0, __rev1))));
66947   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66948   return __ret;
66949 }
__noswap_vabdl_s8(int8x8_t __p0,int8x8_t __p1)66950 __ai int16x8_t __noswap_vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66951   int16x8_t __ret;
66952   __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__p0, __p1))));
66953   return __ret;
66954 }
66955 #endif
66956 
66957 #ifdef __LITTLE_ENDIAN__
vabdl_s32(int32x2_t __p0,int32x2_t __p1)66958 __ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66959   int64x2_t __ret;
66960   __ret = (int64x2_t)(vmovl_u32((uint32x2_t)(vabd_s32(__p0, __p1))));
66961   return __ret;
66962 }
66963 #else
vabdl_s32(int32x2_t __p0,int32x2_t __p1)66964 __ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66965   int64x2_t __ret;
66966   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66967   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66968   __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__rev0, __rev1))));
66969   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66970   return __ret;
66971 }
__noswap_vabdl_s32(int32x2_t __p0,int32x2_t __p1)66972 __ai int64x2_t __noswap_vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66973   int64x2_t __ret;
66974   __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__p0, __p1))));
66975   return __ret;
66976 }
66977 #endif
66978 
66979 #ifdef __LITTLE_ENDIAN__
vabdl_s16(int16x4_t __p0,int16x4_t __p1)66980 __ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66981   int32x4_t __ret;
66982   __ret = (int32x4_t)(vmovl_u16((uint16x4_t)(vabd_s16(__p0, __p1))));
66983   return __ret;
66984 }
66985 #else
vabdl_s16(int16x4_t __p0,int16x4_t __p1)66986 __ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66987   int32x4_t __ret;
66988   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66989   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66990   __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__rev0, __rev1))));
66991   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66992   return __ret;
66993 }
__noswap_vabdl_s16(int16x4_t __p0,int16x4_t __p1)66994 __ai int32x4_t __noswap_vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66995   int32x4_t __ret;
66996   __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__p0, __p1))));
66997   return __ret;
66998 }
66999 #endif
67000 
67001 #ifdef __LITTLE_ENDIAN__
vaddl_u8(uint8x8_t __p0,uint8x8_t __p1)67002 __ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
67003   uint16x8_t __ret;
67004   __ret = vmovl_u8(__p0) + vmovl_u8(__p1);
67005   return __ret;
67006 }
67007 #else
vaddl_u8(uint8x8_t __p0,uint8x8_t __p1)67008 __ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
67009   uint16x8_t __ret;
67010   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67011   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67012   __ret = __noswap_vmovl_u8(__rev0) + __noswap_vmovl_u8(__rev1);
67013   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67014   return __ret;
67015 }
67016 #endif
67017 
67018 #ifdef __LITTLE_ENDIAN__
vaddl_u32(uint32x2_t __p0,uint32x2_t __p1)67019 __ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
67020   uint64x2_t __ret;
67021   __ret = vmovl_u32(__p0) + vmovl_u32(__p1);
67022   return __ret;
67023 }
67024 #else
vaddl_u32(uint32x2_t __p0,uint32x2_t __p1)67025 __ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
67026   uint64x2_t __ret;
67027   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67028   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67029   __ret = __noswap_vmovl_u32(__rev0) + __noswap_vmovl_u32(__rev1);
67030   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67031   return __ret;
67032 }
67033 #endif
67034 
67035 #ifdef __LITTLE_ENDIAN__
vaddl_u16(uint16x4_t __p0,uint16x4_t __p1)67036 __ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
67037   uint32x4_t __ret;
67038   __ret = vmovl_u16(__p0) + vmovl_u16(__p1);
67039   return __ret;
67040 }
67041 #else
vaddl_u16(uint16x4_t __p0,uint16x4_t __p1)67042 __ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
67043   uint32x4_t __ret;
67044   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67045   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67046   __ret = __noswap_vmovl_u16(__rev0) + __noswap_vmovl_u16(__rev1);
67047   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67048   return __ret;
67049 }
67050 #endif
67051 
67052 #ifdef __LITTLE_ENDIAN__
vaddl_s8(int8x8_t __p0,int8x8_t __p1)67053 __ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
67054   int16x8_t __ret;
67055   __ret = vmovl_s8(__p0) + vmovl_s8(__p1);
67056   return __ret;
67057 }
67058 #else
vaddl_s8(int8x8_t __p0,int8x8_t __p1)67059 __ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
67060   int16x8_t __ret;
67061   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67062   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67063   __ret = __noswap_vmovl_s8(__rev0) + __noswap_vmovl_s8(__rev1);
67064   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67065   return __ret;
67066 }
67067 #endif
67068 
67069 #ifdef __LITTLE_ENDIAN__
vaddl_s32(int32x2_t __p0,int32x2_t __p1)67070 __ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
67071   int64x2_t __ret;
67072   __ret = vmovl_s32(__p0) + vmovl_s32(__p1);
67073   return __ret;
67074 }
67075 #else
vaddl_s32(int32x2_t __p0,int32x2_t __p1)67076 __ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
67077   int64x2_t __ret;
67078   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67079   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67080   __ret = __noswap_vmovl_s32(__rev0) + __noswap_vmovl_s32(__rev1);
67081   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67082   return __ret;
67083 }
67084 #endif
67085 
67086 #ifdef __LITTLE_ENDIAN__
vaddl_s16(int16x4_t __p0,int16x4_t __p1)67087 __ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
67088   int32x4_t __ret;
67089   __ret = vmovl_s16(__p0) + vmovl_s16(__p1);
67090   return __ret;
67091 }
67092 #else
vaddl_s16(int16x4_t __p0,int16x4_t __p1)67093 __ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
67094   int32x4_t __ret;
67095   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67096   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67097   __ret = __noswap_vmovl_s16(__rev0) + __noswap_vmovl_s16(__rev1);
67098   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67099   return __ret;
67100 }
67101 #endif
67102 
67103 #ifdef __LITTLE_ENDIAN__
vaddw_u8(uint16x8_t __p0,uint8x8_t __p1)67104 __ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
67105   uint16x8_t __ret;
67106   __ret = __p0 + vmovl_u8(__p1);
67107   return __ret;
67108 }
67109 #else
vaddw_u8(uint16x8_t __p0,uint8x8_t __p1)67110 __ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
67111   uint16x8_t __ret;
67112   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67113   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67114   __ret = __rev0 + __noswap_vmovl_u8(__rev1);
67115   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67116   return __ret;
67117 }
67118 #endif
67119 
67120 #ifdef __LITTLE_ENDIAN__
vaddw_u32(uint64x2_t __p0,uint32x2_t __p1)67121 __ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
67122   uint64x2_t __ret;
67123   __ret = __p0 + vmovl_u32(__p1);
67124   return __ret;
67125 }
67126 #else
vaddw_u32(uint64x2_t __p0,uint32x2_t __p1)67127 __ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
67128   uint64x2_t __ret;
67129   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67130   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67131   __ret = __rev0 + __noswap_vmovl_u32(__rev1);
67132   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67133   return __ret;
67134 }
67135 #endif
67136 
67137 #ifdef __LITTLE_ENDIAN__
vaddw_u16(uint32x4_t __p0,uint16x4_t __p1)67138 __ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
67139   uint32x4_t __ret;
67140   __ret = __p0 + vmovl_u16(__p1);
67141   return __ret;
67142 }
67143 #else
vaddw_u16(uint32x4_t __p0,uint16x4_t __p1)67144 __ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
67145   uint32x4_t __ret;
67146   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67147   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67148   __ret = __rev0 + __noswap_vmovl_u16(__rev1);
67149   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67150   return __ret;
67151 }
67152 #endif
67153 
67154 #ifdef __LITTLE_ENDIAN__
vaddw_s8(int16x8_t __p0,int8x8_t __p1)67155 __ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
67156   int16x8_t __ret;
67157   __ret = __p0 + vmovl_s8(__p1);
67158   return __ret;
67159 }
67160 #else
vaddw_s8(int16x8_t __p0,int8x8_t __p1)67161 __ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
67162   int16x8_t __ret;
67163   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67164   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67165   __ret = __rev0 + __noswap_vmovl_s8(__rev1);
67166   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67167   return __ret;
67168 }
67169 #endif
67170 
67171 #ifdef __LITTLE_ENDIAN__
vaddw_s32(int64x2_t __p0,int32x2_t __p1)67172 __ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
67173   int64x2_t __ret;
67174   __ret = __p0 + vmovl_s32(__p1);
67175   return __ret;
67176 }
67177 #else
vaddw_s32(int64x2_t __p0,int32x2_t __p1)67178 __ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
67179   int64x2_t __ret;
67180   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67181   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67182   __ret = __rev0 + __noswap_vmovl_s32(__rev1);
67183   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67184   return __ret;
67185 }
67186 #endif
67187 
67188 #ifdef __LITTLE_ENDIAN__
vaddw_s16(int32x4_t __p0,int16x4_t __p1)67189 __ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
67190   int32x4_t __ret;
67191   __ret = __p0 + vmovl_s16(__p1);
67192   return __ret;
67193 }
67194 #else
vaddw_s16(int32x4_t __p0,int16x4_t __p1)67195 __ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
67196   int32x4_t __ret;
67197   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67198   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67199   __ret = __rev0 + __noswap_vmovl_s16(__rev1);
67200   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67201   return __ret;
67202 }
67203 #endif
67204 
67205 #ifdef __LITTLE_ENDIAN__
67206 #define vget_lane_f16(__p0_847, __p1_847) __extension__ ({ \
67207   float16_t __ret_847; \
67208   float16x4_t __s0_847 = __p0_847; \
67209 float16x4_t __reint_847 = __s0_847; \
67210 int16_t __reint1_847 = vget_lane_s16(*(int16x4_t *) &__reint_847, __p1_847); \
67211   __ret_847 = *(float16_t *) &__reint1_847; \
67212   __ret_847; \
67213 })
67214 #else
67215 #define vget_lane_f16(__p0_848, __p1_848) __extension__ ({ \
67216   float16_t __ret_848; \
67217   float16x4_t __s0_848 = __p0_848; \
67218   float16x4_t __rev0_848;  __rev0_848 = __builtin_shufflevector(__s0_848, __s0_848, 3, 2, 1, 0); \
67219 float16x4_t __reint_848 = __rev0_848; \
67220 int16_t __reint1_848 = __noswap_vget_lane_s16(*(int16x4_t *) &__reint_848, __p1_848); \
67221   __ret_848 = *(float16_t *) &__reint1_848; \
67222   __ret_848; \
67223 })
67224 #define __noswap_vget_lane_f16(__p0_849, __p1_849) __extension__ ({ \
67225   float16_t __ret_849; \
67226   float16x4_t __s0_849 = __p0_849; \
67227 float16x4_t __reint_849 = __s0_849; \
67228 int16_t __reint1_849 = __noswap_vget_lane_s16(*(int16x4_t *) &__reint_849, __p1_849); \
67229   __ret_849 = *(float16_t *) &__reint1_849; \
67230   __ret_849; \
67231 })
67232 #endif
67233 
67234 #ifdef __LITTLE_ENDIAN__
67235 #define vgetq_lane_f16(__p0_850, __p1_850) __extension__ ({ \
67236   float16_t __ret_850; \
67237   float16x8_t __s0_850 = __p0_850; \
67238 float16x8_t __reint_850 = __s0_850; \
67239 int16_t __reint1_850 = vgetq_lane_s16(*(int16x8_t *) &__reint_850, __p1_850); \
67240   __ret_850 = *(float16_t *) &__reint1_850; \
67241   __ret_850; \
67242 })
67243 #else
67244 #define vgetq_lane_f16(__p0_851, __p1_851) __extension__ ({ \
67245   float16_t __ret_851; \
67246   float16x8_t __s0_851 = __p0_851; \
67247   float16x8_t __rev0_851;  __rev0_851 = __builtin_shufflevector(__s0_851, __s0_851, 7, 6, 5, 4, 3, 2, 1, 0); \
67248 float16x8_t __reint_851 = __rev0_851; \
67249 int16_t __reint1_851 = __noswap_vgetq_lane_s16(*(int16x8_t *) &__reint_851, __p1_851); \
67250   __ret_851 = *(float16_t *) &__reint1_851; \
67251   __ret_851; \
67252 })
67253 #define __noswap_vgetq_lane_f16(__p0_852, __p1_852) __extension__ ({ \
67254   float16_t __ret_852; \
67255   float16x8_t __s0_852 = __p0_852; \
67256 float16x8_t __reint_852 = __s0_852; \
67257 int16_t __reint1_852 = __noswap_vgetq_lane_s16(*(int16x8_t *) &__reint_852, __p1_852); \
67258   __ret_852 = *(float16_t *) &__reint1_852; \
67259   __ret_852; \
67260 })
67261 #endif
67262 
67263 #ifdef __LITTLE_ENDIAN__
vmlal_u8(uint16x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)67264 __ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67265   uint16x8_t __ret;
67266   __ret = __p0 + vmull_u8(__p1, __p2);
67267   return __ret;
67268 }
67269 #else
vmlal_u8(uint16x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)67270 __ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67271   uint16x8_t __ret;
67272   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67273   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67274   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67275   __ret = __rev0 + __noswap_vmull_u8(__rev1, __rev2);
67276   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67277   return __ret;
67278 }
__noswap_vmlal_u8(uint16x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)67279 __ai uint16x8_t __noswap_vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67280   uint16x8_t __ret;
67281   __ret = __p0 + __noswap_vmull_u8(__p1, __p2);
67282   return __ret;
67283 }
67284 #endif
67285 
67286 #ifdef __LITTLE_ENDIAN__
vmlal_u32(uint64x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)67287 __ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67288   uint64x2_t __ret;
67289   __ret = __p0 + vmull_u32(__p1, __p2);
67290   return __ret;
67291 }
67292 #else
vmlal_u32(uint64x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)67293 __ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67294   uint64x2_t __ret;
67295   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67296   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67297   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67298   __ret = __rev0 + __noswap_vmull_u32(__rev1, __rev2);
67299   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67300   return __ret;
67301 }
__noswap_vmlal_u32(uint64x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)67302 __ai uint64x2_t __noswap_vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67303   uint64x2_t __ret;
67304   __ret = __p0 + __noswap_vmull_u32(__p1, __p2);
67305   return __ret;
67306 }
67307 #endif
67308 
67309 #ifdef __LITTLE_ENDIAN__
vmlal_u16(uint32x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)67310 __ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67311   uint32x4_t __ret;
67312   __ret = __p0 + vmull_u16(__p1, __p2);
67313   return __ret;
67314 }
67315 #else
vmlal_u16(uint32x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)67316 __ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67317   uint32x4_t __ret;
67318   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67319   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67320   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67321   __ret = __rev0 + __noswap_vmull_u16(__rev1, __rev2);
67322   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67323   return __ret;
67324 }
__noswap_vmlal_u16(uint32x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)67325 __ai uint32x4_t __noswap_vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67326   uint32x4_t __ret;
67327   __ret = __p0 + __noswap_vmull_u16(__p1, __p2);
67328   return __ret;
67329 }
67330 #endif
67331 
67332 #ifdef __LITTLE_ENDIAN__
vmlal_s8(int16x8_t __p0,int8x8_t __p1,int8x8_t __p2)67333 __ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67334   int16x8_t __ret;
67335   __ret = __p0 + vmull_s8(__p1, __p2);
67336   return __ret;
67337 }
67338 #else
vmlal_s8(int16x8_t __p0,int8x8_t __p1,int8x8_t __p2)67339 __ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67340   int16x8_t __ret;
67341   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67342   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67343   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67344   __ret = __rev0 + __noswap_vmull_s8(__rev1, __rev2);
67345   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67346   return __ret;
67347 }
__noswap_vmlal_s8(int16x8_t __p0,int8x8_t __p1,int8x8_t __p2)67348 __ai int16x8_t __noswap_vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67349   int16x8_t __ret;
67350   __ret = __p0 + __noswap_vmull_s8(__p1, __p2);
67351   return __ret;
67352 }
67353 #endif
67354 
67355 #ifdef __LITTLE_ENDIAN__
vmlal_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)67356 __ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67357   int64x2_t __ret;
67358   __ret = __p0 + vmull_s32(__p1, __p2);
67359   return __ret;
67360 }
67361 #else
vmlal_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)67362 __ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67363   int64x2_t __ret;
67364   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67365   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67366   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67367   __ret = __rev0 + __noswap_vmull_s32(__rev1, __rev2);
67368   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67369   return __ret;
67370 }
__noswap_vmlal_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)67371 __ai int64x2_t __noswap_vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67372   int64x2_t __ret;
67373   __ret = __p0 + __noswap_vmull_s32(__p1, __p2);
67374   return __ret;
67375 }
67376 #endif
67377 
67378 #ifdef __LITTLE_ENDIAN__
vmlal_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)67379 __ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67380   int32x4_t __ret;
67381   __ret = __p0 + vmull_s16(__p1, __p2);
67382   return __ret;
67383 }
67384 #else
vmlal_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)67385 __ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67386   int32x4_t __ret;
67387   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67388   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67389   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67390   __ret = __rev0 + __noswap_vmull_s16(__rev1, __rev2);
67391   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67392   return __ret;
67393 }
__noswap_vmlal_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)67394 __ai int32x4_t __noswap_vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67395   int32x4_t __ret;
67396   __ret = __p0 + __noswap_vmull_s16(__p1, __p2);
67397   return __ret;
67398 }
67399 #endif
67400 
67401 #ifdef __LITTLE_ENDIAN__
67402 #define vmlal_lane_u32(__p0_853, __p1_853, __p2_853, __p3_853) __extension__ ({ \
67403   uint64x2_t __ret_853; \
67404   uint64x2_t __s0_853 = __p0_853; \
67405   uint32x2_t __s1_853 = __p1_853; \
67406   uint32x2_t __s2_853 = __p2_853; \
67407   __ret_853 = __s0_853 + vmull_u32(__s1_853, splat_lane_u32(__s2_853, __p3_853)); \
67408   __ret_853; \
67409 })
67410 #else
67411 #define vmlal_lane_u32(__p0_854, __p1_854, __p2_854, __p3_854) __extension__ ({ \
67412   uint64x2_t __ret_854; \
67413   uint64x2_t __s0_854 = __p0_854; \
67414   uint32x2_t __s1_854 = __p1_854; \
67415   uint32x2_t __s2_854 = __p2_854; \
67416   uint64x2_t __rev0_854;  __rev0_854 = __builtin_shufflevector(__s0_854, __s0_854, 1, 0); \
67417   uint32x2_t __rev1_854;  __rev1_854 = __builtin_shufflevector(__s1_854, __s1_854, 1, 0); \
67418   uint32x2_t __rev2_854;  __rev2_854 = __builtin_shufflevector(__s2_854, __s2_854, 1, 0); \
67419   __ret_854 = __rev0_854 + __noswap_vmull_u32(__rev1_854, __noswap_splat_lane_u32(__rev2_854, __p3_854)); \
67420   __ret_854 = __builtin_shufflevector(__ret_854, __ret_854, 1, 0); \
67421   __ret_854; \
67422 })
67423 #endif
67424 
67425 #ifdef __LITTLE_ENDIAN__
67426 #define vmlal_lane_u16(__p0_855, __p1_855, __p2_855, __p3_855) __extension__ ({ \
67427   uint32x4_t __ret_855; \
67428   uint32x4_t __s0_855 = __p0_855; \
67429   uint16x4_t __s1_855 = __p1_855; \
67430   uint16x4_t __s2_855 = __p2_855; \
67431   __ret_855 = __s0_855 + vmull_u16(__s1_855, splat_lane_u16(__s2_855, __p3_855)); \
67432   __ret_855; \
67433 })
67434 #else
67435 #define vmlal_lane_u16(__p0_856, __p1_856, __p2_856, __p3_856) __extension__ ({ \
67436   uint32x4_t __ret_856; \
67437   uint32x4_t __s0_856 = __p0_856; \
67438   uint16x4_t __s1_856 = __p1_856; \
67439   uint16x4_t __s2_856 = __p2_856; \
67440   uint32x4_t __rev0_856;  __rev0_856 = __builtin_shufflevector(__s0_856, __s0_856, 3, 2, 1, 0); \
67441   uint16x4_t __rev1_856;  __rev1_856 = __builtin_shufflevector(__s1_856, __s1_856, 3, 2, 1, 0); \
67442   uint16x4_t __rev2_856;  __rev2_856 = __builtin_shufflevector(__s2_856, __s2_856, 3, 2, 1, 0); \
67443   __ret_856 = __rev0_856 + __noswap_vmull_u16(__rev1_856, __noswap_splat_lane_u16(__rev2_856, __p3_856)); \
67444   __ret_856 = __builtin_shufflevector(__ret_856, __ret_856, 3, 2, 1, 0); \
67445   __ret_856; \
67446 })
67447 #endif
67448 
67449 #ifdef __LITTLE_ENDIAN__
67450 #define vmlal_lane_s32(__p0_857, __p1_857, __p2_857, __p3_857) __extension__ ({ \
67451   int64x2_t __ret_857; \
67452   int64x2_t __s0_857 = __p0_857; \
67453   int32x2_t __s1_857 = __p1_857; \
67454   int32x2_t __s2_857 = __p2_857; \
67455   __ret_857 = __s0_857 + vmull_s32(__s1_857, splat_lane_s32(__s2_857, __p3_857)); \
67456   __ret_857; \
67457 })
67458 #else
67459 #define vmlal_lane_s32(__p0_858, __p1_858, __p2_858, __p3_858) __extension__ ({ \
67460   int64x2_t __ret_858; \
67461   int64x2_t __s0_858 = __p0_858; \
67462   int32x2_t __s1_858 = __p1_858; \
67463   int32x2_t __s2_858 = __p2_858; \
67464   int64x2_t __rev0_858;  __rev0_858 = __builtin_shufflevector(__s0_858, __s0_858, 1, 0); \
67465   int32x2_t __rev1_858;  __rev1_858 = __builtin_shufflevector(__s1_858, __s1_858, 1, 0); \
67466   int32x2_t __rev2_858;  __rev2_858 = __builtin_shufflevector(__s2_858, __s2_858, 1, 0); \
67467   __ret_858 = __rev0_858 + __noswap_vmull_s32(__rev1_858, __noswap_splat_lane_s32(__rev2_858, __p3_858)); \
67468   __ret_858 = __builtin_shufflevector(__ret_858, __ret_858, 1, 0); \
67469   __ret_858; \
67470 })
67471 #endif
67472 
67473 #ifdef __LITTLE_ENDIAN__
67474 #define vmlal_lane_s16(__p0_859, __p1_859, __p2_859, __p3_859) __extension__ ({ \
67475   int32x4_t __ret_859; \
67476   int32x4_t __s0_859 = __p0_859; \
67477   int16x4_t __s1_859 = __p1_859; \
67478   int16x4_t __s2_859 = __p2_859; \
67479   __ret_859 = __s0_859 + vmull_s16(__s1_859, splat_lane_s16(__s2_859, __p3_859)); \
67480   __ret_859; \
67481 })
67482 #else
67483 #define vmlal_lane_s16(__p0_860, __p1_860, __p2_860, __p3_860) __extension__ ({ \
67484   int32x4_t __ret_860; \
67485   int32x4_t __s0_860 = __p0_860; \
67486   int16x4_t __s1_860 = __p1_860; \
67487   int16x4_t __s2_860 = __p2_860; \
67488   int32x4_t __rev0_860;  __rev0_860 = __builtin_shufflevector(__s0_860, __s0_860, 3, 2, 1, 0); \
67489   int16x4_t __rev1_860;  __rev1_860 = __builtin_shufflevector(__s1_860, __s1_860, 3, 2, 1, 0); \
67490   int16x4_t __rev2_860;  __rev2_860 = __builtin_shufflevector(__s2_860, __s2_860, 3, 2, 1, 0); \
67491   __ret_860 = __rev0_860 + __noswap_vmull_s16(__rev1_860, __noswap_splat_lane_s16(__rev2_860, __p3_860)); \
67492   __ret_860 = __builtin_shufflevector(__ret_860, __ret_860, 3, 2, 1, 0); \
67493   __ret_860; \
67494 })
67495 #endif
67496 
67497 #ifdef __LITTLE_ENDIAN__
vmlal_n_u32(uint64x2_t __p0,uint32x2_t __p1,uint32_t __p2)67498 __ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67499   uint64x2_t __ret;
67500   __ret = __p0 + vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67501   return __ret;
67502 }
67503 #else
vmlal_n_u32(uint64x2_t __p0,uint32x2_t __p1,uint32_t __p2)67504 __ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67505   uint64x2_t __ret;
67506   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67507   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67508   __ret = __rev0 + __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
67509   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67510   return __ret;
67511 }
__noswap_vmlal_n_u32(uint64x2_t __p0,uint32x2_t __p1,uint32_t __p2)67512 __ai uint64x2_t __noswap_vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67513   uint64x2_t __ret;
67514   __ret = __p0 + __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67515   return __ret;
67516 }
67517 #endif
67518 
67519 #ifdef __LITTLE_ENDIAN__
vmlal_n_u16(uint32x4_t __p0,uint16x4_t __p1,uint16_t __p2)67520 __ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67521   uint32x4_t __ret;
67522   __ret = __p0 + vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67523   return __ret;
67524 }
67525 #else
vmlal_n_u16(uint32x4_t __p0,uint16x4_t __p1,uint16_t __p2)67526 __ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67527   uint32x4_t __ret;
67528   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67529   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67530   __ret = __rev0 + __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67531   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67532   return __ret;
67533 }
__noswap_vmlal_n_u16(uint32x4_t __p0,uint16x4_t __p1,uint16_t __p2)67534 __ai uint32x4_t __noswap_vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67535   uint32x4_t __ret;
67536   __ret = __p0 + __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67537   return __ret;
67538 }
67539 #endif
67540 
67541 #ifdef __LITTLE_ENDIAN__
vmlal_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)67542 __ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67543   int64x2_t __ret;
67544   __ret = __p0 + vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67545   return __ret;
67546 }
67547 #else
vmlal_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)67548 __ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67549   int64x2_t __ret;
67550   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67551   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67552   __ret = __rev0 + __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
67553   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67554   return __ret;
67555 }
__noswap_vmlal_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)67556 __ai int64x2_t __noswap_vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67557   int64x2_t __ret;
67558   __ret = __p0 + __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67559   return __ret;
67560 }
67561 #endif
67562 
67563 #ifdef __LITTLE_ENDIAN__
vmlal_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)67564 __ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67565   int32x4_t __ret;
67566   __ret = __p0 + vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67567   return __ret;
67568 }
67569 #else
vmlal_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)67570 __ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67571   int32x4_t __ret;
67572   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67573   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67574   __ret = __rev0 + __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
67575   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67576   return __ret;
67577 }
__noswap_vmlal_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)67578 __ai int32x4_t __noswap_vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67579   int32x4_t __ret;
67580   __ret = __p0 + __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67581   return __ret;
67582 }
67583 #endif
67584 
67585 #ifdef __LITTLE_ENDIAN__
vmlsl_u8(uint16x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)67586 __ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67587   uint16x8_t __ret;
67588   __ret = __p0 - vmull_u8(__p1, __p2);
67589   return __ret;
67590 }
67591 #else
vmlsl_u8(uint16x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)67592 __ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67593   uint16x8_t __ret;
67594   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67595   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67596   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67597   __ret = __rev0 - __noswap_vmull_u8(__rev1, __rev2);
67598   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67599   return __ret;
67600 }
__noswap_vmlsl_u8(uint16x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)67601 __ai uint16x8_t __noswap_vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67602   uint16x8_t __ret;
67603   __ret = __p0 - __noswap_vmull_u8(__p1, __p2);
67604   return __ret;
67605 }
67606 #endif
67607 
67608 #ifdef __LITTLE_ENDIAN__
vmlsl_u32(uint64x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)67609 __ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67610   uint64x2_t __ret;
67611   __ret = __p0 - vmull_u32(__p1, __p2);
67612   return __ret;
67613 }
67614 #else
vmlsl_u32(uint64x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)67615 __ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67616   uint64x2_t __ret;
67617   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67618   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67619   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67620   __ret = __rev0 - __noswap_vmull_u32(__rev1, __rev2);
67621   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67622   return __ret;
67623 }
__noswap_vmlsl_u32(uint64x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)67624 __ai uint64x2_t __noswap_vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67625   uint64x2_t __ret;
67626   __ret = __p0 - __noswap_vmull_u32(__p1, __p2);
67627   return __ret;
67628 }
67629 #endif
67630 
67631 #ifdef __LITTLE_ENDIAN__
vmlsl_u16(uint32x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)67632 __ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67633   uint32x4_t __ret;
67634   __ret = __p0 - vmull_u16(__p1, __p2);
67635   return __ret;
67636 }
67637 #else
vmlsl_u16(uint32x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)67638 __ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67639   uint32x4_t __ret;
67640   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67641   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67642   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67643   __ret = __rev0 - __noswap_vmull_u16(__rev1, __rev2);
67644   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67645   return __ret;
67646 }
__noswap_vmlsl_u16(uint32x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)67647 __ai uint32x4_t __noswap_vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67648   uint32x4_t __ret;
67649   __ret = __p0 - __noswap_vmull_u16(__p1, __p2);
67650   return __ret;
67651 }
67652 #endif
67653 
67654 #ifdef __LITTLE_ENDIAN__
vmlsl_s8(int16x8_t __p0,int8x8_t __p1,int8x8_t __p2)67655 __ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67656   int16x8_t __ret;
67657   __ret = __p0 - vmull_s8(__p1, __p2);
67658   return __ret;
67659 }
67660 #else
vmlsl_s8(int16x8_t __p0,int8x8_t __p1,int8x8_t __p2)67661 __ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67662   int16x8_t __ret;
67663   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67664   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67665   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67666   __ret = __rev0 - __noswap_vmull_s8(__rev1, __rev2);
67667   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67668   return __ret;
67669 }
__noswap_vmlsl_s8(int16x8_t __p0,int8x8_t __p1,int8x8_t __p2)67670 __ai int16x8_t __noswap_vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67671   int16x8_t __ret;
67672   __ret = __p0 - __noswap_vmull_s8(__p1, __p2);
67673   return __ret;
67674 }
67675 #endif
67676 
67677 #ifdef __LITTLE_ENDIAN__
vmlsl_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)67678 __ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67679   int64x2_t __ret;
67680   __ret = __p0 - vmull_s32(__p1, __p2);
67681   return __ret;
67682 }
67683 #else
vmlsl_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)67684 __ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67685   int64x2_t __ret;
67686   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67687   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67688   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67689   __ret = __rev0 - __noswap_vmull_s32(__rev1, __rev2);
67690   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67691   return __ret;
67692 }
__noswap_vmlsl_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)67693 __ai int64x2_t __noswap_vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67694   int64x2_t __ret;
67695   __ret = __p0 - __noswap_vmull_s32(__p1, __p2);
67696   return __ret;
67697 }
67698 #endif
67699 
67700 #ifdef __LITTLE_ENDIAN__
vmlsl_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)67701 __ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67702   int32x4_t __ret;
67703   __ret = __p0 - vmull_s16(__p1, __p2);
67704   return __ret;
67705 }
67706 #else
vmlsl_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)67707 __ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67708   int32x4_t __ret;
67709   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67710   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67711   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67712   __ret = __rev0 - __noswap_vmull_s16(__rev1, __rev2);
67713   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67714   return __ret;
67715 }
__noswap_vmlsl_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)67716 __ai int32x4_t __noswap_vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67717   int32x4_t __ret;
67718   __ret = __p0 - __noswap_vmull_s16(__p1, __p2);
67719   return __ret;
67720 }
67721 #endif
67722 
67723 #ifdef __LITTLE_ENDIAN__
67724 #define vmlsl_lane_u32(__p0_861, __p1_861, __p2_861, __p3_861) __extension__ ({ \
67725   uint64x2_t __ret_861; \
67726   uint64x2_t __s0_861 = __p0_861; \
67727   uint32x2_t __s1_861 = __p1_861; \
67728   uint32x2_t __s2_861 = __p2_861; \
67729   __ret_861 = __s0_861 - vmull_u32(__s1_861, splat_lane_u32(__s2_861, __p3_861)); \
67730   __ret_861; \
67731 })
67732 #else
67733 #define vmlsl_lane_u32(__p0_862, __p1_862, __p2_862, __p3_862) __extension__ ({ \
67734   uint64x2_t __ret_862; \
67735   uint64x2_t __s0_862 = __p0_862; \
67736   uint32x2_t __s1_862 = __p1_862; \
67737   uint32x2_t __s2_862 = __p2_862; \
67738   uint64x2_t __rev0_862;  __rev0_862 = __builtin_shufflevector(__s0_862, __s0_862, 1, 0); \
67739   uint32x2_t __rev1_862;  __rev1_862 = __builtin_shufflevector(__s1_862, __s1_862, 1, 0); \
67740   uint32x2_t __rev2_862;  __rev2_862 = __builtin_shufflevector(__s2_862, __s2_862, 1, 0); \
67741   __ret_862 = __rev0_862 - __noswap_vmull_u32(__rev1_862, __noswap_splat_lane_u32(__rev2_862, __p3_862)); \
67742   __ret_862 = __builtin_shufflevector(__ret_862, __ret_862, 1, 0); \
67743   __ret_862; \
67744 })
67745 #endif
67746 
67747 #ifdef __LITTLE_ENDIAN__
67748 #define vmlsl_lane_u16(__p0_863, __p1_863, __p2_863, __p3_863) __extension__ ({ \
67749   uint32x4_t __ret_863; \
67750   uint32x4_t __s0_863 = __p0_863; \
67751   uint16x4_t __s1_863 = __p1_863; \
67752   uint16x4_t __s2_863 = __p2_863; \
67753   __ret_863 = __s0_863 - vmull_u16(__s1_863, splat_lane_u16(__s2_863, __p3_863)); \
67754   __ret_863; \
67755 })
67756 #else
67757 #define vmlsl_lane_u16(__p0_864, __p1_864, __p2_864, __p3_864) __extension__ ({ \
67758   uint32x4_t __ret_864; \
67759   uint32x4_t __s0_864 = __p0_864; \
67760   uint16x4_t __s1_864 = __p1_864; \
67761   uint16x4_t __s2_864 = __p2_864; \
67762   uint32x4_t __rev0_864;  __rev0_864 = __builtin_shufflevector(__s0_864, __s0_864, 3, 2, 1, 0); \
67763   uint16x4_t __rev1_864;  __rev1_864 = __builtin_shufflevector(__s1_864, __s1_864, 3, 2, 1, 0); \
67764   uint16x4_t __rev2_864;  __rev2_864 = __builtin_shufflevector(__s2_864, __s2_864, 3, 2, 1, 0); \
67765   __ret_864 = __rev0_864 - __noswap_vmull_u16(__rev1_864, __noswap_splat_lane_u16(__rev2_864, __p3_864)); \
67766   __ret_864 = __builtin_shufflevector(__ret_864, __ret_864, 3, 2, 1, 0); \
67767   __ret_864; \
67768 })
67769 #endif
67770 
67771 #ifdef __LITTLE_ENDIAN__
67772 #define vmlsl_lane_s32(__p0_865, __p1_865, __p2_865, __p3_865) __extension__ ({ \
67773   int64x2_t __ret_865; \
67774   int64x2_t __s0_865 = __p0_865; \
67775   int32x2_t __s1_865 = __p1_865; \
67776   int32x2_t __s2_865 = __p2_865; \
67777   __ret_865 = __s0_865 - vmull_s32(__s1_865, splat_lane_s32(__s2_865, __p3_865)); \
67778   __ret_865; \
67779 })
67780 #else
67781 #define vmlsl_lane_s32(__p0_866, __p1_866, __p2_866, __p3_866) __extension__ ({ \
67782   int64x2_t __ret_866; \
67783   int64x2_t __s0_866 = __p0_866; \
67784   int32x2_t __s1_866 = __p1_866; \
67785   int32x2_t __s2_866 = __p2_866; \
67786   int64x2_t __rev0_866;  __rev0_866 = __builtin_shufflevector(__s0_866, __s0_866, 1, 0); \
67787   int32x2_t __rev1_866;  __rev1_866 = __builtin_shufflevector(__s1_866, __s1_866, 1, 0); \
67788   int32x2_t __rev2_866;  __rev2_866 = __builtin_shufflevector(__s2_866, __s2_866, 1, 0); \
67789   __ret_866 = __rev0_866 - __noswap_vmull_s32(__rev1_866, __noswap_splat_lane_s32(__rev2_866, __p3_866)); \
67790   __ret_866 = __builtin_shufflevector(__ret_866, __ret_866, 1, 0); \
67791   __ret_866; \
67792 })
67793 #endif
67794 
67795 #ifdef __LITTLE_ENDIAN__
67796 #define vmlsl_lane_s16(__p0_867, __p1_867, __p2_867, __p3_867) __extension__ ({ \
67797   int32x4_t __ret_867; \
67798   int32x4_t __s0_867 = __p0_867; \
67799   int16x4_t __s1_867 = __p1_867; \
67800   int16x4_t __s2_867 = __p2_867; \
67801   __ret_867 = __s0_867 - vmull_s16(__s1_867, splat_lane_s16(__s2_867, __p3_867)); \
67802   __ret_867; \
67803 })
67804 #else
67805 #define vmlsl_lane_s16(__p0_868, __p1_868, __p2_868, __p3_868) __extension__ ({ \
67806   int32x4_t __ret_868; \
67807   int32x4_t __s0_868 = __p0_868; \
67808   int16x4_t __s1_868 = __p1_868; \
67809   int16x4_t __s2_868 = __p2_868; \
67810   int32x4_t __rev0_868;  __rev0_868 = __builtin_shufflevector(__s0_868, __s0_868, 3, 2, 1, 0); \
67811   int16x4_t __rev1_868;  __rev1_868 = __builtin_shufflevector(__s1_868, __s1_868, 3, 2, 1, 0); \
67812   int16x4_t __rev2_868;  __rev2_868 = __builtin_shufflevector(__s2_868, __s2_868, 3, 2, 1, 0); \
67813   __ret_868 = __rev0_868 - __noswap_vmull_s16(__rev1_868, __noswap_splat_lane_s16(__rev2_868, __p3_868)); \
67814   __ret_868 = __builtin_shufflevector(__ret_868, __ret_868, 3, 2, 1, 0); \
67815   __ret_868; \
67816 })
67817 #endif
67818 
67819 #ifdef __LITTLE_ENDIAN__
vmlsl_n_u32(uint64x2_t __p0,uint32x2_t __p1,uint32_t __p2)67820 __ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67821   uint64x2_t __ret;
67822   __ret = __p0 - vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67823   return __ret;
67824 }
67825 #else
vmlsl_n_u32(uint64x2_t __p0,uint32x2_t __p1,uint32_t __p2)67826 __ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67827   uint64x2_t __ret;
67828   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67829   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67830   __ret = __rev0 - __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
67831   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67832   return __ret;
67833 }
__noswap_vmlsl_n_u32(uint64x2_t __p0,uint32x2_t __p1,uint32_t __p2)67834 __ai uint64x2_t __noswap_vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67835   uint64x2_t __ret;
67836   __ret = __p0 - __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67837   return __ret;
67838 }
67839 #endif
67840 
67841 #ifdef __LITTLE_ENDIAN__
vmlsl_n_u16(uint32x4_t __p0,uint16x4_t __p1,uint16_t __p2)67842 __ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67843   uint32x4_t __ret;
67844   __ret = __p0 - vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67845   return __ret;
67846 }
67847 #else
vmlsl_n_u16(uint32x4_t __p0,uint16x4_t __p1,uint16_t __p2)67848 __ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67849   uint32x4_t __ret;
67850   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67851   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67852   __ret = __rev0 - __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67853   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67854   return __ret;
67855 }
__noswap_vmlsl_n_u16(uint32x4_t __p0,uint16x4_t __p1,uint16_t __p2)67856 __ai uint32x4_t __noswap_vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67857   uint32x4_t __ret;
67858   __ret = __p0 - __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67859   return __ret;
67860 }
67861 #endif
67862 
67863 #ifdef __LITTLE_ENDIAN__
vmlsl_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)67864 __ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67865   int64x2_t __ret;
67866   __ret = __p0 - vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67867   return __ret;
67868 }
67869 #else
vmlsl_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)67870 __ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67871   int64x2_t __ret;
67872   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67873   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67874   __ret = __rev0 - __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
67875   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67876   return __ret;
67877 }
__noswap_vmlsl_n_s32(int64x2_t __p0,int32x2_t __p1,int32_t __p2)67878 __ai int64x2_t __noswap_vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67879   int64x2_t __ret;
67880   __ret = __p0 - __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67881   return __ret;
67882 }
67883 #endif
67884 
67885 #ifdef __LITTLE_ENDIAN__
vmlsl_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)67886 __ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67887   int32x4_t __ret;
67888   __ret = __p0 - vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67889   return __ret;
67890 }
67891 #else
vmlsl_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)67892 __ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67893   int32x4_t __ret;
67894   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67895   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67896   __ret = __rev0 - __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
67897   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67898   return __ret;
67899 }
__noswap_vmlsl_n_s16(int32x4_t __p0,int16x4_t __p1,int16_t __p2)67900 __ai int32x4_t __noswap_vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67901   int32x4_t __ret;
67902   __ret = __p0 - __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67903   return __ret;
67904 }
67905 #endif
67906 
67907 #ifdef __LITTLE_ENDIAN__
67908 #define vset_lane_f16(__p0_869, __p1_869, __p2_869) __extension__ ({ \
67909   float16x4_t __ret_869; \
67910   float16_t __s0_869 = __p0_869; \
67911   float16x4_t __s1_869 = __p1_869; \
67912 float16_t __reint_869 = __s0_869; \
67913 float16x4_t __reint1_869 = __s1_869; \
67914 int16x4_t __reint2_869 = vset_lane_s16(*(int16_t *) &__reint_869, *(int16x4_t *) &__reint1_869, __p2_869); \
67915   __ret_869 = *(float16x4_t *) &__reint2_869; \
67916   __ret_869; \
67917 })
67918 #else
67919 #define vset_lane_f16(__p0_870, __p1_870, __p2_870) __extension__ ({ \
67920   float16x4_t __ret_870; \
67921   float16_t __s0_870 = __p0_870; \
67922   float16x4_t __s1_870 = __p1_870; \
67923   float16x4_t __rev1_870;  __rev1_870 = __builtin_shufflevector(__s1_870, __s1_870, 3, 2, 1, 0); \
67924 float16_t __reint_870 = __s0_870; \
67925 float16x4_t __reint1_870 = __rev1_870; \
67926 int16x4_t __reint2_870 = __noswap_vset_lane_s16(*(int16_t *) &__reint_870, *(int16x4_t *) &__reint1_870, __p2_870); \
67927   __ret_870 = *(float16x4_t *) &__reint2_870; \
67928   __ret_870 = __builtin_shufflevector(__ret_870, __ret_870, 3, 2, 1, 0); \
67929   __ret_870; \
67930 })
67931 #endif
67932 
67933 #ifdef __LITTLE_ENDIAN__
67934 #define vsetq_lane_f16(__p0_871, __p1_871, __p2_871) __extension__ ({ \
67935   float16x8_t __ret_871; \
67936   float16_t __s0_871 = __p0_871; \
67937   float16x8_t __s1_871 = __p1_871; \
67938 float16_t __reint_871 = __s0_871; \
67939 float16x8_t __reint1_871 = __s1_871; \
67940 int16x8_t __reint2_871 = vsetq_lane_s16(*(int16_t *) &__reint_871, *(int16x8_t *) &__reint1_871, __p2_871); \
67941   __ret_871 = *(float16x8_t *) &__reint2_871; \
67942   __ret_871; \
67943 })
67944 #else
67945 #define vsetq_lane_f16(__p0_872, __p1_872, __p2_872) __extension__ ({ \
67946   float16x8_t __ret_872; \
67947   float16_t __s0_872 = __p0_872; \
67948   float16x8_t __s1_872 = __p1_872; \
67949   float16x8_t __rev1_872;  __rev1_872 = __builtin_shufflevector(__s1_872, __s1_872, 7, 6, 5, 4, 3, 2, 1, 0); \
67950 float16_t __reint_872 = __s0_872; \
67951 float16x8_t __reint1_872 = __rev1_872; \
67952 int16x8_t __reint2_872 = __noswap_vsetq_lane_s16(*(int16_t *) &__reint_872, *(int16x8_t *) &__reint1_872, __p2_872); \
67953   __ret_872 = *(float16x8_t *) &__reint2_872; \
67954   __ret_872 = __builtin_shufflevector(__ret_872, __ret_872, 7, 6, 5, 4, 3, 2, 1, 0); \
67955   __ret_872; \
67956 })
67957 #endif
67958 
67959 #ifdef __LITTLE_ENDIAN__
67960 #define vbfmlalbq_lane_f32(__p0_873, __p1_873, __p2_873, __p3_873) __extension__ ({ \
67961   float32x4_t __ret_873; \
67962   float32x4_t __s0_873 = __p0_873; \
67963   bfloat16x8_t __s1_873 = __p1_873; \
67964   bfloat16x4_t __s2_873 = __p2_873; \
67965   __ret_873 = vbfmlalbq_f32(__s0_873, __s1_873, (bfloat16x8_t) {vget_lane_bf16(__s2_873, __p3_873), vget_lane_bf16(__s2_873, __p3_873), vget_lane_bf16(__s2_873, __p3_873), vget_lane_bf16(__s2_873, __p3_873), vget_lane_bf16(__s2_873, __p3_873), vget_lane_bf16(__s2_873, __p3_873), vget_lane_bf16(__s2_873, __p3_873), vget_lane_bf16(__s2_873, __p3_873)}); \
67966   __ret_873; \
67967 })
67968 #else
67969 #define vbfmlalbq_lane_f32(__p0_874, __p1_874, __p2_874, __p3_874) __extension__ ({ \
67970   float32x4_t __ret_874; \
67971   float32x4_t __s0_874 = __p0_874; \
67972   bfloat16x8_t __s1_874 = __p1_874; \
67973   bfloat16x4_t __s2_874 = __p2_874; \
67974   float32x4_t __rev0_874;  __rev0_874 = __builtin_shufflevector(__s0_874, __s0_874, 3, 2, 1, 0); \
67975   bfloat16x8_t __rev1_874;  __rev1_874 = __builtin_shufflevector(__s1_874, __s1_874, 7, 6, 5, 4, 3, 2, 1, 0); \
67976   bfloat16x4_t __rev2_874;  __rev2_874 = __builtin_shufflevector(__s2_874, __s2_874, 3, 2, 1, 0); \
67977   __ret_874 = __noswap_vbfmlalbq_f32(__rev0_874, __rev1_874, (bfloat16x8_t) {__noswap_vget_lane_bf16(__rev2_874, __p3_874), __noswap_vget_lane_bf16(__rev2_874, __p3_874), __noswap_vget_lane_bf16(__rev2_874, __p3_874), __noswap_vget_lane_bf16(__rev2_874, __p3_874), __noswap_vget_lane_bf16(__rev2_874, __p3_874), __noswap_vget_lane_bf16(__rev2_874, __p3_874), __noswap_vget_lane_bf16(__rev2_874, __p3_874), __noswap_vget_lane_bf16(__rev2_874, __p3_874)}); \
67978   __ret_874 = __builtin_shufflevector(__ret_874, __ret_874, 3, 2, 1, 0); \
67979   __ret_874; \
67980 })
67981 #endif
67982 
67983 #ifdef __LITTLE_ENDIAN__
67984 #define vbfmlalbq_laneq_f32(__p0_875, __p1_875, __p2_875, __p3_875) __extension__ ({ \
67985   float32x4_t __ret_875; \
67986   float32x4_t __s0_875 = __p0_875; \
67987   bfloat16x8_t __s1_875 = __p1_875; \
67988   bfloat16x8_t __s2_875 = __p2_875; \
67989   __ret_875 = vbfmlalbq_f32(__s0_875, __s1_875, (bfloat16x8_t) {vgetq_lane_bf16(__s2_875, __p3_875), vgetq_lane_bf16(__s2_875, __p3_875), vgetq_lane_bf16(__s2_875, __p3_875), vgetq_lane_bf16(__s2_875, __p3_875), vgetq_lane_bf16(__s2_875, __p3_875), vgetq_lane_bf16(__s2_875, __p3_875), vgetq_lane_bf16(__s2_875, __p3_875), vgetq_lane_bf16(__s2_875, __p3_875)}); \
67990   __ret_875; \
67991 })
67992 #else
67993 #define vbfmlalbq_laneq_f32(__p0_876, __p1_876, __p2_876, __p3_876) __extension__ ({ \
67994   float32x4_t __ret_876; \
67995   float32x4_t __s0_876 = __p0_876; \
67996   bfloat16x8_t __s1_876 = __p1_876; \
67997   bfloat16x8_t __s2_876 = __p2_876; \
67998   float32x4_t __rev0_876;  __rev0_876 = __builtin_shufflevector(__s0_876, __s0_876, 3, 2, 1, 0); \
67999   bfloat16x8_t __rev1_876;  __rev1_876 = __builtin_shufflevector(__s1_876, __s1_876, 7, 6, 5, 4, 3, 2, 1, 0); \
68000   bfloat16x8_t __rev2_876;  __rev2_876 = __builtin_shufflevector(__s2_876, __s2_876, 7, 6, 5, 4, 3, 2, 1, 0); \
68001   __ret_876 = __noswap_vbfmlalbq_f32(__rev0_876, __rev1_876, (bfloat16x8_t) {__noswap_vgetq_lane_bf16(__rev2_876, __p3_876), __noswap_vgetq_lane_bf16(__rev2_876, __p3_876), __noswap_vgetq_lane_bf16(__rev2_876, __p3_876), __noswap_vgetq_lane_bf16(__rev2_876, __p3_876), __noswap_vgetq_lane_bf16(__rev2_876, __p3_876), __noswap_vgetq_lane_bf16(__rev2_876, __p3_876), __noswap_vgetq_lane_bf16(__rev2_876, __p3_876), __noswap_vgetq_lane_bf16(__rev2_876, __p3_876)}); \
68002   __ret_876 = __builtin_shufflevector(__ret_876, __ret_876, 3, 2, 1, 0); \
68003   __ret_876; \
68004 })
68005 #endif
68006 
68007 #ifdef __LITTLE_ENDIAN__
68008 #define vbfmlaltq_lane_f32(__p0_877, __p1_877, __p2_877, __p3_877) __extension__ ({ \
68009   float32x4_t __ret_877; \
68010   float32x4_t __s0_877 = __p0_877; \
68011   bfloat16x8_t __s1_877 = __p1_877; \
68012   bfloat16x4_t __s2_877 = __p2_877; \
68013   __ret_877 = vbfmlaltq_f32(__s0_877, __s1_877, (bfloat16x8_t) {vget_lane_bf16(__s2_877, __p3_877), vget_lane_bf16(__s2_877, __p3_877), vget_lane_bf16(__s2_877, __p3_877), vget_lane_bf16(__s2_877, __p3_877), vget_lane_bf16(__s2_877, __p3_877), vget_lane_bf16(__s2_877, __p3_877), vget_lane_bf16(__s2_877, __p3_877), vget_lane_bf16(__s2_877, __p3_877)}); \
68014   __ret_877; \
68015 })
68016 #else
68017 #define vbfmlaltq_lane_f32(__p0_878, __p1_878, __p2_878, __p3_878) __extension__ ({ \
68018   float32x4_t __ret_878; \
68019   float32x4_t __s0_878 = __p0_878; \
68020   bfloat16x8_t __s1_878 = __p1_878; \
68021   bfloat16x4_t __s2_878 = __p2_878; \
68022   float32x4_t __rev0_878;  __rev0_878 = __builtin_shufflevector(__s0_878, __s0_878, 3, 2, 1, 0); \
68023   bfloat16x8_t __rev1_878;  __rev1_878 = __builtin_shufflevector(__s1_878, __s1_878, 7, 6, 5, 4, 3, 2, 1, 0); \
68024   bfloat16x4_t __rev2_878;  __rev2_878 = __builtin_shufflevector(__s2_878, __s2_878, 3, 2, 1, 0); \
68025   __ret_878 = __noswap_vbfmlaltq_f32(__rev0_878, __rev1_878, (bfloat16x8_t) {__noswap_vget_lane_bf16(__rev2_878, __p3_878), __noswap_vget_lane_bf16(__rev2_878, __p3_878), __noswap_vget_lane_bf16(__rev2_878, __p3_878), __noswap_vget_lane_bf16(__rev2_878, __p3_878), __noswap_vget_lane_bf16(__rev2_878, __p3_878), __noswap_vget_lane_bf16(__rev2_878, __p3_878), __noswap_vget_lane_bf16(__rev2_878, __p3_878), __noswap_vget_lane_bf16(__rev2_878, __p3_878)}); \
68026   __ret_878 = __builtin_shufflevector(__ret_878, __ret_878, 3, 2, 1, 0); \
68027   __ret_878; \
68028 })
68029 #endif
68030 
68031 #ifdef __LITTLE_ENDIAN__
68032 #define vbfmlaltq_laneq_f32(__p0_879, __p1_879, __p2_879, __p3_879) __extension__ ({ \
68033   float32x4_t __ret_879; \
68034   float32x4_t __s0_879 = __p0_879; \
68035   bfloat16x8_t __s1_879 = __p1_879; \
68036   bfloat16x8_t __s2_879 = __p2_879; \
68037   __ret_879 = vbfmlaltq_f32(__s0_879, __s1_879, (bfloat16x8_t) {vgetq_lane_bf16(__s2_879, __p3_879), vgetq_lane_bf16(__s2_879, __p3_879), vgetq_lane_bf16(__s2_879, __p3_879), vgetq_lane_bf16(__s2_879, __p3_879), vgetq_lane_bf16(__s2_879, __p3_879), vgetq_lane_bf16(__s2_879, __p3_879), vgetq_lane_bf16(__s2_879, __p3_879), vgetq_lane_bf16(__s2_879, __p3_879)}); \
68038   __ret_879; \
68039 })
68040 #else
68041 #define vbfmlaltq_laneq_f32(__p0_880, __p1_880, __p2_880, __p3_880) __extension__ ({ \
68042   float32x4_t __ret_880; \
68043   float32x4_t __s0_880 = __p0_880; \
68044   bfloat16x8_t __s1_880 = __p1_880; \
68045   bfloat16x8_t __s2_880 = __p2_880; \
68046   float32x4_t __rev0_880;  __rev0_880 = __builtin_shufflevector(__s0_880, __s0_880, 3, 2, 1, 0); \
68047   bfloat16x8_t __rev1_880;  __rev1_880 = __builtin_shufflevector(__s1_880, __s1_880, 7, 6, 5, 4, 3, 2, 1, 0); \
68048   bfloat16x8_t __rev2_880;  __rev2_880 = __builtin_shufflevector(__s2_880, __s2_880, 7, 6, 5, 4, 3, 2, 1, 0); \
68049   __ret_880 = __noswap_vbfmlaltq_f32(__rev0_880, __rev1_880, (bfloat16x8_t) {__noswap_vgetq_lane_bf16(__rev2_880, __p3_880), __noswap_vgetq_lane_bf16(__rev2_880, __p3_880), __noswap_vgetq_lane_bf16(__rev2_880, __p3_880), __noswap_vgetq_lane_bf16(__rev2_880, __p3_880), __noswap_vgetq_lane_bf16(__rev2_880, __p3_880), __noswap_vgetq_lane_bf16(__rev2_880, __p3_880), __noswap_vgetq_lane_bf16(__rev2_880, __p3_880), __noswap_vgetq_lane_bf16(__rev2_880, __p3_880)}); \
68050   __ret_880 = __builtin_shufflevector(__ret_880, __ret_880, 3, 2, 1, 0); \
68051   __ret_880; \
68052 })
68053 #endif
68054 
68055 #ifdef __LITTLE_ENDIAN__
vcvtq_high_f32_bf16(bfloat16x8_t __p0)68056 __ai __attribute__((target("bf16"))) float32x4_t vcvtq_high_f32_bf16(bfloat16x8_t __p0) {
68057   float32x4_t __ret;
68058   __ret = vcvt_f32_bf16(vget_high_bf16(__p0));
68059   return __ret;
68060 }
68061 #else
vcvtq_high_f32_bf16(bfloat16x8_t __p0)68062 __ai __attribute__((target("bf16"))) float32x4_t vcvtq_high_f32_bf16(bfloat16x8_t __p0) {
68063   float32x4_t __ret;
68064   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68065   __ret = __noswap_vcvt_f32_bf16(__noswap_vget_high_bf16(__rev0));
68066   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68067   return __ret;
68068 }
68069 #endif
68070 
68071 #ifdef __LITTLE_ENDIAN__
vcvtq_low_f32_bf16(bfloat16x8_t __p0)68072 __ai __attribute__((target("bf16"))) float32x4_t vcvtq_low_f32_bf16(bfloat16x8_t __p0) {
68073   float32x4_t __ret;
68074   __ret = vcvt_f32_bf16(vget_low_bf16(__p0));
68075   return __ret;
68076 }
68077 #else
vcvtq_low_f32_bf16(bfloat16x8_t __p0)68078 __ai __attribute__((target("bf16"))) float32x4_t vcvtq_low_f32_bf16(bfloat16x8_t __p0) {
68079   float32x4_t __ret;
68080   bfloat16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68081   __ret = __noswap_vcvt_f32_bf16(__noswap_vget_low_bf16(__rev0));
68082   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68083   return __ret;
68084 }
68085 #endif
68086 
68087 #ifdef __LITTLE_ENDIAN__
68088 #define vsudotq_lane_s32(__p0_881, __p1_881, __p2_881, __p3_881) __extension__ ({ \
68089   int32x4_t __ret_881; \
68090   int32x4_t __s0_881 = __p0_881; \
68091   int8x16_t __s1_881 = __p1_881; \
68092   uint8x8_t __s2_881 = __p2_881; \
68093 uint8x8_t __reint_881 = __s2_881; \
68094   __ret_881 = vusdotq_s32(__s0_881, (uint8x16_t)(splatq_lane_s32(*(int32x2_t *) &__reint_881, __p3_881)), __s1_881); \
68095   __ret_881; \
68096 })
68097 #else
68098 #define vsudotq_lane_s32(__p0_882, __p1_882, __p2_882, __p3_882) __extension__ ({ \
68099   int32x4_t __ret_882; \
68100   int32x4_t __s0_882 = __p0_882; \
68101   int8x16_t __s1_882 = __p1_882; \
68102   uint8x8_t __s2_882 = __p2_882; \
68103   int32x4_t __rev0_882;  __rev0_882 = __builtin_shufflevector(__s0_882, __s0_882, 3, 2, 1, 0); \
68104   int8x16_t __rev1_882;  __rev1_882 = __builtin_shufflevector(__s1_882, __s1_882, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
68105   uint8x8_t __rev2_882;  __rev2_882 = __builtin_shufflevector(__s2_882, __s2_882, 7, 6, 5, 4, 3, 2, 1, 0); \
68106 uint8x8_t __reint_882 = __rev2_882; \
68107   __ret_882 = __noswap_vusdotq_s32(__rev0_882, (uint8x16_t)(__noswap_splatq_lane_s32(*(int32x2_t *) &__reint_882, __p3_882)), __rev1_882); \
68108   __ret_882 = __builtin_shufflevector(__ret_882, __ret_882, 3, 2, 1, 0); \
68109   __ret_882; \
68110 })
68111 #endif
68112 
68113 #ifdef __LITTLE_ENDIAN__
68114 #define vsudot_lane_s32(__p0_883, __p1_883, __p2_883, __p3_883) __extension__ ({ \
68115   int32x2_t __ret_883; \
68116   int32x2_t __s0_883 = __p0_883; \
68117   int8x8_t __s1_883 = __p1_883; \
68118   uint8x8_t __s2_883 = __p2_883; \
68119 uint8x8_t __reint_883 = __s2_883; \
68120   __ret_883 = vusdot_s32(__s0_883, (uint8x8_t)(splat_lane_s32(*(int32x2_t *) &__reint_883, __p3_883)), __s1_883); \
68121   __ret_883; \
68122 })
68123 #else
68124 #define vsudot_lane_s32(__p0_884, __p1_884, __p2_884, __p3_884) __extension__ ({ \
68125   int32x2_t __ret_884; \
68126   int32x2_t __s0_884 = __p0_884; \
68127   int8x8_t __s1_884 = __p1_884; \
68128   uint8x8_t __s2_884 = __p2_884; \
68129   int32x2_t __rev0_884;  __rev0_884 = __builtin_shufflevector(__s0_884, __s0_884, 1, 0); \
68130   int8x8_t __rev1_884;  __rev1_884 = __builtin_shufflevector(__s1_884, __s1_884, 7, 6, 5, 4, 3, 2, 1, 0); \
68131   uint8x8_t __rev2_884;  __rev2_884 = __builtin_shufflevector(__s2_884, __s2_884, 7, 6, 5, 4, 3, 2, 1, 0); \
68132 uint8x8_t __reint_884 = __rev2_884; \
68133   __ret_884 = __noswap_vusdot_s32(__rev0_884, (uint8x8_t)(__noswap_splat_lane_s32(*(int32x2_t *) &__reint_884, __p3_884)), __rev1_884); \
68134   __ret_884 = __builtin_shufflevector(__ret_884, __ret_884, 1, 0); \
68135   __ret_884; \
68136 })
68137 #endif
68138 
68139 #if defined(__aarch64__)
68140 #ifdef __LITTLE_ENDIAN__
vabdl_high_u8(uint8x16_t __p0,uint8x16_t __p1)68141 __ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68142   uint16x8_t __ret;
68143   __ret = vabdl_u8(vget_high_u8(__p0), vget_high_u8(__p1));
68144   return __ret;
68145 }
68146 #else
vabdl_high_u8(uint8x16_t __p0,uint8x16_t __p1)68147 __ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68148   uint16x8_t __ret;
68149   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68150   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68151   __ret = __noswap_vabdl_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
68152   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68153   return __ret;
68154 }
68155 #endif
68156 
68157 #ifdef __LITTLE_ENDIAN__
vabdl_high_u32(uint32x4_t __p0,uint32x4_t __p1)68158 __ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68159   uint64x2_t __ret;
68160   __ret = vabdl_u32(vget_high_u32(__p0), vget_high_u32(__p1));
68161   return __ret;
68162 }
68163 #else
vabdl_high_u32(uint32x4_t __p0,uint32x4_t __p1)68164 __ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68165   uint64x2_t __ret;
68166   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68167   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68168   __ret = __noswap_vabdl_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
68169   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68170   return __ret;
68171 }
68172 #endif
68173 
68174 #ifdef __LITTLE_ENDIAN__
vabdl_high_u16(uint16x8_t __p0,uint16x8_t __p1)68175 __ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68176   uint32x4_t __ret;
68177   __ret = vabdl_u16(vget_high_u16(__p0), vget_high_u16(__p1));
68178   return __ret;
68179 }
68180 #else
vabdl_high_u16(uint16x8_t __p0,uint16x8_t __p1)68181 __ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68182   uint32x4_t __ret;
68183   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68184   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68185   __ret = __noswap_vabdl_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
68186   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68187   return __ret;
68188 }
68189 #endif
68190 
68191 #ifdef __LITTLE_ENDIAN__
vabdl_high_s8(int8x16_t __p0,int8x16_t __p1)68192 __ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68193   int16x8_t __ret;
68194   __ret = vabdl_s8(vget_high_s8(__p0), vget_high_s8(__p1));
68195   return __ret;
68196 }
68197 #else
vabdl_high_s8(int8x16_t __p0,int8x16_t __p1)68198 __ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68199   int16x8_t __ret;
68200   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68201   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68202   __ret = __noswap_vabdl_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
68203   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68204   return __ret;
68205 }
68206 #endif
68207 
68208 #ifdef __LITTLE_ENDIAN__
vabdl_high_s32(int32x4_t __p0,int32x4_t __p1)68209 __ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68210   int64x2_t __ret;
68211   __ret = vabdl_s32(vget_high_s32(__p0), vget_high_s32(__p1));
68212   return __ret;
68213 }
68214 #else
vabdl_high_s32(int32x4_t __p0,int32x4_t __p1)68215 __ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68216   int64x2_t __ret;
68217   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68218   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68219   __ret = __noswap_vabdl_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
68220   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68221   return __ret;
68222 }
68223 #endif
68224 
68225 #ifdef __LITTLE_ENDIAN__
vabdl_high_s16(int16x8_t __p0,int16x8_t __p1)68226 __ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68227   int32x4_t __ret;
68228   __ret = vabdl_s16(vget_high_s16(__p0), vget_high_s16(__p1));
68229   return __ret;
68230 }
68231 #else
vabdl_high_s16(int16x8_t __p0,int16x8_t __p1)68232 __ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68233   int32x4_t __ret;
68234   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68235   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68236   __ret = __noswap_vabdl_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
68237   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68238   return __ret;
68239 }
68240 #endif
68241 
68242 #ifdef __LITTLE_ENDIAN__
vaddl_high_u8(uint8x16_t __p0,uint8x16_t __p1)68243 __ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68244   uint16x8_t __ret;
68245   __ret = vmovl_high_u8(__p0) + vmovl_high_u8(__p1);
68246   return __ret;
68247 }
68248 #else
vaddl_high_u8(uint8x16_t __p0,uint8x16_t __p1)68249 __ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68250   uint16x8_t __ret;
68251   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68252   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68253   __ret = __noswap_vmovl_high_u8(__rev0) + __noswap_vmovl_high_u8(__rev1);
68254   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68255   return __ret;
68256 }
68257 #endif
68258 
68259 #ifdef __LITTLE_ENDIAN__
vaddl_high_u32(uint32x4_t __p0,uint32x4_t __p1)68260 __ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68261   uint64x2_t __ret;
68262   __ret = vmovl_high_u32(__p0) + vmovl_high_u32(__p1);
68263   return __ret;
68264 }
68265 #else
vaddl_high_u32(uint32x4_t __p0,uint32x4_t __p1)68266 __ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68267   uint64x2_t __ret;
68268   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68269   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68270   __ret = __noswap_vmovl_high_u32(__rev0) + __noswap_vmovl_high_u32(__rev1);
68271   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68272   return __ret;
68273 }
68274 #endif
68275 
68276 #ifdef __LITTLE_ENDIAN__
vaddl_high_u16(uint16x8_t __p0,uint16x8_t __p1)68277 __ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68278   uint32x4_t __ret;
68279   __ret = vmovl_high_u16(__p0) + vmovl_high_u16(__p1);
68280   return __ret;
68281 }
68282 #else
vaddl_high_u16(uint16x8_t __p0,uint16x8_t __p1)68283 __ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68284   uint32x4_t __ret;
68285   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68286   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68287   __ret = __noswap_vmovl_high_u16(__rev0) + __noswap_vmovl_high_u16(__rev1);
68288   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68289   return __ret;
68290 }
68291 #endif
68292 
68293 #ifdef __LITTLE_ENDIAN__
vaddl_high_s8(int8x16_t __p0,int8x16_t __p1)68294 __ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68295   int16x8_t __ret;
68296   __ret = vmovl_high_s8(__p0) + vmovl_high_s8(__p1);
68297   return __ret;
68298 }
68299 #else
vaddl_high_s8(int8x16_t __p0,int8x16_t __p1)68300 __ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68301   int16x8_t __ret;
68302   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68303   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68304   __ret = __noswap_vmovl_high_s8(__rev0) + __noswap_vmovl_high_s8(__rev1);
68305   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68306   return __ret;
68307 }
68308 #endif
68309 
68310 #ifdef __LITTLE_ENDIAN__
vaddl_high_s32(int32x4_t __p0,int32x4_t __p1)68311 __ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68312   int64x2_t __ret;
68313   __ret = vmovl_high_s32(__p0) + vmovl_high_s32(__p1);
68314   return __ret;
68315 }
68316 #else
vaddl_high_s32(int32x4_t __p0,int32x4_t __p1)68317 __ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68318   int64x2_t __ret;
68319   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68320   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68321   __ret = __noswap_vmovl_high_s32(__rev0) + __noswap_vmovl_high_s32(__rev1);
68322   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68323   return __ret;
68324 }
68325 #endif
68326 
68327 #ifdef __LITTLE_ENDIAN__
vaddl_high_s16(int16x8_t __p0,int16x8_t __p1)68328 __ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68329   int32x4_t __ret;
68330   __ret = vmovl_high_s16(__p0) + vmovl_high_s16(__p1);
68331   return __ret;
68332 }
68333 #else
vaddl_high_s16(int16x8_t __p0,int16x8_t __p1)68334 __ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68335   int32x4_t __ret;
68336   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68337   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68338   __ret = __noswap_vmovl_high_s16(__rev0) + __noswap_vmovl_high_s16(__rev1);
68339   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68340   return __ret;
68341 }
68342 #endif
68343 
68344 #ifdef __LITTLE_ENDIAN__
vaddw_high_u8(uint16x8_t __p0,uint8x16_t __p1)68345 __ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
68346   uint16x8_t __ret;
68347   __ret = __p0 + vmovl_high_u8(__p1);
68348   return __ret;
68349 }
68350 #else
vaddw_high_u8(uint16x8_t __p0,uint8x16_t __p1)68351 __ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
68352   uint16x8_t __ret;
68353   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68354   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68355   __ret = __rev0 + __noswap_vmovl_high_u8(__rev1);
68356   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68357   return __ret;
68358 }
68359 #endif
68360 
68361 #ifdef __LITTLE_ENDIAN__
vaddw_high_u32(uint64x2_t __p0,uint32x4_t __p1)68362 __ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
68363   uint64x2_t __ret;
68364   __ret = __p0 + vmovl_high_u32(__p1);
68365   return __ret;
68366 }
68367 #else
vaddw_high_u32(uint64x2_t __p0,uint32x4_t __p1)68368 __ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
68369   uint64x2_t __ret;
68370   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68371   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68372   __ret = __rev0 + __noswap_vmovl_high_u32(__rev1);
68373   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68374   return __ret;
68375 }
68376 #endif
68377 
68378 #ifdef __LITTLE_ENDIAN__
vaddw_high_u16(uint32x4_t __p0,uint16x8_t __p1)68379 __ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
68380   uint32x4_t __ret;
68381   __ret = __p0 + vmovl_high_u16(__p1);
68382   return __ret;
68383 }
68384 #else
vaddw_high_u16(uint32x4_t __p0,uint16x8_t __p1)68385 __ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
68386   uint32x4_t __ret;
68387   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68388   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68389   __ret = __rev0 + __noswap_vmovl_high_u16(__rev1);
68390   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68391   return __ret;
68392 }
68393 #endif
68394 
68395 #ifdef __LITTLE_ENDIAN__
vaddw_high_s8(int16x8_t __p0,int8x16_t __p1)68396 __ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
68397   int16x8_t __ret;
68398   __ret = __p0 + vmovl_high_s8(__p1);
68399   return __ret;
68400 }
68401 #else
vaddw_high_s8(int16x8_t __p0,int8x16_t __p1)68402 __ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
68403   int16x8_t __ret;
68404   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68405   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68406   __ret = __rev0 + __noswap_vmovl_high_s8(__rev1);
68407   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68408   return __ret;
68409 }
68410 #endif
68411 
68412 #ifdef __LITTLE_ENDIAN__
vaddw_high_s32(int64x2_t __p0,int32x4_t __p1)68413 __ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
68414   int64x2_t __ret;
68415   __ret = __p0 + vmovl_high_s32(__p1);
68416   return __ret;
68417 }
68418 #else
vaddw_high_s32(int64x2_t __p0,int32x4_t __p1)68419 __ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
68420   int64x2_t __ret;
68421   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68422   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68423   __ret = __rev0 + __noswap_vmovl_high_s32(__rev1);
68424   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68425   return __ret;
68426 }
68427 #endif
68428 
68429 #ifdef __LITTLE_ENDIAN__
vaddw_high_s16(int32x4_t __p0,int16x8_t __p1)68430 __ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
68431   int32x4_t __ret;
68432   __ret = __p0 + vmovl_high_s16(__p1);
68433   return __ret;
68434 }
68435 #else
vaddw_high_s16(int32x4_t __p0,int16x8_t __p1)68436 __ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
68437   int32x4_t __ret;
68438   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68439   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68440   __ret = __rev0 + __noswap_vmovl_high_s16(__rev1);
68441   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68442   return __ret;
68443 }
68444 #endif
68445 
68446 #ifdef __LITTLE_ENDIAN__
68447 #define vcopyq_lane_p64(__p0_885, __p1_885, __p2_885, __p3_885) __extension__ ({ \
68448   poly64x2_t __ret_885; \
68449   poly64x2_t __s0_885 = __p0_885; \
68450   poly64x1_t __s2_885 = __p2_885; \
68451   __ret_885 = vsetq_lane_p64(vget_lane_p64(__s2_885, __p3_885), __s0_885, __p1_885); \
68452   __ret_885; \
68453 })
68454 #else
68455 #define vcopyq_lane_p64(__p0_886, __p1_886, __p2_886, __p3_886) __extension__ ({ \
68456   poly64x2_t __ret_886; \
68457   poly64x2_t __s0_886 = __p0_886; \
68458   poly64x1_t __s2_886 = __p2_886; \
68459   poly64x2_t __rev0_886;  __rev0_886 = __builtin_shufflevector(__s0_886, __s0_886, 1, 0); \
68460   __ret_886 = __noswap_vsetq_lane_p64(vget_lane_p64(__s2_886, __p3_886), __rev0_886, __p1_886); \
68461   __ret_886 = __builtin_shufflevector(__ret_886, __ret_886, 1, 0); \
68462   __ret_886; \
68463 })
68464 #endif
68465 
68466 #ifdef __LITTLE_ENDIAN__
68467 #define vcopyq_lane_f64(__p0_887, __p1_887, __p2_887, __p3_887) __extension__ ({ \
68468   float64x2_t __ret_887; \
68469   float64x2_t __s0_887 = __p0_887; \
68470   float64x1_t __s2_887 = __p2_887; \
68471   __ret_887 = vsetq_lane_f64(vget_lane_f64(__s2_887, __p3_887), __s0_887, __p1_887); \
68472   __ret_887; \
68473 })
68474 #else
68475 #define vcopyq_lane_f64(__p0_888, __p1_888, __p2_888, __p3_888) __extension__ ({ \
68476   float64x2_t __ret_888; \
68477   float64x2_t __s0_888 = __p0_888; \
68478   float64x1_t __s2_888 = __p2_888; \
68479   float64x2_t __rev0_888;  __rev0_888 = __builtin_shufflevector(__s0_888, __s0_888, 1, 0); \
68480   __ret_888 = __noswap_vsetq_lane_f64(vget_lane_f64(__s2_888, __p3_888), __rev0_888, __p1_888); \
68481   __ret_888 = __builtin_shufflevector(__ret_888, __ret_888, 1, 0); \
68482   __ret_888; \
68483 })
68484 #endif
68485 
68486 #define vcopy_lane_p64(__p0_889, __p1_889, __p2_889, __p3_889) __extension__ ({ \
68487   poly64x1_t __ret_889; \
68488   poly64x1_t __s0_889 = __p0_889; \
68489   poly64x1_t __s2_889 = __p2_889; \
68490   __ret_889 = vset_lane_p64(vget_lane_p64(__s2_889, __p3_889), __s0_889, __p1_889); \
68491   __ret_889; \
68492 })
68493 #define vcopy_lane_f64(__p0_890, __p1_890, __p2_890, __p3_890) __extension__ ({ \
68494   float64x1_t __ret_890; \
68495   float64x1_t __s0_890 = __p0_890; \
68496   float64x1_t __s2_890 = __p2_890; \
68497   __ret_890 = vset_lane_f64(vget_lane_f64(__s2_890, __p3_890), __s0_890, __p1_890); \
68498   __ret_890; \
68499 })
68500 #ifdef __LITTLE_ENDIAN__
68501 #define vcopyq_laneq_p64(__p0_891, __p1_891, __p2_891, __p3_891) __extension__ ({ \
68502   poly64x2_t __ret_891; \
68503   poly64x2_t __s0_891 = __p0_891; \
68504   poly64x2_t __s2_891 = __p2_891; \
68505   __ret_891 = vsetq_lane_p64(vgetq_lane_p64(__s2_891, __p3_891), __s0_891, __p1_891); \
68506   __ret_891; \
68507 })
68508 #else
68509 #define vcopyq_laneq_p64(__p0_892, __p1_892, __p2_892, __p3_892) __extension__ ({ \
68510   poly64x2_t __ret_892; \
68511   poly64x2_t __s0_892 = __p0_892; \
68512   poly64x2_t __s2_892 = __p2_892; \
68513   poly64x2_t __rev0_892;  __rev0_892 = __builtin_shufflevector(__s0_892, __s0_892, 1, 0); \
68514   poly64x2_t __rev2_892;  __rev2_892 = __builtin_shufflevector(__s2_892, __s2_892, 1, 0); \
68515   __ret_892 = __noswap_vsetq_lane_p64(__noswap_vgetq_lane_p64(__rev2_892, __p3_892), __rev0_892, __p1_892); \
68516   __ret_892 = __builtin_shufflevector(__ret_892, __ret_892, 1, 0); \
68517   __ret_892; \
68518 })
68519 #endif
68520 
68521 #ifdef __LITTLE_ENDIAN__
68522 #define vcopyq_laneq_f64(__p0_893, __p1_893, __p2_893, __p3_893) __extension__ ({ \
68523   float64x2_t __ret_893; \
68524   float64x2_t __s0_893 = __p0_893; \
68525   float64x2_t __s2_893 = __p2_893; \
68526   __ret_893 = vsetq_lane_f64(vgetq_lane_f64(__s2_893, __p3_893), __s0_893, __p1_893); \
68527   __ret_893; \
68528 })
68529 #else
68530 #define vcopyq_laneq_f64(__p0_894, __p1_894, __p2_894, __p3_894) __extension__ ({ \
68531   float64x2_t __ret_894; \
68532   float64x2_t __s0_894 = __p0_894; \
68533   float64x2_t __s2_894 = __p2_894; \
68534   float64x2_t __rev0_894;  __rev0_894 = __builtin_shufflevector(__s0_894, __s0_894, 1, 0); \
68535   float64x2_t __rev2_894;  __rev2_894 = __builtin_shufflevector(__s2_894, __s2_894, 1, 0); \
68536   __ret_894 = __noswap_vsetq_lane_f64(__noswap_vgetq_lane_f64(__rev2_894, __p3_894), __rev0_894, __p1_894); \
68537   __ret_894 = __builtin_shufflevector(__ret_894, __ret_894, 1, 0); \
68538   __ret_894; \
68539 })
68540 #endif
68541 
68542 #ifdef __LITTLE_ENDIAN__
68543 #define vcopy_laneq_p64(__p0_895, __p1_895, __p2_895, __p3_895) __extension__ ({ \
68544   poly64x1_t __ret_895; \
68545   poly64x1_t __s0_895 = __p0_895; \
68546   poly64x2_t __s2_895 = __p2_895; \
68547   __ret_895 = vset_lane_p64(vgetq_lane_p64(__s2_895, __p3_895), __s0_895, __p1_895); \
68548   __ret_895; \
68549 })
68550 #else
68551 #define vcopy_laneq_p64(__p0_896, __p1_896, __p2_896, __p3_896) __extension__ ({ \
68552   poly64x1_t __ret_896; \
68553   poly64x1_t __s0_896 = __p0_896; \
68554   poly64x2_t __s2_896 = __p2_896; \
68555   poly64x2_t __rev2_896;  __rev2_896 = __builtin_shufflevector(__s2_896, __s2_896, 1, 0); \
68556   __ret_896 = vset_lane_p64(__noswap_vgetq_lane_p64(__rev2_896, __p3_896), __s0_896, __p1_896); \
68557   __ret_896; \
68558 })
68559 #endif
68560 
68561 #ifdef __LITTLE_ENDIAN__
68562 #define vcopy_laneq_f64(__p0_897, __p1_897, __p2_897, __p3_897) __extension__ ({ \
68563   float64x1_t __ret_897; \
68564   float64x1_t __s0_897 = __p0_897; \
68565   float64x2_t __s2_897 = __p2_897; \
68566   __ret_897 = vset_lane_f64(vgetq_lane_f64(__s2_897, __p3_897), __s0_897, __p1_897); \
68567   __ret_897; \
68568 })
68569 #else
68570 #define vcopy_laneq_f64(__p0_898, __p1_898, __p2_898, __p3_898) __extension__ ({ \
68571   float64x1_t __ret_898; \
68572   float64x1_t __s0_898 = __p0_898; \
68573   float64x2_t __s2_898 = __p2_898; \
68574   float64x2_t __rev2_898;  __rev2_898 = __builtin_shufflevector(__s2_898, __s2_898, 1, 0); \
68575   __ret_898 = vset_lane_f64(__noswap_vgetq_lane_f64(__rev2_898, __p3_898), __s0_898, __p1_898); \
68576   __ret_898; \
68577 })
68578 #endif
68579 
68580 #ifdef __LITTLE_ENDIAN__
vmlal_high_u8(uint16x8_t __p0,uint8x16_t __p1,uint8x16_t __p2)68581 __ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68582   uint16x8_t __ret;
68583   __ret = vmlal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
68584   return __ret;
68585 }
68586 #else
vmlal_high_u8(uint16x8_t __p0,uint8x16_t __p1,uint8x16_t __p2)68587 __ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68588   uint16x8_t __ret;
68589   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68590   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68591   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68592   __ret = __noswap_vmlal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
68593   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68594   return __ret;
68595 }
68596 #endif
68597 
68598 #ifdef __LITTLE_ENDIAN__
vmlal_high_u32(uint64x2_t __p0,uint32x4_t __p1,uint32x4_t __p2)68599 __ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68600   uint64x2_t __ret;
68601   __ret = vmlal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
68602   return __ret;
68603 }
68604 #else
vmlal_high_u32(uint64x2_t __p0,uint32x4_t __p1,uint32x4_t __p2)68605 __ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68606   uint64x2_t __ret;
68607   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68608   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68609   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68610   __ret = __noswap_vmlal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
68611   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68612   return __ret;
68613 }
68614 #endif
68615 
68616 #ifdef __LITTLE_ENDIAN__
vmlal_high_u16(uint32x4_t __p0,uint16x8_t __p1,uint16x8_t __p2)68617 __ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68618   uint32x4_t __ret;
68619   __ret = vmlal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
68620   return __ret;
68621 }
68622 #else
vmlal_high_u16(uint32x4_t __p0,uint16x8_t __p1,uint16x8_t __p2)68623 __ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68624   uint32x4_t __ret;
68625   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68626   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68627   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68628   __ret = __noswap_vmlal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
68629   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68630   return __ret;
68631 }
68632 #endif
68633 
68634 #ifdef __LITTLE_ENDIAN__
vmlal_high_s8(int16x8_t __p0,int8x16_t __p1,int8x16_t __p2)68635 __ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68636   int16x8_t __ret;
68637   __ret = vmlal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
68638   return __ret;
68639 }
68640 #else
vmlal_high_s8(int16x8_t __p0,int8x16_t __p1,int8x16_t __p2)68641 __ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68642   int16x8_t __ret;
68643   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68644   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68645   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68646   __ret = __noswap_vmlal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
68647   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68648   return __ret;
68649 }
68650 #endif
68651 
68652 #ifdef __LITTLE_ENDIAN__
vmlal_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)68653 __ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68654   int64x2_t __ret;
68655   __ret = vmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
68656   return __ret;
68657 }
68658 #else
vmlal_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)68659 __ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68660   int64x2_t __ret;
68661   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68662   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68663   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68664   __ret = __noswap_vmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
68665   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68666   return __ret;
68667 }
68668 #endif
68669 
68670 #ifdef __LITTLE_ENDIAN__
vmlal_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)68671 __ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68672   int32x4_t __ret;
68673   __ret = vmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
68674   return __ret;
68675 }
68676 #else
vmlal_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)68677 __ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68678   int32x4_t __ret;
68679   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68680   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68681   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68682   __ret = __noswap_vmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68683   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68684   return __ret;
68685 }
68686 #endif
68687 
68688 #ifdef __LITTLE_ENDIAN__
vmlal_high_n_u32(uint64x2_t __p0,uint32x4_t __p1,uint32_t __p2)68689 __ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68690   uint64x2_t __ret;
68691   __ret = vmlal_n_u32(__p0, vget_high_u32(__p1), __p2);
68692   return __ret;
68693 }
68694 #else
vmlal_high_n_u32(uint64x2_t __p0,uint32x4_t __p1,uint32_t __p2)68695 __ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68696   uint64x2_t __ret;
68697   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68698   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68699   __ret = __noswap_vmlal_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
68700   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68701   return __ret;
68702 }
68703 #endif
68704 
68705 #ifdef __LITTLE_ENDIAN__
vmlal_high_n_u16(uint32x4_t __p0,uint16x8_t __p1,uint16_t __p2)68706 __ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68707   uint32x4_t __ret;
68708   __ret = vmlal_n_u16(__p0, vget_high_u16(__p1), __p2);
68709   return __ret;
68710 }
68711 #else
vmlal_high_n_u16(uint32x4_t __p0,uint16x8_t __p1,uint16_t __p2)68712 __ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68713   uint32x4_t __ret;
68714   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68715   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68716   __ret = __noswap_vmlal_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
68717   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68718   return __ret;
68719 }
68720 #endif
68721 
68722 #ifdef __LITTLE_ENDIAN__
vmlal_high_n_s32(int64x2_t __p0,int32x4_t __p1,int32_t __p2)68723 __ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68724   int64x2_t __ret;
68725   __ret = vmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
68726   return __ret;
68727 }
68728 #else
vmlal_high_n_s32(int64x2_t __p0,int32x4_t __p1,int32_t __p2)68729 __ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68730   int64x2_t __ret;
68731   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68732   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68733   __ret = __noswap_vmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
68734   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68735   return __ret;
68736 }
68737 #endif
68738 
68739 #ifdef __LITTLE_ENDIAN__
vmlal_high_n_s16(int32x4_t __p0,int16x8_t __p1,int16_t __p2)68740 __ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68741   int32x4_t __ret;
68742   __ret = vmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
68743   return __ret;
68744 }
68745 #else
vmlal_high_n_s16(int32x4_t __p0,int16x8_t __p1,int16_t __p2)68746 __ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68747   int32x4_t __ret;
68748   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68749   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68750   __ret = __noswap_vmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
68751   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68752   return __ret;
68753 }
68754 #endif
68755 
68756 #ifdef __LITTLE_ENDIAN__
vmlsl_high_u8(uint16x8_t __p0,uint8x16_t __p1,uint8x16_t __p2)68757 __ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68758   uint16x8_t __ret;
68759   __ret = vmlsl_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
68760   return __ret;
68761 }
68762 #else
vmlsl_high_u8(uint16x8_t __p0,uint8x16_t __p1,uint8x16_t __p2)68763 __ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68764   uint16x8_t __ret;
68765   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68766   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68767   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68768   __ret = __noswap_vmlsl_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
68769   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68770   return __ret;
68771 }
68772 #endif
68773 
68774 #ifdef __LITTLE_ENDIAN__
vmlsl_high_u32(uint64x2_t __p0,uint32x4_t __p1,uint32x4_t __p2)68775 __ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68776   uint64x2_t __ret;
68777   __ret = vmlsl_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
68778   return __ret;
68779 }
68780 #else
vmlsl_high_u32(uint64x2_t __p0,uint32x4_t __p1,uint32x4_t __p2)68781 __ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68782   uint64x2_t __ret;
68783   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68784   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68785   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68786   __ret = __noswap_vmlsl_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
68787   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68788   return __ret;
68789 }
68790 #endif
68791 
68792 #ifdef __LITTLE_ENDIAN__
vmlsl_high_u16(uint32x4_t __p0,uint16x8_t __p1,uint16x8_t __p2)68793 __ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68794   uint32x4_t __ret;
68795   __ret = vmlsl_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
68796   return __ret;
68797 }
68798 #else
vmlsl_high_u16(uint32x4_t __p0,uint16x8_t __p1,uint16x8_t __p2)68799 __ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68800   uint32x4_t __ret;
68801   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68802   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68803   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68804   __ret = __noswap_vmlsl_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
68805   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68806   return __ret;
68807 }
68808 #endif
68809 
68810 #ifdef __LITTLE_ENDIAN__
vmlsl_high_s8(int16x8_t __p0,int8x16_t __p1,int8x16_t __p2)68811 __ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68812   int16x8_t __ret;
68813   __ret = vmlsl_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
68814   return __ret;
68815 }
68816 #else
vmlsl_high_s8(int16x8_t __p0,int8x16_t __p1,int8x16_t __p2)68817 __ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68818   int16x8_t __ret;
68819   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68820   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68821   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68822   __ret = __noswap_vmlsl_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
68823   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68824   return __ret;
68825 }
68826 #endif
68827 
68828 #ifdef __LITTLE_ENDIAN__
vmlsl_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)68829 __ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68830   int64x2_t __ret;
68831   __ret = vmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
68832   return __ret;
68833 }
68834 #else
vmlsl_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)68835 __ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68836   int64x2_t __ret;
68837   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68838   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68839   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68840   __ret = __noswap_vmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
68841   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68842   return __ret;
68843 }
68844 #endif
68845 
68846 #ifdef __LITTLE_ENDIAN__
vmlsl_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)68847 __ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68848   int32x4_t __ret;
68849   __ret = vmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
68850   return __ret;
68851 }
68852 #else
vmlsl_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)68853 __ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68854   int32x4_t __ret;
68855   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68856   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68857   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68858   __ret = __noswap_vmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68859   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68860   return __ret;
68861 }
68862 #endif
68863 
68864 #ifdef __LITTLE_ENDIAN__
vmlsl_high_n_u32(uint64x2_t __p0,uint32x4_t __p1,uint32_t __p2)68865 __ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68866   uint64x2_t __ret;
68867   __ret = vmlsl_n_u32(__p0, vget_high_u32(__p1), __p2);
68868   return __ret;
68869 }
68870 #else
vmlsl_high_n_u32(uint64x2_t __p0,uint32x4_t __p1,uint32_t __p2)68871 __ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68872   uint64x2_t __ret;
68873   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68874   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68875   __ret = __noswap_vmlsl_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
68876   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68877   return __ret;
68878 }
68879 #endif
68880 
68881 #ifdef __LITTLE_ENDIAN__
vmlsl_high_n_u16(uint32x4_t __p0,uint16x8_t __p1,uint16_t __p2)68882 __ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68883   uint32x4_t __ret;
68884   __ret = vmlsl_n_u16(__p0, vget_high_u16(__p1), __p2);
68885   return __ret;
68886 }
68887 #else
vmlsl_high_n_u16(uint32x4_t __p0,uint16x8_t __p1,uint16_t __p2)68888 __ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68889   uint32x4_t __ret;
68890   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68891   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68892   __ret = __noswap_vmlsl_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
68893   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68894   return __ret;
68895 }
68896 #endif
68897 
68898 #ifdef __LITTLE_ENDIAN__
vmlsl_high_n_s32(int64x2_t __p0,int32x4_t __p1,int32_t __p2)68899 __ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68900   int64x2_t __ret;
68901   __ret = vmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
68902   return __ret;
68903 }
68904 #else
vmlsl_high_n_s32(int64x2_t __p0,int32x4_t __p1,int32_t __p2)68905 __ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68906   int64x2_t __ret;
68907   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68908   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68909   __ret = __noswap_vmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
68910   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68911   return __ret;
68912 }
68913 #endif
68914 
68915 #ifdef __LITTLE_ENDIAN__
vmlsl_high_n_s16(int32x4_t __p0,int16x8_t __p1,int16_t __p2)68916 __ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68917   int32x4_t __ret;
68918   __ret = vmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
68919   return __ret;
68920 }
68921 #else
vmlsl_high_n_s16(int32x4_t __p0,int16x8_t __p1,int16_t __p2)68922 __ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68923   int32x4_t __ret;
68924   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68925   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68926   __ret = __noswap_vmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
68927   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68928   return __ret;
68929 }
68930 #endif
68931 
68932 #define vmulx_lane_f64(__p0_899, __p1_899, __p2_899) __extension__ ({ \
68933   float64x1_t __ret_899; \
68934   float64x1_t __s0_899 = __p0_899; \
68935   float64x1_t __s1_899 = __p1_899; \
68936   float64_t __x_899 = vget_lane_f64(__s0_899, 0); \
68937   float64_t __y_899 = vget_lane_f64(__s1_899, __p2_899); \
68938   float64_t __z_899 = vmulxd_f64(__x_899, __y_899); \
68939   __ret_899 = vset_lane_f64(__z_899, __s0_899, __p2_899); \
68940   __ret_899; \
68941 })
68942 #ifdef __LITTLE_ENDIAN__
68943 #define vmulx_laneq_f64(__p0_900, __p1_900, __p2_900) __extension__ ({ \
68944   float64x1_t __ret_900; \
68945   float64x1_t __s0_900 = __p0_900; \
68946   float64x2_t __s1_900 = __p1_900; \
68947   float64_t __x_900 = vget_lane_f64(__s0_900, 0); \
68948   float64_t __y_900 = vgetq_lane_f64(__s1_900, __p2_900); \
68949   float64_t __z_900 = vmulxd_f64(__x_900, __y_900); \
68950   __ret_900 = vset_lane_f64(__z_900, __s0_900, 0); \
68951   __ret_900; \
68952 })
68953 #else
68954 #define vmulx_laneq_f64(__p0_901, __p1_901, __p2_901) __extension__ ({ \
68955   float64x1_t __ret_901; \
68956   float64x1_t __s0_901 = __p0_901; \
68957   float64x2_t __s1_901 = __p1_901; \
68958   float64x2_t __rev1_901;  __rev1_901 = __builtin_shufflevector(__s1_901, __s1_901, 1, 0); \
68959   float64_t __x_901 = vget_lane_f64(__s0_901, 0); \
68960   float64_t __y_901 = __noswap_vgetq_lane_f64(__rev1_901, __p2_901); \
68961   float64_t __z_901 = vmulxd_f64(__x_901, __y_901); \
68962   __ret_901 = vset_lane_f64(__z_901, __s0_901, 0); \
68963   __ret_901; \
68964 })
68965 #endif
68966 
68967 #ifdef __LITTLE_ENDIAN__
68968 #define vfmlalq_lane_high_f16(__p0_902, __p1_902, __p2_902, __p3_902) __extension__ ({ \
68969   float32x4_t __ret_902; \
68970   float32x4_t __s0_902 = __p0_902; \
68971   float16x8_t __s1_902 = __p1_902; \
68972   float16x4_t __s2_902 = __p2_902; \
68973   __ret_902 = vfmlalq_high_f16(__s0_902, __s1_902, (float16x8_t) {vget_lane_f16(__s2_902, __p3_902), vget_lane_f16(__s2_902, __p3_902), vget_lane_f16(__s2_902, __p3_902), vget_lane_f16(__s2_902, __p3_902), vget_lane_f16(__s2_902, __p3_902), vget_lane_f16(__s2_902, __p3_902), vget_lane_f16(__s2_902, __p3_902), vget_lane_f16(__s2_902, __p3_902)}); \
68974   __ret_902; \
68975 })
68976 #else
68977 #define vfmlalq_lane_high_f16(__p0_903, __p1_903, __p2_903, __p3_903) __extension__ ({ \
68978   float32x4_t __ret_903; \
68979   float32x4_t __s0_903 = __p0_903; \
68980   float16x8_t __s1_903 = __p1_903; \
68981   float16x4_t __s2_903 = __p2_903; \
68982   float32x4_t __rev0_903;  __rev0_903 = __builtin_shufflevector(__s0_903, __s0_903, 3, 2, 1, 0); \
68983   float16x8_t __rev1_903;  __rev1_903 = __builtin_shufflevector(__s1_903, __s1_903, 7, 6, 5, 4, 3, 2, 1, 0); \
68984   float16x4_t __rev2_903;  __rev2_903 = __builtin_shufflevector(__s2_903, __s2_903, 3, 2, 1, 0); \
68985   __ret_903 = __noswap_vfmlalq_high_f16(__rev0_903, __rev1_903, (float16x8_t) {__noswap_vget_lane_f16(__rev2_903, __p3_903), __noswap_vget_lane_f16(__rev2_903, __p3_903), __noswap_vget_lane_f16(__rev2_903, __p3_903), __noswap_vget_lane_f16(__rev2_903, __p3_903), __noswap_vget_lane_f16(__rev2_903, __p3_903), __noswap_vget_lane_f16(__rev2_903, __p3_903), __noswap_vget_lane_f16(__rev2_903, __p3_903), __noswap_vget_lane_f16(__rev2_903, __p3_903)}); \
68986   __ret_903 = __builtin_shufflevector(__ret_903, __ret_903, 3, 2, 1, 0); \
68987   __ret_903; \
68988 })
68989 #endif
68990 
68991 #ifdef __LITTLE_ENDIAN__
68992 #define vfmlal_lane_high_f16(__p0_904, __p1_904, __p2_904, __p3_904) __extension__ ({ \
68993   float32x2_t __ret_904; \
68994   float32x2_t __s0_904 = __p0_904; \
68995   float16x4_t __s1_904 = __p1_904; \
68996   float16x4_t __s2_904 = __p2_904; \
68997   __ret_904 = vfmlal_high_f16(__s0_904, __s1_904, (float16x4_t) {vget_lane_f16(__s2_904, __p3_904), vget_lane_f16(__s2_904, __p3_904), vget_lane_f16(__s2_904, __p3_904), vget_lane_f16(__s2_904, __p3_904)}); \
68998   __ret_904; \
68999 })
69000 #else
69001 #define vfmlal_lane_high_f16(__p0_905, __p1_905, __p2_905, __p3_905) __extension__ ({ \
69002   float32x2_t __ret_905; \
69003   float32x2_t __s0_905 = __p0_905; \
69004   float16x4_t __s1_905 = __p1_905; \
69005   float16x4_t __s2_905 = __p2_905; \
69006   float32x2_t __rev0_905;  __rev0_905 = __builtin_shufflevector(__s0_905, __s0_905, 1, 0); \
69007   float16x4_t __rev1_905;  __rev1_905 = __builtin_shufflevector(__s1_905, __s1_905, 3, 2, 1, 0); \
69008   float16x4_t __rev2_905;  __rev2_905 = __builtin_shufflevector(__s2_905, __s2_905, 3, 2, 1, 0); \
69009   __ret_905 = __noswap_vfmlal_high_f16(__rev0_905, __rev1_905, (float16x4_t) {__noswap_vget_lane_f16(__rev2_905, __p3_905), __noswap_vget_lane_f16(__rev2_905, __p3_905), __noswap_vget_lane_f16(__rev2_905, __p3_905), __noswap_vget_lane_f16(__rev2_905, __p3_905)}); \
69010   __ret_905 = __builtin_shufflevector(__ret_905, __ret_905, 1, 0); \
69011   __ret_905; \
69012 })
69013 #endif
69014 
69015 #ifdef __LITTLE_ENDIAN__
69016 #define vfmlalq_lane_low_f16(__p0_906, __p1_906, __p2_906, __p3_906) __extension__ ({ \
69017   float32x4_t __ret_906; \
69018   float32x4_t __s0_906 = __p0_906; \
69019   float16x8_t __s1_906 = __p1_906; \
69020   float16x4_t __s2_906 = __p2_906; \
69021   __ret_906 = vfmlalq_low_f16(__s0_906, __s1_906, (float16x8_t) {vget_lane_f16(__s2_906, __p3_906), vget_lane_f16(__s2_906, __p3_906), vget_lane_f16(__s2_906, __p3_906), vget_lane_f16(__s2_906, __p3_906), vget_lane_f16(__s2_906, __p3_906), vget_lane_f16(__s2_906, __p3_906), vget_lane_f16(__s2_906, __p3_906), vget_lane_f16(__s2_906, __p3_906)}); \
69022   __ret_906; \
69023 })
69024 #else
69025 #define vfmlalq_lane_low_f16(__p0_907, __p1_907, __p2_907, __p3_907) __extension__ ({ \
69026   float32x4_t __ret_907; \
69027   float32x4_t __s0_907 = __p0_907; \
69028   float16x8_t __s1_907 = __p1_907; \
69029   float16x4_t __s2_907 = __p2_907; \
69030   float32x4_t __rev0_907;  __rev0_907 = __builtin_shufflevector(__s0_907, __s0_907, 3, 2, 1, 0); \
69031   float16x8_t __rev1_907;  __rev1_907 = __builtin_shufflevector(__s1_907, __s1_907, 7, 6, 5, 4, 3, 2, 1, 0); \
69032   float16x4_t __rev2_907;  __rev2_907 = __builtin_shufflevector(__s2_907, __s2_907, 3, 2, 1, 0); \
69033   __ret_907 = __noswap_vfmlalq_low_f16(__rev0_907, __rev1_907, (float16x8_t) {__noswap_vget_lane_f16(__rev2_907, __p3_907), __noswap_vget_lane_f16(__rev2_907, __p3_907), __noswap_vget_lane_f16(__rev2_907, __p3_907), __noswap_vget_lane_f16(__rev2_907, __p3_907), __noswap_vget_lane_f16(__rev2_907, __p3_907), __noswap_vget_lane_f16(__rev2_907, __p3_907), __noswap_vget_lane_f16(__rev2_907, __p3_907), __noswap_vget_lane_f16(__rev2_907, __p3_907)}); \
69034   __ret_907 = __builtin_shufflevector(__ret_907, __ret_907, 3, 2, 1, 0); \
69035   __ret_907; \
69036 })
69037 #endif
69038 
69039 #ifdef __LITTLE_ENDIAN__
69040 #define vfmlal_lane_low_f16(__p0_908, __p1_908, __p2_908, __p3_908) __extension__ ({ \
69041   float32x2_t __ret_908; \
69042   float32x2_t __s0_908 = __p0_908; \
69043   float16x4_t __s1_908 = __p1_908; \
69044   float16x4_t __s2_908 = __p2_908; \
69045   __ret_908 = vfmlal_low_f16(__s0_908, __s1_908, (float16x4_t) {vget_lane_f16(__s2_908, __p3_908), vget_lane_f16(__s2_908, __p3_908), vget_lane_f16(__s2_908, __p3_908), vget_lane_f16(__s2_908, __p3_908)}); \
69046   __ret_908; \
69047 })
69048 #else
69049 #define vfmlal_lane_low_f16(__p0_909, __p1_909, __p2_909, __p3_909) __extension__ ({ \
69050   float32x2_t __ret_909; \
69051   float32x2_t __s0_909 = __p0_909; \
69052   float16x4_t __s1_909 = __p1_909; \
69053   float16x4_t __s2_909 = __p2_909; \
69054   float32x2_t __rev0_909;  __rev0_909 = __builtin_shufflevector(__s0_909, __s0_909, 1, 0); \
69055   float16x4_t __rev1_909;  __rev1_909 = __builtin_shufflevector(__s1_909, __s1_909, 3, 2, 1, 0); \
69056   float16x4_t __rev2_909;  __rev2_909 = __builtin_shufflevector(__s2_909, __s2_909, 3, 2, 1, 0); \
69057   __ret_909 = __noswap_vfmlal_low_f16(__rev0_909, __rev1_909, (float16x4_t) {__noswap_vget_lane_f16(__rev2_909, __p3_909), __noswap_vget_lane_f16(__rev2_909, __p3_909), __noswap_vget_lane_f16(__rev2_909, __p3_909), __noswap_vget_lane_f16(__rev2_909, __p3_909)}); \
69058   __ret_909 = __builtin_shufflevector(__ret_909, __ret_909, 1, 0); \
69059   __ret_909; \
69060 })
69061 #endif
69062 
69063 #ifdef __LITTLE_ENDIAN__
69064 #define vfmlalq_laneq_high_f16(__p0_910, __p1_910, __p2_910, __p3_910) __extension__ ({ \
69065   float32x4_t __ret_910; \
69066   float32x4_t __s0_910 = __p0_910; \
69067   float16x8_t __s1_910 = __p1_910; \
69068   float16x8_t __s2_910 = __p2_910; \
69069   __ret_910 = vfmlalq_high_f16(__s0_910, __s1_910, (float16x8_t) {vgetq_lane_f16(__s2_910, __p3_910), vgetq_lane_f16(__s2_910, __p3_910), vgetq_lane_f16(__s2_910, __p3_910), vgetq_lane_f16(__s2_910, __p3_910), vgetq_lane_f16(__s2_910, __p3_910), vgetq_lane_f16(__s2_910, __p3_910), vgetq_lane_f16(__s2_910, __p3_910), vgetq_lane_f16(__s2_910, __p3_910)}); \
69070   __ret_910; \
69071 })
69072 #else
69073 #define vfmlalq_laneq_high_f16(__p0_911, __p1_911, __p2_911, __p3_911) __extension__ ({ \
69074   float32x4_t __ret_911; \
69075   float32x4_t __s0_911 = __p0_911; \
69076   float16x8_t __s1_911 = __p1_911; \
69077   float16x8_t __s2_911 = __p2_911; \
69078   float32x4_t __rev0_911;  __rev0_911 = __builtin_shufflevector(__s0_911, __s0_911, 3, 2, 1, 0); \
69079   float16x8_t __rev1_911;  __rev1_911 = __builtin_shufflevector(__s1_911, __s1_911, 7, 6, 5, 4, 3, 2, 1, 0); \
69080   float16x8_t __rev2_911;  __rev2_911 = __builtin_shufflevector(__s2_911, __s2_911, 7, 6, 5, 4, 3, 2, 1, 0); \
69081   __ret_911 = __noswap_vfmlalq_high_f16(__rev0_911, __rev1_911, (float16x8_t) {__noswap_vgetq_lane_f16(__rev2_911, __p3_911), __noswap_vgetq_lane_f16(__rev2_911, __p3_911), __noswap_vgetq_lane_f16(__rev2_911, __p3_911), __noswap_vgetq_lane_f16(__rev2_911, __p3_911), __noswap_vgetq_lane_f16(__rev2_911, __p3_911), __noswap_vgetq_lane_f16(__rev2_911, __p3_911), __noswap_vgetq_lane_f16(__rev2_911, __p3_911), __noswap_vgetq_lane_f16(__rev2_911, __p3_911)}); \
69082   __ret_911 = __builtin_shufflevector(__ret_911, __ret_911, 3, 2, 1, 0); \
69083   __ret_911; \
69084 })
69085 #endif
69086 
69087 #ifdef __LITTLE_ENDIAN__
69088 #define vfmlal_laneq_high_f16(__p0_912, __p1_912, __p2_912, __p3_912) __extension__ ({ \
69089   float32x2_t __ret_912; \
69090   float32x2_t __s0_912 = __p0_912; \
69091   float16x4_t __s1_912 = __p1_912; \
69092   float16x8_t __s2_912 = __p2_912; \
69093   __ret_912 = vfmlal_high_f16(__s0_912, __s1_912, (float16x4_t) {vgetq_lane_f16(__s2_912, __p3_912), vgetq_lane_f16(__s2_912, __p3_912), vgetq_lane_f16(__s2_912, __p3_912), vgetq_lane_f16(__s2_912, __p3_912)}); \
69094   __ret_912; \
69095 })
69096 #else
69097 #define vfmlal_laneq_high_f16(__p0_913, __p1_913, __p2_913, __p3_913) __extension__ ({ \
69098   float32x2_t __ret_913; \
69099   float32x2_t __s0_913 = __p0_913; \
69100   float16x4_t __s1_913 = __p1_913; \
69101   float16x8_t __s2_913 = __p2_913; \
69102   float32x2_t __rev0_913;  __rev0_913 = __builtin_shufflevector(__s0_913, __s0_913, 1, 0); \
69103   float16x4_t __rev1_913;  __rev1_913 = __builtin_shufflevector(__s1_913, __s1_913, 3, 2, 1, 0); \
69104   float16x8_t __rev2_913;  __rev2_913 = __builtin_shufflevector(__s2_913, __s2_913, 7, 6, 5, 4, 3, 2, 1, 0); \
69105   __ret_913 = __noswap_vfmlal_high_f16(__rev0_913, __rev1_913, (float16x4_t) {__noswap_vgetq_lane_f16(__rev2_913, __p3_913), __noswap_vgetq_lane_f16(__rev2_913, __p3_913), __noswap_vgetq_lane_f16(__rev2_913, __p3_913), __noswap_vgetq_lane_f16(__rev2_913, __p3_913)}); \
69106   __ret_913 = __builtin_shufflevector(__ret_913, __ret_913, 1, 0); \
69107   __ret_913; \
69108 })
69109 #endif
69110 
69111 #ifdef __LITTLE_ENDIAN__
69112 #define vfmlalq_laneq_low_f16(__p0_914, __p1_914, __p2_914, __p3_914) __extension__ ({ \
69113   float32x4_t __ret_914; \
69114   float32x4_t __s0_914 = __p0_914; \
69115   float16x8_t __s1_914 = __p1_914; \
69116   float16x8_t __s2_914 = __p2_914; \
69117   __ret_914 = vfmlalq_low_f16(__s0_914, __s1_914, (float16x8_t) {vgetq_lane_f16(__s2_914, __p3_914), vgetq_lane_f16(__s2_914, __p3_914), vgetq_lane_f16(__s2_914, __p3_914), vgetq_lane_f16(__s2_914, __p3_914), vgetq_lane_f16(__s2_914, __p3_914), vgetq_lane_f16(__s2_914, __p3_914), vgetq_lane_f16(__s2_914, __p3_914), vgetq_lane_f16(__s2_914, __p3_914)}); \
69118   __ret_914; \
69119 })
69120 #else
69121 #define vfmlalq_laneq_low_f16(__p0_915, __p1_915, __p2_915, __p3_915) __extension__ ({ \
69122   float32x4_t __ret_915; \
69123   float32x4_t __s0_915 = __p0_915; \
69124   float16x8_t __s1_915 = __p1_915; \
69125   float16x8_t __s2_915 = __p2_915; \
69126   float32x4_t __rev0_915;  __rev0_915 = __builtin_shufflevector(__s0_915, __s0_915, 3, 2, 1, 0); \
69127   float16x8_t __rev1_915;  __rev1_915 = __builtin_shufflevector(__s1_915, __s1_915, 7, 6, 5, 4, 3, 2, 1, 0); \
69128   float16x8_t __rev2_915;  __rev2_915 = __builtin_shufflevector(__s2_915, __s2_915, 7, 6, 5, 4, 3, 2, 1, 0); \
69129   __ret_915 = __noswap_vfmlalq_low_f16(__rev0_915, __rev1_915, (float16x8_t) {__noswap_vgetq_lane_f16(__rev2_915, __p3_915), __noswap_vgetq_lane_f16(__rev2_915, __p3_915), __noswap_vgetq_lane_f16(__rev2_915, __p3_915), __noswap_vgetq_lane_f16(__rev2_915, __p3_915), __noswap_vgetq_lane_f16(__rev2_915, __p3_915), __noswap_vgetq_lane_f16(__rev2_915, __p3_915), __noswap_vgetq_lane_f16(__rev2_915, __p3_915), __noswap_vgetq_lane_f16(__rev2_915, __p3_915)}); \
69130   __ret_915 = __builtin_shufflevector(__ret_915, __ret_915, 3, 2, 1, 0); \
69131   __ret_915; \
69132 })
69133 #endif
69134 
69135 #ifdef __LITTLE_ENDIAN__
69136 #define vfmlal_laneq_low_f16(__p0_916, __p1_916, __p2_916, __p3_916) __extension__ ({ \
69137   float32x2_t __ret_916; \
69138   float32x2_t __s0_916 = __p0_916; \
69139   float16x4_t __s1_916 = __p1_916; \
69140   float16x8_t __s2_916 = __p2_916; \
69141   __ret_916 = vfmlal_low_f16(__s0_916, __s1_916, (float16x4_t) {vgetq_lane_f16(__s2_916, __p3_916), vgetq_lane_f16(__s2_916, __p3_916), vgetq_lane_f16(__s2_916, __p3_916), vgetq_lane_f16(__s2_916, __p3_916)}); \
69142   __ret_916; \
69143 })
69144 #else
69145 #define vfmlal_laneq_low_f16(__p0_917, __p1_917, __p2_917, __p3_917) __extension__ ({ \
69146   float32x2_t __ret_917; \
69147   float32x2_t __s0_917 = __p0_917; \
69148   float16x4_t __s1_917 = __p1_917; \
69149   float16x8_t __s2_917 = __p2_917; \
69150   float32x2_t __rev0_917;  __rev0_917 = __builtin_shufflevector(__s0_917, __s0_917, 1, 0); \
69151   float16x4_t __rev1_917;  __rev1_917 = __builtin_shufflevector(__s1_917, __s1_917, 3, 2, 1, 0); \
69152   float16x8_t __rev2_917;  __rev2_917 = __builtin_shufflevector(__s2_917, __s2_917, 7, 6, 5, 4, 3, 2, 1, 0); \
69153   __ret_917 = __noswap_vfmlal_low_f16(__rev0_917, __rev1_917, (float16x4_t) {__noswap_vgetq_lane_f16(__rev2_917, __p3_917), __noswap_vgetq_lane_f16(__rev2_917, __p3_917), __noswap_vgetq_lane_f16(__rev2_917, __p3_917), __noswap_vgetq_lane_f16(__rev2_917, __p3_917)}); \
69154   __ret_917 = __builtin_shufflevector(__ret_917, __ret_917, 1, 0); \
69155   __ret_917; \
69156 })
69157 #endif
69158 
69159 #ifdef __LITTLE_ENDIAN__
69160 #define vfmlslq_lane_high_f16(__p0_918, __p1_918, __p2_918, __p3_918) __extension__ ({ \
69161   float32x4_t __ret_918; \
69162   float32x4_t __s0_918 = __p0_918; \
69163   float16x8_t __s1_918 = __p1_918; \
69164   float16x4_t __s2_918 = __p2_918; \
69165   __ret_918 = vfmlslq_high_f16(__s0_918, __s1_918, (float16x8_t) {vget_lane_f16(__s2_918, __p3_918), vget_lane_f16(__s2_918, __p3_918), vget_lane_f16(__s2_918, __p3_918), vget_lane_f16(__s2_918, __p3_918), vget_lane_f16(__s2_918, __p3_918), vget_lane_f16(__s2_918, __p3_918), vget_lane_f16(__s2_918, __p3_918), vget_lane_f16(__s2_918, __p3_918)}); \
69166   __ret_918; \
69167 })
69168 #else
69169 #define vfmlslq_lane_high_f16(__p0_919, __p1_919, __p2_919, __p3_919) __extension__ ({ \
69170   float32x4_t __ret_919; \
69171   float32x4_t __s0_919 = __p0_919; \
69172   float16x8_t __s1_919 = __p1_919; \
69173   float16x4_t __s2_919 = __p2_919; \
69174   float32x4_t __rev0_919;  __rev0_919 = __builtin_shufflevector(__s0_919, __s0_919, 3, 2, 1, 0); \
69175   float16x8_t __rev1_919;  __rev1_919 = __builtin_shufflevector(__s1_919, __s1_919, 7, 6, 5, 4, 3, 2, 1, 0); \
69176   float16x4_t __rev2_919;  __rev2_919 = __builtin_shufflevector(__s2_919, __s2_919, 3, 2, 1, 0); \
69177   __ret_919 = __noswap_vfmlslq_high_f16(__rev0_919, __rev1_919, (float16x8_t) {__noswap_vget_lane_f16(__rev2_919, __p3_919), __noswap_vget_lane_f16(__rev2_919, __p3_919), __noswap_vget_lane_f16(__rev2_919, __p3_919), __noswap_vget_lane_f16(__rev2_919, __p3_919), __noswap_vget_lane_f16(__rev2_919, __p3_919), __noswap_vget_lane_f16(__rev2_919, __p3_919), __noswap_vget_lane_f16(__rev2_919, __p3_919), __noswap_vget_lane_f16(__rev2_919, __p3_919)}); \
69178   __ret_919 = __builtin_shufflevector(__ret_919, __ret_919, 3, 2, 1, 0); \
69179   __ret_919; \
69180 })
69181 #endif
69182 
69183 #ifdef __LITTLE_ENDIAN__
69184 #define vfmlsl_lane_high_f16(__p0_920, __p1_920, __p2_920, __p3_920) __extension__ ({ \
69185   float32x2_t __ret_920; \
69186   float32x2_t __s0_920 = __p0_920; \
69187   float16x4_t __s1_920 = __p1_920; \
69188   float16x4_t __s2_920 = __p2_920; \
69189   __ret_920 = vfmlsl_high_f16(__s0_920, __s1_920, (float16x4_t) {vget_lane_f16(__s2_920, __p3_920), vget_lane_f16(__s2_920, __p3_920), vget_lane_f16(__s2_920, __p3_920), vget_lane_f16(__s2_920, __p3_920)}); \
69190   __ret_920; \
69191 })
69192 #else
69193 #define vfmlsl_lane_high_f16(__p0_921, __p1_921, __p2_921, __p3_921) __extension__ ({ \
69194   float32x2_t __ret_921; \
69195   float32x2_t __s0_921 = __p0_921; \
69196   float16x4_t __s1_921 = __p1_921; \
69197   float16x4_t __s2_921 = __p2_921; \
69198   float32x2_t __rev0_921;  __rev0_921 = __builtin_shufflevector(__s0_921, __s0_921, 1, 0); \
69199   float16x4_t __rev1_921;  __rev1_921 = __builtin_shufflevector(__s1_921, __s1_921, 3, 2, 1, 0); \
69200   float16x4_t __rev2_921;  __rev2_921 = __builtin_shufflevector(__s2_921, __s2_921, 3, 2, 1, 0); \
69201   __ret_921 = __noswap_vfmlsl_high_f16(__rev0_921, __rev1_921, (float16x4_t) {__noswap_vget_lane_f16(__rev2_921, __p3_921), __noswap_vget_lane_f16(__rev2_921, __p3_921), __noswap_vget_lane_f16(__rev2_921, __p3_921), __noswap_vget_lane_f16(__rev2_921, __p3_921)}); \
69202   __ret_921 = __builtin_shufflevector(__ret_921, __ret_921, 1, 0); \
69203   __ret_921; \
69204 })
69205 #endif
69206 
69207 #ifdef __LITTLE_ENDIAN__
69208 #define vfmlslq_lane_low_f16(__p0_922, __p1_922, __p2_922, __p3_922) __extension__ ({ \
69209   float32x4_t __ret_922; \
69210   float32x4_t __s0_922 = __p0_922; \
69211   float16x8_t __s1_922 = __p1_922; \
69212   float16x4_t __s2_922 = __p2_922; \
69213   __ret_922 = vfmlslq_low_f16(__s0_922, __s1_922, (float16x8_t) {vget_lane_f16(__s2_922, __p3_922), vget_lane_f16(__s2_922, __p3_922), vget_lane_f16(__s2_922, __p3_922), vget_lane_f16(__s2_922, __p3_922), vget_lane_f16(__s2_922, __p3_922), vget_lane_f16(__s2_922, __p3_922), vget_lane_f16(__s2_922, __p3_922), vget_lane_f16(__s2_922, __p3_922)}); \
69214   __ret_922; \
69215 })
69216 #else
69217 #define vfmlslq_lane_low_f16(__p0_923, __p1_923, __p2_923, __p3_923) __extension__ ({ \
69218   float32x4_t __ret_923; \
69219   float32x4_t __s0_923 = __p0_923; \
69220   float16x8_t __s1_923 = __p1_923; \
69221   float16x4_t __s2_923 = __p2_923; \
69222   float32x4_t __rev0_923;  __rev0_923 = __builtin_shufflevector(__s0_923, __s0_923, 3, 2, 1, 0); \
69223   float16x8_t __rev1_923;  __rev1_923 = __builtin_shufflevector(__s1_923, __s1_923, 7, 6, 5, 4, 3, 2, 1, 0); \
69224   float16x4_t __rev2_923;  __rev2_923 = __builtin_shufflevector(__s2_923, __s2_923, 3, 2, 1, 0); \
69225   __ret_923 = __noswap_vfmlslq_low_f16(__rev0_923, __rev1_923, (float16x8_t) {__noswap_vget_lane_f16(__rev2_923, __p3_923), __noswap_vget_lane_f16(__rev2_923, __p3_923), __noswap_vget_lane_f16(__rev2_923, __p3_923), __noswap_vget_lane_f16(__rev2_923, __p3_923), __noswap_vget_lane_f16(__rev2_923, __p3_923), __noswap_vget_lane_f16(__rev2_923, __p3_923), __noswap_vget_lane_f16(__rev2_923, __p3_923), __noswap_vget_lane_f16(__rev2_923, __p3_923)}); \
69226   __ret_923 = __builtin_shufflevector(__ret_923, __ret_923, 3, 2, 1, 0); \
69227   __ret_923; \
69228 })
69229 #endif
69230 
69231 #ifdef __LITTLE_ENDIAN__
69232 #define vfmlsl_lane_low_f16(__p0_924, __p1_924, __p2_924, __p3_924) __extension__ ({ \
69233   float32x2_t __ret_924; \
69234   float32x2_t __s0_924 = __p0_924; \
69235   float16x4_t __s1_924 = __p1_924; \
69236   float16x4_t __s2_924 = __p2_924; \
69237   __ret_924 = vfmlsl_low_f16(__s0_924, __s1_924, (float16x4_t) {vget_lane_f16(__s2_924, __p3_924), vget_lane_f16(__s2_924, __p3_924), vget_lane_f16(__s2_924, __p3_924), vget_lane_f16(__s2_924, __p3_924)}); \
69238   __ret_924; \
69239 })
69240 #else
69241 #define vfmlsl_lane_low_f16(__p0_925, __p1_925, __p2_925, __p3_925) __extension__ ({ \
69242   float32x2_t __ret_925; \
69243   float32x2_t __s0_925 = __p0_925; \
69244   float16x4_t __s1_925 = __p1_925; \
69245   float16x4_t __s2_925 = __p2_925; \
69246   float32x2_t __rev0_925;  __rev0_925 = __builtin_shufflevector(__s0_925, __s0_925, 1, 0); \
69247   float16x4_t __rev1_925;  __rev1_925 = __builtin_shufflevector(__s1_925, __s1_925, 3, 2, 1, 0); \
69248   float16x4_t __rev2_925;  __rev2_925 = __builtin_shufflevector(__s2_925, __s2_925, 3, 2, 1, 0); \
69249   __ret_925 = __noswap_vfmlsl_low_f16(__rev0_925, __rev1_925, (float16x4_t) {__noswap_vget_lane_f16(__rev2_925, __p3_925), __noswap_vget_lane_f16(__rev2_925, __p3_925), __noswap_vget_lane_f16(__rev2_925, __p3_925), __noswap_vget_lane_f16(__rev2_925, __p3_925)}); \
69250   __ret_925 = __builtin_shufflevector(__ret_925, __ret_925, 1, 0); \
69251   __ret_925; \
69252 })
69253 #endif
69254 
69255 #ifdef __LITTLE_ENDIAN__
69256 #define vfmlslq_laneq_high_f16(__p0_926, __p1_926, __p2_926, __p3_926) __extension__ ({ \
69257   float32x4_t __ret_926; \
69258   float32x4_t __s0_926 = __p0_926; \
69259   float16x8_t __s1_926 = __p1_926; \
69260   float16x8_t __s2_926 = __p2_926; \
69261   __ret_926 = vfmlslq_high_f16(__s0_926, __s1_926, (float16x8_t) {vgetq_lane_f16(__s2_926, __p3_926), vgetq_lane_f16(__s2_926, __p3_926), vgetq_lane_f16(__s2_926, __p3_926), vgetq_lane_f16(__s2_926, __p3_926), vgetq_lane_f16(__s2_926, __p3_926), vgetq_lane_f16(__s2_926, __p3_926), vgetq_lane_f16(__s2_926, __p3_926), vgetq_lane_f16(__s2_926, __p3_926)}); \
69262   __ret_926; \
69263 })
69264 #else
69265 #define vfmlslq_laneq_high_f16(__p0_927, __p1_927, __p2_927, __p3_927) __extension__ ({ \
69266   float32x4_t __ret_927; \
69267   float32x4_t __s0_927 = __p0_927; \
69268   float16x8_t __s1_927 = __p1_927; \
69269   float16x8_t __s2_927 = __p2_927; \
69270   float32x4_t __rev0_927;  __rev0_927 = __builtin_shufflevector(__s0_927, __s0_927, 3, 2, 1, 0); \
69271   float16x8_t __rev1_927;  __rev1_927 = __builtin_shufflevector(__s1_927, __s1_927, 7, 6, 5, 4, 3, 2, 1, 0); \
69272   float16x8_t __rev2_927;  __rev2_927 = __builtin_shufflevector(__s2_927, __s2_927, 7, 6, 5, 4, 3, 2, 1, 0); \
69273   __ret_927 = __noswap_vfmlslq_high_f16(__rev0_927, __rev1_927, (float16x8_t) {__noswap_vgetq_lane_f16(__rev2_927, __p3_927), __noswap_vgetq_lane_f16(__rev2_927, __p3_927), __noswap_vgetq_lane_f16(__rev2_927, __p3_927), __noswap_vgetq_lane_f16(__rev2_927, __p3_927), __noswap_vgetq_lane_f16(__rev2_927, __p3_927), __noswap_vgetq_lane_f16(__rev2_927, __p3_927), __noswap_vgetq_lane_f16(__rev2_927, __p3_927), __noswap_vgetq_lane_f16(__rev2_927, __p3_927)}); \
69274   __ret_927 = __builtin_shufflevector(__ret_927, __ret_927, 3, 2, 1, 0); \
69275   __ret_927; \
69276 })
69277 #endif
69278 
69279 #ifdef __LITTLE_ENDIAN__
69280 #define vfmlsl_laneq_high_f16(__p0_928, __p1_928, __p2_928, __p3_928) __extension__ ({ \
69281   float32x2_t __ret_928; \
69282   float32x2_t __s0_928 = __p0_928; \
69283   float16x4_t __s1_928 = __p1_928; \
69284   float16x8_t __s2_928 = __p2_928; \
69285   __ret_928 = vfmlsl_high_f16(__s0_928, __s1_928, (float16x4_t) {vgetq_lane_f16(__s2_928, __p3_928), vgetq_lane_f16(__s2_928, __p3_928), vgetq_lane_f16(__s2_928, __p3_928), vgetq_lane_f16(__s2_928, __p3_928)}); \
69286   __ret_928; \
69287 })
69288 #else
69289 #define vfmlsl_laneq_high_f16(__p0_929, __p1_929, __p2_929, __p3_929) __extension__ ({ \
69290   float32x2_t __ret_929; \
69291   float32x2_t __s0_929 = __p0_929; \
69292   float16x4_t __s1_929 = __p1_929; \
69293   float16x8_t __s2_929 = __p2_929; \
69294   float32x2_t __rev0_929;  __rev0_929 = __builtin_shufflevector(__s0_929, __s0_929, 1, 0); \
69295   float16x4_t __rev1_929;  __rev1_929 = __builtin_shufflevector(__s1_929, __s1_929, 3, 2, 1, 0); \
69296   float16x8_t __rev2_929;  __rev2_929 = __builtin_shufflevector(__s2_929, __s2_929, 7, 6, 5, 4, 3, 2, 1, 0); \
69297   __ret_929 = __noswap_vfmlsl_high_f16(__rev0_929, __rev1_929, (float16x4_t) {__noswap_vgetq_lane_f16(__rev2_929, __p3_929), __noswap_vgetq_lane_f16(__rev2_929, __p3_929), __noswap_vgetq_lane_f16(__rev2_929, __p3_929), __noswap_vgetq_lane_f16(__rev2_929, __p3_929)}); \
69298   __ret_929 = __builtin_shufflevector(__ret_929, __ret_929, 1, 0); \
69299   __ret_929; \
69300 })
69301 #endif
69302 
69303 #ifdef __LITTLE_ENDIAN__
69304 #define vfmlslq_laneq_low_f16(__p0_930, __p1_930, __p2_930, __p3_930) __extension__ ({ \
69305   float32x4_t __ret_930; \
69306   float32x4_t __s0_930 = __p0_930; \
69307   float16x8_t __s1_930 = __p1_930; \
69308   float16x8_t __s2_930 = __p2_930; \
69309   __ret_930 = vfmlslq_low_f16(__s0_930, __s1_930, (float16x8_t) {vgetq_lane_f16(__s2_930, __p3_930), vgetq_lane_f16(__s2_930, __p3_930), vgetq_lane_f16(__s2_930, __p3_930), vgetq_lane_f16(__s2_930, __p3_930), vgetq_lane_f16(__s2_930, __p3_930), vgetq_lane_f16(__s2_930, __p3_930), vgetq_lane_f16(__s2_930, __p3_930), vgetq_lane_f16(__s2_930, __p3_930)}); \
69310   __ret_930; \
69311 })
69312 #else
69313 #define vfmlslq_laneq_low_f16(__p0_931, __p1_931, __p2_931, __p3_931) __extension__ ({ \
69314   float32x4_t __ret_931; \
69315   float32x4_t __s0_931 = __p0_931; \
69316   float16x8_t __s1_931 = __p1_931; \
69317   float16x8_t __s2_931 = __p2_931; \
69318   float32x4_t __rev0_931;  __rev0_931 = __builtin_shufflevector(__s0_931, __s0_931, 3, 2, 1, 0); \
69319   float16x8_t __rev1_931;  __rev1_931 = __builtin_shufflevector(__s1_931, __s1_931, 7, 6, 5, 4, 3, 2, 1, 0); \
69320   float16x8_t __rev2_931;  __rev2_931 = __builtin_shufflevector(__s2_931, __s2_931, 7, 6, 5, 4, 3, 2, 1, 0); \
69321   __ret_931 = __noswap_vfmlslq_low_f16(__rev0_931, __rev1_931, (float16x8_t) {__noswap_vgetq_lane_f16(__rev2_931, __p3_931), __noswap_vgetq_lane_f16(__rev2_931, __p3_931), __noswap_vgetq_lane_f16(__rev2_931, __p3_931), __noswap_vgetq_lane_f16(__rev2_931, __p3_931), __noswap_vgetq_lane_f16(__rev2_931, __p3_931), __noswap_vgetq_lane_f16(__rev2_931, __p3_931), __noswap_vgetq_lane_f16(__rev2_931, __p3_931), __noswap_vgetq_lane_f16(__rev2_931, __p3_931)}); \
69322   __ret_931 = __builtin_shufflevector(__ret_931, __ret_931, 3, 2, 1, 0); \
69323   __ret_931; \
69324 })
69325 #endif
69326 
69327 #ifdef __LITTLE_ENDIAN__
69328 #define vfmlsl_laneq_low_f16(__p0_932, __p1_932, __p2_932, __p3_932) __extension__ ({ \
69329   float32x2_t __ret_932; \
69330   float32x2_t __s0_932 = __p0_932; \
69331   float16x4_t __s1_932 = __p1_932; \
69332   float16x8_t __s2_932 = __p2_932; \
69333   __ret_932 = vfmlsl_low_f16(__s0_932, __s1_932, (float16x4_t) {vgetq_lane_f16(__s2_932, __p3_932), vgetq_lane_f16(__s2_932, __p3_932), vgetq_lane_f16(__s2_932, __p3_932), vgetq_lane_f16(__s2_932, __p3_932)}); \
69334   __ret_932; \
69335 })
69336 #else
69337 #define vfmlsl_laneq_low_f16(__p0_933, __p1_933, __p2_933, __p3_933) __extension__ ({ \
69338   float32x2_t __ret_933; \
69339   float32x2_t __s0_933 = __p0_933; \
69340   float16x4_t __s1_933 = __p1_933; \
69341   float16x8_t __s2_933 = __p2_933; \
69342   float32x2_t __rev0_933;  __rev0_933 = __builtin_shufflevector(__s0_933, __s0_933, 1, 0); \
69343   float16x4_t __rev1_933;  __rev1_933 = __builtin_shufflevector(__s1_933, __s1_933, 3, 2, 1, 0); \
69344   float16x8_t __rev2_933;  __rev2_933 = __builtin_shufflevector(__s2_933, __s2_933, 7, 6, 5, 4, 3, 2, 1, 0); \
69345   __ret_933 = __noswap_vfmlsl_low_f16(__rev0_933, __rev1_933, (float16x4_t) {__noswap_vgetq_lane_f16(__rev2_933, __p3_933), __noswap_vgetq_lane_f16(__rev2_933, __p3_933), __noswap_vgetq_lane_f16(__rev2_933, __p3_933), __noswap_vgetq_lane_f16(__rev2_933, __p3_933)}); \
69346   __ret_933 = __builtin_shufflevector(__ret_933, __ret_933, 1, 0); \
69347   __ret_933; \
69348 })
69349 #endif
69350 
69351 #ifdef __LITTLE_ENDIAN__
69352 #define vmulh_lane_f16(__p0_934, __p1_934, __p2_934) __extension__ ({ \
69353   float16_t __ret_934; \
69354   float16_t __s0_934 = __p0_934; \
69355   float16x4_t __s1_934 = __p1_934; \
69356   __ret_934 = __s0_934 * vget_lane_f16(__s1_934, __p2_934); \
69357   __ret_934; \
69358 })
69359 #else
69360 #define vmulh_lane_f16(__p0_935, __p1_935, __p2_935) __extension__ ({ \
69361   float16_t __ret_935; \
69362   float16_t __s0_935 = __p0_935; \
69363   float16x4_t __s1_935 = __p1_935; \
69364   float16x4_t __rev1_935;  __rev1_935 = __builtin_shufflevector(__s1_935, __s1_935, 3, 2, 1, 0); \
69365   __ret_935 = __s0_935 * __noswap_vget_lane_f16(__rev1_935, __p2_935); \
69366   __ret_935; \
69367 })
69368 #endif
69369 
69370 #ifdef __LITTLE_ENDIAN__
69371 #define vmulh_laneq_f16(__p0_936, __p1_936, __p2_936) __extension__ ({ \
69372   float16_t __ret_936; \
69373   float16_t __s0_936 = __p0_936; \
69374   float16x8_t __s1_936 = __p1_936; \
69375   __ret_936 = __s0_936 * vgetq_lane_f16(__s1_936, __p2_936); \
69376   __ret_936; \
69377 })
69378 #else
69379 #define vmulh_laneq_f16(__p0_937, __p1_937, __p2_937) __extension__ ({ \
69380   float16_t __ret_937; \
69381   float16_t __s0_937 = __p0_937; \
69382   float16x8_t __s1_937 = __p1_937; \
69383   float16x8_t __rev1_937;  __rev1_937 = __builtin_shufflevector(__s1_937, __s1_937, 7, 6, 5, 4, 3, 2, 1, 0); \
69384   __ret_937 = __s0_937 * __noswap_vgetq_lane_f16(__rev1_937, __p2_937); \
69385   __ret_937; \
69386 })
69387 #endif
69388 
69389 #endif
69390 #ifdef __LITTLE_ENDIAN__
vabal_u8(uint16x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)69391 __ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
69392   uint16x8_t __ret;
69393   __ret = __p0 + vabdl_u8(__p1, __p2);
69394   return __ret;
69395 }
69396 #else
vabal_u8(uint16x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)69397 __ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
69398   uint16x8_t __ret;
69399   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69400   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69401   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69402   __ret = __rev0 + __noswap_vabdl_u8(__rev1, __rev2);
69403   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69404   return __ret;
69405 }
__noswap_vabal_u8(uint16x8_t __p0,uint8x8_t __p1,uint8x8_t __p2)69406 __ai uint16x8_t __noswap_vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
69407   uint16x8_t __ret;
69408   __ret = __p0 + __noswap_vabdl_u8(__p1, __p2);
69409   return __ret;
69410 }
69411 #endif
69412 
69413 #ifdef __LITTLE_ENDIAN__
vabal_u32(uint64x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)69414 __ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69415   uint64x2_t __ret;
69416   __ret = __p0 + vabdl_u32(__p1, __p2);
69417   return __ret;
69418 }
69419 #else
vabal_u32(uint64x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)69420 __ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69421   uint64x2_t __ret;
69422   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69423   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
69424   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
69425   __ret = __rev0 + __noswap_vabdl_u32(__rev1, __rev2);
69426   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69427   return __ret;
69428 }
__noswap_vabal_u32(uint64x2_t __p0,uint32x2_t __p1,uint32x2_t __p2)69429 __ai uint64x2_t __noswap_vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69430   uint64x2_t __ret;
69431   __ret = __p0 + __noswap_vabdl_u32(__p1, __p2);
69432   return __ret;
69433 }
69434 #endif
69435 
69436 #ifdef __LITTLE_ENDIAN__
vabal_u16(uint32x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)69437 __ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69438   uint32x4_t __ret;
69439   __ret = __p0 + vabdl_u16(__p1, __p2);
69440   return __ret;
69441 }
69442 #else
vabal_u16(uint32x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)69443 __ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69444   uint32x4_t __ret;
69445   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69446   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69447   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69448   __ret = __rev0 + __noswap_vabdl_u16(__rev1, __rev2);
69449   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69450   return __ret;
69451 }
__noswap_vabal_u16(uint32x4_t __p0,uint16x4_t __p1,uint16x4_t __p2)69452 __ai uint32x4_t __noswap_vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69453   uint32x4_t __ret;
69454   __ret = __p0 + __noswap_vabdl_u16(__p1, __p2);
69455   return __ret;
69456 }
69457 #endif
69458 
69459 #ifdef __LITTLE_ENDIAN__
vabal_s8(int16x8_t __p0,int8x8_t __p1,int8x8_t __p2)69460 __ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69461   int16x8_t __ret;
69462   __ret = __p0 + vabdl_s8(__p1, __p2);
69463   return __ret;
69464 }
69465 #else
vabal_s8(int16x8_t __p0,int8x8_t __p1,int8x8_t __p2)69466 __ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69467   int16x8_t __ret;
69468   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69469   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69470   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69471   __ret = __rev0 + __noswap_vabdl_s8(__rev1, __rev2);
69472   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69473   return __ret;
69474 }
__noswap_vabal_s8(int16x8_t __p0,int8x8_t __p1,int8x8_t __p2)69475 __ai int16x8_t __noswap_vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69476   int16x8_t __ret;
69477   __ret = __p0 + __noswap_vabdl_s8(__p1, __p2);
69478   return __ret;
69479 }
69480 #endif
69481 
69482 #ifdef __LITTLE_ENDIAN__
vabal_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)69483 __ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69484   int64x2_t __ret;
69485   __ret = __p0 + vabdl_s32(__p1, __p2);
69486   return __ret;
69487 }
69488 #else
vabal_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)69489 __ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69490   int64x2_t __ret;
69491   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69492   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
69493   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
69494   __ret = __rev0 + __noswap_vabdl_s32(__rev1, __rev2);
69495   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69496   return __ret;
69497 }
__noswap_vabal_s32(int64x2_t __p0,int32x2_t __p1,int32x2_t __p2)69498 __ai int64x2_t __noswap_vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69499   int64x2_t __ret;
69500   __ret = __p0 + __noswap_vabdl_s32(__p1, __p2);
69501   return __ret;
69502 }
69503 #endif
69504 
69505 #ifdef __LITTLE_ENDIAN__
vabal_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)69506 __ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69507   int32x4_t __ret;
69508   __ret = __p0 + vabdl_s16(__p1, __p2);
69509   return __ret;
69510 }
69511 #else
vabal_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)69512 __ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69513   int32x4_t __ret;
69514   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69515   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69516   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69517   __ret = __rev0 + __noswap_vabdl_s16(__rev1, __rev2);
69518   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69519   return __ret;
69520 }
__noswap_vabal_s16(int32x4_t __p0,int16x4_t __p1,int16x4_t __p2)69521 __ai int32x4_t __noswap_vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69522   int32x4_t __ret;
69523   __ret = __p0 + __noswap_vabdl_s16(__p1, __p2);
69524   return __ret;
69525 }
69526 #endif
69527 
69528 #if defined(__aarch64__)
69529 #ifdef __LITTLE_ENDIAN__
vabal_high_u8(uint16x8_t __p0,uint8x16_t __p1,uint8x16_t __p2)69530 __ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
69531   uint16x8_t __ret;
69532   __ret = vabal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
69533   return __ret;
69534 }
69535 #else
vabal_high_u8(uint16x8_t __p0,uint8x16_t __p1,uint8x16_t __p2)69536 __ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
69537   uint16x8_t __ret;
69538   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69539   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69540   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69541   __ret = __noswap_vabal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
69542   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69543   return __ret;
69544 }
69545 #endif
69546 
69547 #ifdef __LITTLE_ENDIAN__
vabal_high_u32(uint64x2_t __p0,uint32x4_t __p1,uint32x4_t __p2)69548 __ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
69549   uint64x2_t __ret;
69550   __ret = vabal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
69551   return __ret;
69552 }
69553 #else
vabal_high_u32(uint64x2_t __p0,uint32x4_t __p1,uint32x4_t __p2)69554 __ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
69555   uint64x2_t __ret;
69556   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69557   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69558   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69559   __ret = __noswap_vabal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
69560   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69561   return __ret;
69562 }
69563 #endif
69564 
69565 #ifdef __LITTLE_ENDIAN__
vabal_high_u16(uint32x4_t __p0,uint16x8_t __p1,uint16x8_t __p2)69566 __ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
69567   uint32x4_t __ret;
69568   __ret = vabal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
69569   return __ret;
69570 }
69571 #else
vabal_high_u16(uint32x4_t __p0,uint16x8_t __p1,uint16x8_t __p2)69572 __ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
69573   uint32x4_t __ret;
69574   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69575   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69576   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69577   __ret = __noswap_vabal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
69578   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69579   return __ret;
69580 }
69581 #endif
69582 
69583 #ifdef __LITTLE_ENDIAN__
vabal_high_s8(int16x8_t __p0,int8x16_t __p1,int8x16_t __p2)69584 __ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
69585   int16x8_t __ret;
69586   __ret = vabal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
69587   return __ret;
69588 }
69589 #else
vabal_high_s8(int16x8_t __p0,int8x16_t __p1,int8x16_t __p2)69590 __ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
69591   int16x8_t __ret;
69592   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69593   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69594   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69595   __ret = __noswap_vabal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
69596   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69597   return __ret;
69598 }
69599 #endif
69600 
69601 #ifdef __LITTLE_ENDIAN__
vabal_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)69602 __ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
69603   int64x2_t __ret;
69604   __ret = vabal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
69605   return __ret;
69606 }
69607 #else
vabal_high_s32(int64x2_t __p0,int32x4_t __p1,int32x4_t __p2)69608 __ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
69609   int64x2_t __ret;
69610   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69611   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69612   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69613   __ret = __noswap_vabal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
69614   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69615   return __ret;
69616 }
69617 #endif
69618 
69619 #ifdef __LITTLE_ENDIAN__
vabal_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)69620 __ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
69621   int32x4_t __ret;
69622   __ret = vabal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
69623   return __ret;
69624 }
69625 #else
vabal_high_s16(int32x4_t __p0,int16x8_t __p1,int16x8_t __p2)69626 __ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
69627   int32x4_t __ret;
69628   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69629   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69630   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69631   __ret = __noswap_vabal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
69632   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69633   return __ret;
69634 }
69635 #endif
69636 
69637 #endif
69638 
69639 #undef __ai
69640 
69641 #endif /* if !defined(__ARM_NEON) */
69642 #endif /* ifndef __ARM_FP */
69643 #endif /* __ARM_NEON_H */
69644