Home | History | Annotate | Download | only in include
      1 /*===---- arm_neon.h - ARM Neon intrinsics ---------------------------------===
      2  *
      3  * Permission is hereby granted, free of charge, to any person obtaining a copy
      4  * of this software and associated documentation files (the "Software"), to deal
      5  * in the Software without restriction, including without limitation the rights
      6  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
      7  * copies of the Software, and to permit persons to whom the Software is
      8  * furnished to do so, subject to the following conditions:
      9  *
     10  * The above copyright notice and this permission notice shall be included in
     11  * all copies or substantial portions of the Software.
     12  *
     13  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     14  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     15  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
     16  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
     17  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
     18  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
     19  * THE SOFTWARE.
     20  *
     21  *===-----------------------------------------------------------------------===
     22  */
     23 
     24 #ifndef __ARM_NEON_H
     25 #define __ARM_NEON_H
     26 
     27 #if !defined(__ARM_NEON)
     28 #error "NEON support not enabled"
     29 #endif
     30 
     31 #include <stdint.h>
     32 
     33 typedef float float32_t;
     34 typedef __fp16 float16_t;
     35 #ifdef __aarch64__
     36 typedef double float64_t;
     37 #endif
     38 
     39 #ifdef __aarch64__
     40 typedef uint8_t poly8_t;
     41 typedef uint16_t poly16_t;
     42 typedef uint64_t poly64_t;
     43 typedef __uint128_t poly128_t;
     44 #else
     45 typedef int8_t poly8_t;
     46 typedef int16_t poly16_t;
     47 #endif
     48 typedef __attribute__((neon_vector_type(8))) int8_t int8x8_t;
     49 typedef __attribute__((neon_vector_type(16))) int8_t int8x16_t;
     50 typedef __attribute__((neon_vector_type(4))) int16_t int16x4_t;
     51 typedef __attribute__((neon_vector_type(8))) int16_t int16x8_t;
     52 typedef __attribute__((neon_vector_type(2))) int32_t int32x2_t;
     53 typedef __attribute__((neon_vector_type(4))) int32_t int32x4_t;
     54 typedef __attribute__((neon_vector_type(1))) int64_t int64x1_t;
     55 typedef __attribute__((neon_vector_type(2))) int64_t int64x2_t;
     56 typedef __attribute__((neon_vector_type(8))) uint8_t uint8x8_t;
     57 typedef __attribute__((neon_vector_type(16))) uint8_t uint8x16_t;
     58 typedef __attribute__((neon_vector_type(4))) uint16_t uint16x4_t;
     59 typedef __attribute__((neon_vector_type(8))) uint16_t uint16x8_t;
     60 typedef __attribute__((neon_vector_type(2))) uint32_t uint32x2_t;
     61 typedef __attribute__((neon_vector_type(4))) uint32_t uint32x4_t;
     62 typedef __attribute__((neon_vector_type(1))) uint64_t uint64x1_t;
     63 typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t;
     64 typedef __attribute__((neon_vector_type(4))) float16_t float16x4_t;
     65 typedef __attribute__((neon_vector_type(8))) float16_t float16x8_t;
     66 typedef __attribute__((neon_vector_type(2))) float32_t float32x2_t;
     67 typedef __attribute__((neon_vector_type(4))) float32_t float32x4_t;
     68 #ifdef __aarch64__
     69 typedef __attribute__((neon_vector_type(1))) float64_t float64x1_t;
     70 typedef __attribute__((neon_vector_type(2))) float64_t float64x2_t;
     71 #endif
     72 typedef __attribute__((neon_polyvector_type(8))) poly8_t poly8x8_t;
     73 typedef __attribute__((neon_polyvector_type(16))) poly8_t poly8x16_t;
     74 typedef __attribute__((neon_polyvector_type(4))) poly16_t poly16x4_t;
     75 typedef __attribute__((neon_polyvector_type(8))) poly16_t poly16x8_t;
     76 #ifdef __aarch64__
     77 typedef __attribute__((neon_polyvector_type(1))) poly64_t poly64x1_t;
     78 typedef __attribute__((neon_polyvector_type(2))) poly64_t poly64x2_t;
     79 #endif
     80 
     81 typedef struct int8x8x2_t {
     82   int8x8_t val[2];
     83 } int8x8x2_t;
     84 
     85 typedef struct int8x16x2_t {
     86   int8x16_t val[2];
     87 } int8x16x2_t;
     88 
     89 typedef struct int16x4x2_t {
     90   int16x4_t val[2];
     91 } int16x4x2_t;
     92 
     93 typedef struct int16x8x2_t {
     94   int16x8_t val[2];
     95 } int16x8x2_t;
     96 
     97 typedef struct int32x2x2_t {
     98   int32x2_t val[2];
     99 } int32x2x2_t;
    100 
    101 typedef struct int32x4x2_t {
    102   int32x4_t val[2];
    103 } int32x4x2_t;
    104 
    105 typedef struct int64x1x2_t {
    106   int64x1_t val[2];
    107 } int64x1x2_t;
    108 
    109 typedef struct int64x2x2_t {
    110   int64x2_t val[2];
    111 } int64x2x2_t;
    112 
    113 typedef struct uint8x8x2_t {
    114   uint8x8_t val[2];
    115 } uint8x8x2_t;
    116 
    117 typedef struct uint8x16x2_t {
    118   uint8x16_t val[2];
    119 } uint8x16x2_t;
    120 
    121 typedef struct uint16x4x2_t {
    122   uint16x4_t val[2];
    123 } uint16x4x2_t;
    124 
    125 typedef struct uint16x8x2_t {
    126   uint16x8_t val[2];
    127 } uint16x8x2_t;
    128 
    129 typedef struct uint32x2x2_t {
    130   uint32x2_t val[2];
    131 } uint32x2x2_t;
    132 
    133 typedef struct uint32x4x2_t {
    134   uint32x4_t val[2];
    135 } uint32x4x2_t;
    136 
    137 typedef struct uint64x1x2_t {
    138   uint64x1_t val[2];
    139 } uint64x1x2_t;
    140 
    141 typedef struct uint64x2x2_t {
    142   uint64x2_t val[2];
    143 } uint64x2x2_t;
    144 
    145 typedef struct float16x4x2_t {
    146   float16x4_t val[2];
    147 } float16x4x2_t;
    148 
    149 typedef struct float16x8x2_t {
    150   float16x8_t val[2];
    151 } float16x8x2_t;
    152 
    153 typedef struct float32x2x2_t {
    154   float32x2_t val[2];
    155 } float32x2x2_t;
    156 
    157 typedef struct float32x4x2_t {
    158   float32x4_t val[2];
    159 } float32x4x2_t;
    160 
    161 #ifdef __aarch64__
    162 typedef struct float64x1x2_t {
    163   float64x1_t val[2];
    164 } float64x1x2_t;
    165 
    166 typedef struct float64x2x2_t {
    167   float64x2_t val[2];
    168 } float64x2x2_t;
    169 
    170 #endif
    171 typedef struct poly8x8x2_t {
    172   poly8x8_t val[2];
    173 } poly8x8x2_t;
    174 
    175 typedef struct poly8x16x2_t {
    176   poly8x16_t val[2];
    177 } poly8x16x2_t;
    178 
    179 typedef struct poly16x4x2_t {
    180   poly16x4_t val[2];
    181 } poly16x4x2_t;
    182 
    183 typedef struct poly16x8x2_t {
    184   poly16x8_t val[2];
    185 } poly16x8x2_t;
    186 
    187 #ifdef __aarch64__
    188 typedef struct poly64x1x2_t {
    189   poly64x1_t val[2];
    190 } poly64x1x2_t;
    191 
    192 typedef struct poly64x2x2_t {
    193   poly64x2_t val[2];
    194 } poly64x2x2_t;
    195 
    196 #endif
    197 typedef struct int8x8x3_t {
    198   int8x8_t val[3];
    199 } int8x8x3_t;
    200 
    201 typedef struct int8x16x3_t {
    202   int8x16_t val[3];
    203 } int8x16x3_t;
    204 
    205 typedef struct int16x4x3_t {
    206   int16x4_t val[3];
    207 } int16x4x3_t;
    208 
    209 typedef struct int16x8x3_t {
    210   int16x8_t val[3];
    211 } int16x8x3_t;
    212 
    213 typedef struct int32x2x3_t {
    214   int32x2_t val[3];
    215 } int32x2x3_t;
    216 
    217 typedef struct int32x4x3_t {
    218   int32x4_t val[3];
    219 } int32x4x3_t;
    220 
    221 typedef struct int64x1x3_t {
    222   int64x1_t val[3];
    223 } int64x1x3_t;
    224 
    225 typedef struct int64x2x3_t {
    226   int64x2_t val[3];
    227 } int64x2x3_t;
    228 
    229 typedef struct uint8x8x3_t {
    230   uint8x8_t val[3];
    231 } uint8x8x3_t;
    232 
    233 typedef struct uint8x16x3_t {
    234   uint8x16_t val[3];
    235 } uint8x16x3_t;
    236 
    237 typedef struct uint16x4x3_t {
    238   uint16x4_t val[3];
    239 } uint16x4x3_t;
    240 
    241 typedef struct uint16x8x3_t {
    242   uint16x8_t val[3];
    243 } uint16x8x3_t;
    244 
    245 typedef struct uint32x2x3_t {
    246   uint32x2_t val[3];
    247 } uint32x2x3_t;
    248 
    249 typedef struct uint32x4x3_t {
    250   uint32x4_t val[3];
    251 } uint32x4x3_t;
    252 
    253 typedef struct uint64x1x3_t {
    254   uint64x1_t val[3];
    255 } uint64x1x3_t;
    256 
    257 typedef struct uint64x2x3_t {
    258   uint64x2_t val[3];
    259 } uint64x2x3_t;
    260 
    261 typedef struct float16x4x3_t {
    262   float16x4_t val[3];
    263 } float16x4x3_t;
    264 
    265 typedef struct float16x8x3_t {
    266   float16x8_t val[3];
    267 } float16x8x3_t;
    268 
    269 typedef struct float32x2x3_t {
    270   float32x2_t val[3];
    271 } float32x2x3_t;
    272 
    273 typedef struct float32x4x3_t {
    274   float32x4_t val[3];
    275 } float32x4x3_t;
    276 
    277 #ifdef __aarch64__
    278 typedef struct float64x1x3_t {
    279   float64x1_t val[3];
    280 } float64x1x3_t;
    281 
    282 typedef struct float64x2x3_t {
    283   float64x2_t val[3];
    284 } float64x2x3_t;
    285 
    286 #endif
    287 typedef struct poly8x8x3_t {
    288   poly8x8_t val[3];
    289 } poly8x8x3_t;
    290 
    291 typedef struct poly8x16x3_t {
    292   poly8x16_t val[3];
    293 } poly8x16x3_t;
    294 
    295 typedef struct poly16x4x3_t {
    296   poly16x4_t val[3];
    297 } poly16x4x3_t;
    298 
    299 typedef struct poly16x8x3_t {
    300   poly16x8_t val[3];
    301 } poly16x8x3_t;
    302 
    303 #ifdef __aarch64__
    304 typedef struct poly64x1x3_t {
    305   poly64x1_t val[3];
    306 } poly64x1x3_t;
    307 
    308 typedef struct poly64x2x3_t {
    309   poly64x2_t val[3];
    310 } poly64x2x3_t;
    311 
    312 #endif
    313 typedef struct int8x8x4_t {
    314   int8x8_t val[4];
    315 } int8x8x4_t;
    316 
    317 typedef struct int8x16x4_t {
    318   int8x16_t val[4];
    319 } int8x16x4_t;
    320 
    321 typedef struct int16x4x4_t {
    322   int16x4_t val[4];
    323 } int16x4x4_t;
    324 
    325 typedef struct int16x8x4_t {
    326   int16x8_t val[4];
    327 } int16x8x4_t;
    328 
    329 typedef struct int32x2x4_t {
    330   int32x2_t val[4];
    331 } int32x2x4_t;
    332 
    333 typedef struct int32x4x4_t {
    334   int32x4_t val[4];
    335 } int32x4x4_t;
    336 
    337 typedef struct int64x1x4_t {
    338   int64x1_t val[4];
    339 } int64x1x4_t;
    340 
    341 typedef struct int64x2x4_t {
    342   int64x2_t val[4];
    343 } int64x2x4_t;
    344 
    345 typedef struct uint8x8x4_t {
    346   uint8x8_t val[4];
    347 } uint8x8x4_t;
    348 
    349 typedef struct uint8x16x4_t {
    350   uint8x16_t val[4];
    351 } uint8x16x4_t;
    352 
    353 typedef struct uint16x4x4_t {
    354   uint16x4_t val[4];
    355 } uint16x4x4_t;
    356 
    357 typedef struct uint16x8x4_t {
    358   uint16x8_t val[4];
    359 } uint16x8x4_t;
    360 
    361 typedef struct uint32x2x4_t {
    362   uint32x2_t val[4];
    363 } uint32x2x4_t;
    364 
    365 typedef struct uint32x4x4_t {
    366   uint32x4_t val[4];
    367 } uint32x4x4_t;
    368 
    369 typedef struct uint64x1x4_t {
    370   uint64x1_t val[4];
    371 } uint64x1x4_t;
    372 
    373 typedef struct uint64x2x4_t {
    374   uint64x2_t val[4];
    375 } uint64x2x4_t;
    376 
    377 typedef struct float16x4x4_t {
    378   float16x4_t val[4];
    379 } float16x4x4_t;
    380 
    381 typedef struct float16x8x4_t {
    382   float16x8_t val[4];
    383 } float16x8x4_t;
    384 
    385 typedef struct float32x2x4_t {
    386   float32x2_t val[4];
    387 } float32x2x4_t;
    388 
    389 typedef struct float32x4x4_t {
    390   float32x4_t val[4];
    391 } float32x4x4_t;
    392 
    393 #ifdef __aarch64__
    394 typedef struct float64x1x4_t {
    395   float64x1_t val[4];
    396 } float64x1x4_t;
    397 
    398 typedef struct float64x2x4_t {
    399   float64x2_t val[4];
    400 } float64x2x4_t;
    401 
    402 #endif
    403 typedef struct poly8x8x4_t {
    404   poly8x8_t val[4];
    405 } poly8x8x4_t;
    406 
    407 typedef struct poly8x16x4_t {
    408   poly8x16_t val[4];
    409 } poly8x16x4_t;
    410 
    411 typedef struct poly16x4x4_t {
    412   poly16x4_t val[4];
    413 } poly16x4x4_t;
    414 
    415 typedef struct poly16x8x4_t {
    416   poly16x8_t val[4];
    417 } poly16x8x4_t;
    418 
    419 #ifdef __aarch64__
    420 typedef struct poly64x1x4_t {
    421   poly64x1_t val[4];
    422 } poly64x1x4_t;
    423 
    424 typedef struct poly64x2x4_t {
    425   poly64x2_t val[4];
    426 } poly64x2x4_t;
    427 
    428 #endif
    429 
    430 #define __ai static inline __attribute__((__always_inline__, __nodebug__))
    431 
    432 #ifdef __LITTLE_ENDIAN__
    433 __ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
    434   uint8x16_t __ret;
    435   __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
    436   return __ret;
    437 }
    438 #else
    439 __ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
    440   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    441   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    442   uint8x16_t __ret;
    443   __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
    444   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    445   return __ret;
    446 }
    447 __ai uint8x16_t __noswap_vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
    448   uint8x16_t __ret;
    449   __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
    450   return __ret;
    451 }
    452 #endif
    453 
    454 #ifdef __LITTLE_ENDIAN__
    455 __ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
    456   uint32x4_t __ret;
    457   __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
    458   return __ret;
    459 }
    460 #else
    461 __ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
    462   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    463   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
    464   uint32x4_t __ret;
    465   __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
    466   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    467   return __ret;
    468 }
    469 __ai uint32x4_t __noswap_vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
    470   uint32x4_t __ret;
    471   __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
    472   return __ret;
    473 }
    474 #endif
    475 
    476 #ifdef __LITTLE_ENDIAN__
    477 __ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
    478   uint16x8_t __ret;
    479   __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
    480   return __ret;
    481 }
    482 #else
    483 __ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
    484   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
    485   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
    486   uint16x8_t __ret;
    487   __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
    488   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
    489   return __ret;
    490 }
    491 __ai uint16x8_t __noswap_vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
    492   uint16x8_t __ret;
    493   __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
    494   return __ret;
    495 }
    496 #endif
    497 
    498 #ifdef __LITTLE_ENDIAN__
    499 __ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
    500   int8x16_t __ret;
    501   __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
    502   return __ret;
    503 }
    504 #else
    505 __ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
    506   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    507   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    508   int8x16_t __ret;
    509   __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
    510   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    511   return __ret;
    512 }
    513 __ai int8x16_t __noswap_vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
    514   int8x16_t __ret;
    515   __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
    516   return __ret;
    517 }
    518 #endif
    519 
    520 #ifdef __LITTLE_ENDIAN__
    521 __ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
    522   float32x4_t __ret;
    523   __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
    524   return __ret;
    525 }
    526 #else
    527 __ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
    528   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    529   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
    530   float32x4_t __ret;
    531   __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
    532   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    533   return __ret;
    534 }
    535 #endif
    536 
    537 #ifdef __LITTLE_ENDIAN__
    538 __ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
    539   int32x4_t __ret;
    540   __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
    541   return __ret;
    542 }
    543 #else
    544 __ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
    545   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    546   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
    547   int32x4_t __ret;
    548   __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
    549   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    550   return __ret;
    551 }
    552 __ai int32x4_t __noswap_vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
    553   int32x4_t __ret;
    554   __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
    555   return __ret;
    556 }
    557 #endif
    558 
    559 #ifdef __LITTLE_ENDIAN__
    560 __ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
    561   int16x8_t __ret;
    562   __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
    563   return __ret;
    564 }
    565 #else
    566 __ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
    567   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
    568   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
    569   int16x8_t __ret;
    570   __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
    571   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
    572   return __ret;
    573 }
    574 __ai int16x8_t __noswap_vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
    575   int16x8_t __ret;
    576   __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
    577   return __ret;
    578 }
    579 #endif
    580 
    581 #ifdef __LITTLE_ENDIAN__
    582 __ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
    583   uint8x8_t __ret;
    584   __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
    585   return __ret;
    586 }
    587 #else
    588 __ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
    589   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
    590   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
    591   uint8x8_t __ret;
    592   __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
    593   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
    594   return __ret;
    595 }
    596 __ai uint8x8_t __noswap_vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
    597   uint8x8_t __ret;
    598   __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
    599   return __ret;
    600 }
    601 #endif
    602 
    603 #ifdef __LITTLE_ENDIAN__
    604 __ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
    605   uint32x2_t __ret;
    606   __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
    607   return __ret;
    608 }
    609 #else
    610 __ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
    611   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
    612   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
    613   uint32x2_t __ret;
    614   __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
    615   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
    616   return __ret;
    617 }
    618 __ai uint32x2_t __noswap_vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
    619   uint32x2_t __ret;
    620   __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
    621   return __ret;
    622 }
    623 #endif
    624 
    625 #ifdef __LITTLE_ENDIAN__
    626 __ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
    627   uint16x4_t __ret;
    628   __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
    629   return __ret;
    630 }
    631 #else
    632 __ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
    633   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    634   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
    635   uint16x4_t __ret;
    636   __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
    637   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    638   return __ret;
    639 }
    640 __ai uint16x4_t __noswap_vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
    641   uint16x4_t __ret;
    642   __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
    643   return __ret;
    644 }
    645 #endif
    646 
    647 #ifdef __LITTLE_ENDIAN__
    648 __ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
    649   int8x8_t __ret;
    650   __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
    651   return __ret;
    652 }
    653 #else
    654 __ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
    655   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
    656   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
    657   int8x8_t __ret;
    658   __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
    659   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
    660   return __ret;
    661 }
    662 __ai int8x8_t __noswap_vabd_s8(int8x8_t __p0, int8x8_t __p1) {
    663   int8x8_t __ret;
    664   __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
    665   return __ret;
    666 }
    667 #endif
    668 
    669 #ifdef __LITTLE_ENDIAN__
    670 __ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
    671   float32x2_t __ret;
    672   __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
    673   return __ret;
    674 }
    675 #else
    676 __ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
    677   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
    678   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
    679   float32x2_t __ret;
    680   __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
    681   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
    682   return __ret;
    683 }
    684 #endif
    685 
    686 #ifdef __LITTLE_ENDIAN__
    687 __ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
    688   int32x2_t __ret;
    689   __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
    690   return __ret;
    691 }
    692 #else
    693 __ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
    694   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
    695   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
    696   int32x2_t __ret;
    697   __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
    698   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
    699   return __ret;
    700 }
    701 __ai int32x2_t __noswap_vabd_s32(int32x2_t __p0, int32x2_t __p1) {
    702   int32x2_t __ret;
    703   __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
    704   return __ret;
    705 }
    706 #endif
    707 
    708 #ifdef __LITTLE_ENDIAN__
    709 __ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
    710   int16x4_t __ret;
    711   __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
    712   return __ret;
    713 }
    714 #else
    715 __ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
    716   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    717   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
    718   int16x4_t __ret;
    719   __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
    720   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    721   return __ret;
    722 }
    723 __ai int16x4_t __noswap_vabd_s16(int16x4_t __p0, int16x4_t __p1) {
    724   int16x4_t __ret;
    725   __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
    726   return __ret;
    727 }
    728 #endif
    729 
    730 #ifdef __LITTLE_ENDIAN__
    731 __ai int8x16_t vabsq_s8(int8x16_t __p0) {
    732   int8x16_t __ret;
    733   __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 32);
    734   return __ret;
    735 }
    736 #else
    737 __ai int8x16_t vabsq_s8(int8x16_t __p0) {
    738   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    739   int8x16_t __ret;
    740   __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 32);
    741   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    742   return __ret;
    743 }
    744 #endif
    745 
    746 #ifdef __LITTLE_ENDIAN__
    747 __ai float32x4_t vabsq_f32(float32x4_t __p0) {
    748   float32x4_t __ret;
    749   __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 41);
    750   return __ret;
    751 }
    752 #else
    753 __ai float32x4_t vabsq_f32(float32x4_t __p0) {
    754   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    755   float32x4_t __ret;
    756   __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 41);
    757   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    758   return __ret;
    759 }
    760 #endif
    761 
    762 #ifdef __LITTLE_ENDIAN__
    763 __ai int32x4_t vabsq_s32(int32x4_t __p0) {
    764   int32x4_t __ret;
    765   __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 34);
    766   return __ret;
    767 }
    768 #else
    769 __ai int32x4_t vabsq_s32(int32x4_t __p0) {
    770   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    771   int32x4_t __ret;
    772   __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 34);
    773   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    774   return __ret;
    775 }
    776 #endif
    777 
    778 #ifdef __LITTLE_ENDIAN__
    779 __ai int16x8_t vabsq_s16(int16x8_t __p0) {
    780   int16x8_t __ret;
    781   __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 33);
    782   return __ret;
    783 }
    784 #else
    785 __ai int16x8_t vabsq_s16(int16x8_t __p0) {
    786   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
    787   int16x8_t __ret;
    788   __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 33);
    789   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
    790   return __ret;
    791 }
    792 #endif
    793 
    794 #ifdef __LITTLE_ENDIAN__
    795 __ai int8x8_t vabs_s8(int8x8_t __p0) {
    796   int8x8_t __ret;
    797   __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__p0, 0);
    798   return __ret;
    799 }
    800 #else
    801 __ai int8x8_t vabs_s8(int8x8_t __p0) {
    802   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
    803   int8x8_t __ret;
    804   __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 0);
    805   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
    806   return __ret;
    807 }
    808 #endif
    809 
    810 #ifdef __LITTLE_ENDIAN__
    811 __ai float32x2_t vabs_f32(float32x2_t __p0) {
    812   float32x2_t __ret;
    813   __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 9);
    814   return __ret;
    815 }
    816 #else
    817 __ai float32x2_t vabs_f32(float32x2_t __p0) {
    818   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
    819   float32x2_t __ret;
    820   __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 9);
    821   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
    822   return __ret;
    823 }
    824 #endif
    825 
    826 #ifdef __LITTLE_ENDIAN__
    827 __ai int32x2_t vabs_s32(int32x2_t __p0) {
    828   int32x2_t __ret;
    829   __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 2);
    830   return __ret;
    831 }
    832 #else
    833 __ai int32x2_t vabs_s32(int32x2_t __p0) {
    834   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
    835   int32x2_t __ret;
    836   __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 2);
    837   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
    838   return __ret;
    839 }
    840 #endif
    841 
    842 #ifdef __LITTLE_ENDIAN__
    843 __ai int16x4_t vabs_s16(int16x4_t __p0) {
    844   int16x4_t __ret;
    845   __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__p0, 1);
    846   return __ret;
    847 }
    848 #else
    849 __ai int16x4_t vabs_s16(int16x4_t __p0) {
    850   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    851   int16x4_t __ret;
    852   __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 1);
    853   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    854   return __ret;
    855 }
    856 #endif
    857 
    858 #ifdef __LITTLE_ENDIAN__
    859 __ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
    860   uint8x16_t __ret;
    861   __ret = __p0 + __p1;
    862   return __ret;
    863 }
    864 #else
    865 __ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
    866   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    867   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    868   uint8x16_t __ret;
    869   __ret = __rev0 + __rev1;
    870   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    871   return __ret;
    872 }
    873 #endif
    874 
    875 #ifdef __LITTLE_ENDIAN__
    876 __ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
    877   uint32x4_t __ret;
    878   __ret = __p0 + __p1;
    879   return __ret;
    880 }
    881 #else
    882 __ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
    883   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    884   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
    885   uint32x4_t __ret;
    886   __ret = __rev0 + __rev1;
    887   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    888   return __ret;
    889 }
    890 #endif
    891 
    892 #ifdef __LITTLE_ENDIAN__
    893 __ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
    894   uint64x2_t __ret;
    895   __ret = __p0 + __p1;
    896   return __ret;
    897 }
    898 #else
    899 __ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
    900   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
    901   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
    902   uint64x2_t __ret;
    903   __ret = __rev0 + __rev1;
    904   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
    905   return __ret;
    906 }
    907 #endif
    908 
    909 #ifdef __LITTLE_ENDIAN__
    910 __ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
    911   uint16x8_t __ret;
    912   __ret = __p0 + __p1;
    913   return __ret;
    914 }
    915 #else
    916 __ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
    917   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
    918   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
    919   uint16x8_t __ret;
    920   __ret = __rev0 + __rev1;
    921   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
    922   return __ret;
    923 }
    924 #endif
    925 
    926 #ifdef __LITTLE_ENDIAN__
    927 __ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
    928   int8x16_t __ret;
    929   __ret = __p0 + __p1;
    930   return __ret;
    931 }
    932 #else
    933 __ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
    934   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    935   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    936   int8x16_t __ret;
    937   __ret = __rev0 + __rev1;
    938   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
    939   return __ret;
    940 }
    941 #endif
    942 
    943 #ifdef __LITTLE_ENDIAN__
    944 __ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
    945   float32x4_t __ret;
    946   __ret = __p0 + __p1;
    947   return __ret;
    948 }
    949 #else
    950 __ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
    951   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    952   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
    953   float32x4_t __ret;
    954   __ret = __rev0 + __rev1;
    955   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    956   return __ret;
    957 }
    958 #endif
    959 
    960 #ifdef __LITTLE_ENDIAN__
    961 __ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
    962   int32x4_t __ret;
    963   __ret = __p0 + __p1;
    964   return __ret;
    965 }
    966 #else
    967 __ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
    968   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
    969   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
    970   int32x4_t __ret;
    971   __ret = __rev0 + __rev1;
    972   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
    973   return __ret;
    974 }
    975 #endif
    976 
    977 #ifdef __LITTLE_ENDIAN__
    978 __ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
    979   int64x2_t __ret;
    980   __ret = __p0 + __p1;
    981   return __ret;
    982 }
    983 #else
    984 __ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
    985   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
    986   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
    987   int64x2_t __ret;
    988   __ret = __rev0 + __rev1;
    989   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
    990   return __ret;
    991 }
    992 #endif
    993 
    994 #ifdef __LITTLE_ENDIAN__
    995 __ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
    996   int16x8_t __ret;
    997   __ret = __p0 + __p1;
    998   return __ret;
    999 }
   1000 #else
   1001 __ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   1002   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1003   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1004   int16x8_t __ret;
   1005   __ret = __rev0 + __rev1;
   1006   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1007   return __ret;
   1008 }
   1009 #endif
   1010 
   1011 #ifdef __LITTLE_ENDIAN__
   1012 __ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   1013   uint8x8_t __ret;
   1014   __ret = __p0 + __p1;
   1015   return __ret;
   1016 }
   1017 #else
   1018 __ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   1019   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1020   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1021   uint8x8_t __ret;
   1022   __ret = __rev0 + __rev1;
   1023   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1024   return __ret;
   1025 }
   1026 #endif
   1027 
   1028 #ifdef __LITTLE_ENDIAN__
   1029 __ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   1030   uint32x2_t __ret;
   1031   __ret = __p0 + __p1;
   1032   return __ret;
   1033 }
   1034 #else
   1035 __ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   1036   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1037   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1038   uint32x2_t __ret;
   1039   __ret = __rev0 + __rev1;
   1040   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1041   return __ret;
   1042 }
   1043 #endif
   1044 
   1045 #ifdef __LITTLE_ENDIAN__
   1046 __ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
   1047   uint64x1_t __ret;
   1048   __ret = __p0 + __p1;
   1049   return __ret;
   1050 }
   1051 #else
   1052 __ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
   1053   uint64x1_t __ret;
   1054   __ret = __p0 + __p1;
   1055   return __ret;
   1056 }
   1057 #endif
   1058 
   1059 #ifdef __LITTLE_ENDIAN__
   1060 __ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   1061   uint16x4_t __ret;
   1062   __ret = __p0 + __p1;
   1063   return __ret;
   1064 }
   1065 #else
   1066 __ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   1067   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1068   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1069   uint16x4_t __ret;
   1070   __ret = __rev0 + __rev1;
   1071   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1072   return __ret;
   1073 }
   1074 #endif
   1075 
   1076 #ifdef __LITTLE_ENDIAN__
   1077 __ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
   1078   int8x8_t __ret;
   1079   __ret = __p0 + __p1;
   1080   return __ret;
   1081 }
   1082 #else
   1083 __ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
   1084   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1085   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1086   int8x8_t __ret;
   1087   __ret = __rev0 + __rev1;
   1088   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1089   return __ret;
   1090 }
   1091 #endif
   1092 
   1093 #ifdef __LITTLE_ENDIAN__
   1094 __ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
   1095   float32x2_t __ret;
   1096   __ret = __p0 + __p1;
   1097   return __ret;
   1098 }
   1099 #else
   1100 __ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
   1101   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1102   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1103   float32x2_t __ret;
   1104   __ret = __rev0 + __rev1;
   1105   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1106   return __ret;
   1107 }
   1108 #endif
   1109 
   1110 #ifdef __LITTLE_ENDIAN__
   1111 __ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
   1112   int32x2_t __ret;
   1113   __ret = __p0 + __p1;
   1114   return __ret;
   1115 }
   1116 #else
   1117 __ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
   1118   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1119   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1120   int32x2_t __ret;
   1121   __ret = __rev0 + __rev1;
   1122   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1123   return __ret;
   1124 }
   1125 #endif
   1126 
   1127 #ifdef __LITTLE_ENDIAN__
   1128 __ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
   1129   int64x1_t __ret;
   1130   __ret = __p0 + __p1;
   1131   return __ret;
   1132 }
   1133 #else
   1134 __ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
   1135   int64x1_t __ret;
   1136   __ret = __p0 + __p1;
   1137   return __ret;
   1138 }
   1139 #endif
   1140 
   1141 #ifdef __LITTLE_ENDIAN__
   1142 __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
   1143   int16x4_t __ret;
   1144   __ret = __p0 + __p1;
   1145   return __ret;
   1146 }
   1147 #else
   1148 __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
   1149   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1150   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1151   int16x4_t __ret;
   1152   __ret = __rev0 + __rev1;
   1153   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1154   return __ret;
   1155 }
   1156 #endif
   1157 
   1158 #ifdef __LITTLE_ENDIAN__
   1159 __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   1160   uint16x4_t __ret;
   1161   __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
   1162   return __ret;
   1163 }
   1164 #else
   1165 __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   1166   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1167   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1168   uint16x4_t __ret;
   1169   __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
   1170   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1171   return __ret;
   1172 }
   1173 __ai uint16x4_t __noswap_vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   1174   uint16x4_t __ret;
   1175   __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
   1176   return __ret;
   1177 }
   1178 #endif
   1179 
   1180 #ifdef __LITTLE_ENDIAN__
   1181 __ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   1182   uint32x2_t __ret;
   1183   __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
   1184   return __ret;
   1185 }
   1186 #else
   1187 __ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   1188   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1189   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1190   uint32x2_t __ret;
   1191   __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
   1192   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1193   return __ret;
   1194 }
   1195 __ai uint32x2_t __noswap_vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   1196   uint32x2_t __ret;
   1197   __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
   1198   return __ret;
   1199 }
   1200 #endif
   1201 
   1202 #ifdef __LITTLE_ENDIAN__
   1203 __ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   1204   uint8x8_t __ret;
   1205   __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
   1206   return __ret;
   1207 }
   1208 #else
   1209 __ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   1210   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1211   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1212   uint8x8_t __ret;
   1213   __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
   1214   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1215   return __ret;
   1216 }
   1217 __ai uint8x8_t __noswap_vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   1218   uint8x8_t __ret;
   1219   __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
   1220   return __ret;
   1221 }
   1222 #endif
   1223 
   1224 #ifdef __LITTLE_ENDIAN__
   1225 __ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
   1226   int16x4_t __ret;
   1227   __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
   1228   return __ret;
   1229 }
   1230 #else
   1231 __ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
   1232   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1233   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1234   int16x4_t __ret;
   1235   __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
   1236   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1237   return __ret;
   1238 }
   1239 __ai int16x4_t __noswap_vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
   1240   int16x4_t __ret;
   1241   __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
   1242   return __ret;
   1243 }
   1244 #endif
   1245 
   1246 #ifdef __LITTLE_ENDIAN__
   1247 __ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
   1248   int32x2_t __ret;
   1249   __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
   1250   return __ret;
   1251 }
   1252 #else
   1253 __ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
   1254   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1255   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1256   int32x2_t __ret;
   1257   __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
   1258   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1259   return __ret;
   1260 }
   1261 __ai int32x2_t __noswap_vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
   1262   int32x2_t __ret;
   1263   __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
   1264   return __ret;
   1265 }
   1266 #endif
   1267 
   1268 #ifdef __LITTLE_ENDIAN__
   1269 __ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
   1270   int8x8_t __ret;
   1271   __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
   1272   return __ret;
   1273 }
   1274 #else
   1275 __ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
   1276   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1277   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1278   int8x8_t __ret;
   1279   __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
   1280   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1281   return __ret;
   1282 }
   1283 __ai int8x8_t __noswap_vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
   1284   int8x8_t __ret;
   1285   __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
   1286   return __ret;
   1287 }
   1288 #endif
   1289 
   1290 #ifdef __LITTLE_ENDIAN__
   1291 __ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   1292   uint8x16_t __ret;
   1293   __ret = __p0 & __p1;
   1294   return __ret;
   1295 }
   1296 #else
   1297 __ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   1298   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1299   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1300   uint8x16_t __ret;
   1301   __ret = __rev0 & __rev1;
   1302   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1303   return __ret;
   1304 }
   1305 #endif
   1306 
   1307 #ifdef __LITTLE_ENDIAN__
   1308 __ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   1309   uint32x4_t __ret;
   1310   __ret = __p0 & __p1;
   1311   return __ret;
   1312 }
   1313 #else
   1314 __ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   1315   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1316   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1317   uint32x4_t __ret;
   1318   __ret = __rev0 & __rev1;
   1319   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1320   return __ret;
   1321 }
   1322 #endif
   1323 
   1324 #ifdef __LITTLE_ENDIAN__
   1325 __ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   1326   uint64x2_t __ret;
   1327   __ret = __p0 & __p1;
   1328   return __ret;
   1329 }
   1330 #else
   1331 __ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   1332   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1333   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1334   uint64x2_t __ret;
   1335   __ret = __rev0 & __rev1;
   1336   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1337   return __ret;
   1338 }
   1339 #endif
   1340 
   1341 #ifdef __LITTLE_ENDIAN__
   1342 __ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   1343   uint16x8_t __ret;
   1344   __ret = __p0 & __p1;
   1345   return __ret;
   1346 }
   1347 #else
   1348 __ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   1349   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1350   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1351   uint16x8_t __ret;
   1352   __ret = __rev0 & __rev1;
   1353   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1354   return __ret;
   1355 }
   1356 #endif
   1357 
   1358 #ifdef __LITTLE_ENDIAN__
   1359 __ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
   1360   int8x16_t __ret;
   1361   __ret = __p0 & __p1;
   1362   return __ret;
   1363 }
   1364 #else
   1365 __ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
   1366   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1367   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1368   int8x16_t __ret;
   1369   __ret = __rev0 & __rev1;
   1370   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1371   return __ret;
   1372 }
   1373 #endif
   1374 
   1375 #ifdef __LITTLE_ENDIAN__
   1376 __ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
   1377   int32x4_t __ret;
   1378   __ret = __p0 & __p1;
   1379   return __ret;
   1380 }
   1381 #else
   1382 __ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
   1383   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1384   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1385   int32x4_t __ret;
   1386   __ret = __rev0 & __rev1;
   1387   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1388   return __ret;
   1389 }
   1390 #endif
   1391 
   1392 #ifdef __LITTLE_ENDIAN__
   1393 __ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
   1394   int64x2_t __ret;
   1395   __ret = __p0 & __p1;
   1396   return __ret;
   1397 }
   1398 #else
   1399 __ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
   1400   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1401   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1402   int64x2_t __ret;
   1403   __ret = __rev0 & __rev1;
   1404   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1405   return __ret;
   1406 }
   1407 #endif
   1408 
   1409 #ifdef __LITTLE_ENDIAN__
   1410 __ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
   1411   int16x8_t __ret;
   1412   __ret = __p0 & __p1;
   1413   return __ret;
   1414 }
   1415 #else
   1416 __ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
   1417   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1418   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1419   int16x8_t __ret;
   1420   __ret = __rev0 & __rev1;
   1421   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1422   return __ret;
   1423 }
   1424 #endif
   1425 
   1426 #ifdef __LITTLE_ENDIAN__
   1427 __ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
   1428   uint8x8_t __ret;
   1429   __ret = __p0 & __p1;
   1430   return __ret;
   1431 }
   1432 #else
   1433 __ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
   1434   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1435   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1436   uint8x8_t __ret;
   1437   __ret = __rev0 & __rev1;
   1438   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1439   return __ret;
   1440 }
   1441 #endif
   1442 
   1443 #ifdef __LITTLE_ENDIAN__
   1444 __ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
   1445   uint32x2_t __ret;
   1446   __ret = __p0 & __p1;
   1447   return __ret;
   1448 }
   1449 #else
   1450 __ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
   1451   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1452   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1453   uint32x2_t __ret;
   1454   __ret = __rev0 & __rev1;
   1455   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1456   return __ret;
   1457 }
   1458 #endif
   1459 
   1460 #ifdef __LITTLE_ENDIAN__
   1461 __ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
   1462   uint64x1_t __ret;
   1463   __ret = __p0 & __p1;
   1464   return __ret;
   1465 }
   1466 #else
   1467 __ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
   1468   uint64x1_t __ret;
   1469   __ret = __p0 & __p1;
   1470   return __ret;
   1471 }
   1472 #endif
   1473 
   1474 #ifdef __LITTLE_ENDIAN__
   1475 __ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
   1476   uint16x4_t __ret;
   1477   __ret = __p0 & __p1;
   1478   return __ret;
   1479 }
   1480 #else
   1481 __ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
   1482   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1483   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1484   uint16x4_t __ret;
   1485   __ret = __rev0 & __rev1;
   1486   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1487   return __ret;
   1488 }
   1489 #endif
   1490 
   1491 #ifdef __LITTLE_ENDIAN__
   1492 __ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
   1493   int8x8_t __ret;
   1494   __ret = __p0 & __p1;
   1495   return __ret;
   1496 }
   1497 #else
   1498 __ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
   1499   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1500   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1501   int8x8_t __ret;
   1502   __ret = __rev0 & __rev1;
   1503   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1504   return __ret;
   1505 }
   1506 #endif
   1507 
   1508 #ifdef __LITTLE_ENDIAN__
   1509 __ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
   1510   int32x2_t __ret;
   1511   __ret = __p0 & __p1;
   1512   return __ret;
   1513 }
   1514 #else
   1515 __ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
   1516   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1517   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1518   int32x2_t __ret;
   1519   __ret = __rev0 & __rev1;
   1520   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1521   return __ret;
   1522 }
   1523 #endif
   1524 
   1525 #ifdef __LITTLE_ENDIAN__
   1526 __ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
   1527   int64x1_t __ret;
   1528   __ret = __p0 & __p1;
   1529   return __ret;
   1530 }
   1531 #else
   1532 __ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
   1533   int64x1_t __ret;
   1534   __ret = __p0 & __p1;
   1535   return __ret;
   1536 }
   1537 #endif
   1538 
   1539 #ifdef __LITTLE_ENDIAN__
   1540 __ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
   1541   int16x4_t __ret;
   1542   __ret = __p0 & __p1;
   1543   return __ret;
   1544 }
   1545 #else
   1546 __ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
   1547   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1548   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1549   int16x4_t __ret;
   1550   __ret = __rev0 & __rev1;
   1551   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1552   return __ret;
   1553 }
   1554 #endif
   1555 
   1556 #ifdef __LITTLE_ENDIAN__
   1557 __ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   1558   uint8x16_t __ret;
   1559   __ret = __p0 & ~__p1;
   1560   return __ret;
   1561 }
   1562 #else
   1563 __ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   1564   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1565   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1566   uint8x16_t __ret;
   1567   __ret = __rev0 & ~__rev1;
   1568   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1569   return __ret;
   1570 }
   1571 #endif
   1572 
   1573 #ifdef __LITTLE_ENDIAN__
   1574 __ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   1575   uint32x4_t __ret;
   1576   __ret = __p0 & ~__p1;
   1577   return __ret;
   1578 }
   1579 #else
   1580 __ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   1581   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1582   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1583   uint32x4_t __ret;
   1584   __ret = __rev0 & ~__rev1;
   1585   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1586   return __ret;
   1587 }
   1588 #endif
   1589 
   1590 #ifdef __LITTLE_ENDIAN__
   1591 __ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   1592   uint64x2_t __ret;
   1593   __ret = __p0 & ~__p1;
   1594   return __ret;
   1595 }
   1596 #else
   1597 __ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   1598   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1599   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1600   uint64x2_t __ret;
   1601   __ret = __rev0 & ~__rev1;
   1602   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1603   return __ret;
   1604 }
   1605 #endif
   1606 
   1607 #ifdef __LITTLE_ENDIAN__
   1608 __ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   1609   uint16x8_t __ret;
   1610   __ret = __p0 & ~__p1;
   1611   return __ret;
   1612 }
   1613 #else
   1614 __ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   1615   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1616   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1617   uint16x8_t __ret;
   1618   __ret = __rev0 & ~__rev1;
   1619   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1620   return __ret;
   1621 }
   1622 #endif
   1623 
   1624 #ifdef __LITTLE_ENDIAN__
   1625 __ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
   1626   int8x16_t __ret;
   1627   __ret = __p0 & ~__p1;
   1628   return __ret;
   1629 }
   1630 #else
   1631 __ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
   1632   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1633   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1634   int8x16_t __ret;
   1635   __ret = __rev0 & ~__rev1;
   1636   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1637   return __ret;
   1638 }
   1639 #endif
   1640 
   1641 #ifdef __LITTLE_ENDIAN__
   1642 __ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
   1643   int32x4_t __ret;
   1644   __ret = __p0 & ~__p1;
   1645   return __ret;
   1646 }
   1647 #else
   1648 __ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
   1649   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1650   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1651   int32x4_t __ret;
   1652   __ret = __rev0 & ~__rev1;
   1653   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1654   return __ret;
   1655 }
   1656 #endif
   1657 
   1658 #ifdef __LITTLE_ENDIAN__
   1659 __ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
   1660   int64x2_t __ret;
   1661   __ret = __p0 & ~__p1;
   1662   return __ret;
   1663 }
   1664 #else
   1665 __ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
   1666   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1667   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1668   int64x2_t __ret;
   1669   __ret = __rev0 & ~__rev1;
   1670   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1671   return __ret;
   1672 }
   1673 #endif
   1674 
   1675 #ifdef __LITTLE_ENDIAN__
   1676 __ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
   1677   int16x8_t __ret;
   1678   __ret = __p0 & ~__p1;
   1679   return __ret;
   1680 }
   1681 #else
   1682 __ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
   1683   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1684   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1685   int16x8_t __ret;
   1686   __ret = __rev0 & ~__rev1;
   1687   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1688   return __ret;
   1689 }
   1690 #endif
   1691 
   1692 #ifdef __LITTLE_ENDIAN__
   1693 __ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
   1694   uint8x8_t __ret;
   1695   __ret = __p0 & ~__p1;
   1696   return __ret;
   1697 }
   1698 #else
   1699 __ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
   1700   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1701   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1702   uint8x8_t __ret;
   1703   __ret = __rev0 & ~__rev1;
   1704   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1705   return __ret;
   1706 }
   1707 #endif
   1708 
   1709 #ifdef __LITTLE_ENDIAN__
   1710 __ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
   1711   uint32x2_t __ret;
   1712   __ret = __p0 & ~__p1;
   1713   return __ret;
   1714 }
   1715 #else
   1716 __ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
   1717   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1718   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1719   uint32x2_t __ret;
   1720   __ret = __rev0 & ~__rev1;
   1721   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1722   return __ret;
   1723 }
   1724 #endif
   1725 
   1726 #ifdef __LITTLE_ENDIAN__
   1727 __ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
   1728   uint64x1_t __ret;
   1729   __ret = __p0 & ~__p1;
   1730   return __ret;
   1731 }
   1732 #else
   1733 __ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
   1734   uint64x1_t __ret;
   1735   __ret = __p0 & ~__p1;
   1736   return __ret;
   1737 }
   1738 #endif
   1739 
   1740 #ifdef __LITTLE_ENDIAN__
   1741 __ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
   1742   uint16x4_t __ret;
   1743   __ret = __p0 & ~__p1;
   1744   return __ret;
   1745 }
   1746 #else
   1747 __ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
   1748   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1749   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1750   uint16x4_t __ret;
   1751   __ret = __rev0 & ~__rev1;
   1752   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1753   return __ret;
   1754 }
   1755 #endif
   1756 
   1757 #ifdef __LITTLE_ENDIAN__
   1758 __ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
   1759   int8x8_t __ret;
   1760   __ret = __p0 & ~__p1;
   1761   return __ret;
   1762 }
   1763 #else
   1764 __ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
   1765   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1766   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1767   int8x8_t __ret;
   1768   __ret = __rev0 & ~__rev1;
   1769   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1770   return __ret;
   1771 }
   1772 #endif
   1773 
   1774 #ifdef __LITTLE_ENDIAN__
   1775 __ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
   1776   int32x2_t __ret;
   1777   __ret = __p0 & ~__p1;
   1778   return __ret;
   1779 }
   1780 #else
   1781 __ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
   1782   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1783   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1784   int32x2_t __ret;
   1785   __ret = __rev0 & ~__rev1;
   1786   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1787   return __ret;
   1788 }
   1789 #endif
   1790 
   1791 #ifdef __LITTLE_ENDIAN__
   1792 __ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
   1793   int64x1_t __ret;
   1794   __ret = __p0 & ~__p1;
   1795   return __ret;
   1796 }
   1797 #else
   1798 __ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
   1799   int64x1_t __ret;
   1800   __ret = __p0 & ~__p1;
   1801   return __ret;
   1802 }
   1803 #endif
   1804 
   1805 #ifdef __LITTLE_ENDIAN__
   1806 __ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
   1807   int16x4_t __ret;
   1808   __ret = __p0 & ~__p1;
   1809   return __ret;
   1810 }
   1811 #else
   1812 __ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
   1813   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1814   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1815   int16x4_t __ret;
   1816   __ret = __rev0 & ~__rev1;
   1817   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1818   return __ret;
   1819 }
   1820 #endif
   1821 
   1822 #ifdef __LITTLE_ENDIAN__
   1823 __ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
   1824   poly8x8_t __ret;
   1825   __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
   1826   return __ret;
   1827 }
   1828 #else
   1829 __ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
   1830   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1831   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1832   poly8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   1833   poly8x8_t __ret;
   1834   __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
   1835   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1836   return __ret;
   1837 }
   1838 #endif
   1839 
   1840 #ifdef __LITTLE_ENDIAN__
   1841 __ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
   1842   poly16x4_t __ret;
   1843   __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 5);
   1844   return __ret;
   1845 }
   1846 #else
   1847 __ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
   1848   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1849   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1850   poly16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   1851   poly16x4_t __ret;
   1852   __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 5);
   1853   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1854   return __ret;
   1855 }
   1856 #endif
   1857 
   1858 #ifdef __LITTLE_ENDIAN__
   1859 __ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
   1860   poly8x16_t __ret;
   1861   __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
   1862   return __ret;
   1863 }
   1864 #else
   1865 __ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
   1866   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1867   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1868   poly8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1869   poly8x16_t __ret;
   1870   __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
   1871   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1872   return __ret;
   1873 }
   1874 #endif
   1875 
   1876 #ifdef __LITTLE_ENDIAN__
   1877 __ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
   1878   poly16x8_t __ret;
   1879   __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 37);
   1880   return __ret;
   1881 }
   1882 #else
   1883 __ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
   1884   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1885   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1886   poly16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   1887   poly16x8_t __ret;
   1888   __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 37);
   1889   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1890   return __ret;
   1891 }
   1892 #endif
   1893 
   1894 #ifdef __LITTLE_ENDIAN__
   1895 __ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   1896   uint8x16_t __ret;
   1897   __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
   1898   return __ret;
   1899 }
   1900 #else
   1901 __ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   1902   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1903   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1904   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1905   uint8x16_t __ret;
   1906   __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
   1907   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1908   return __ret;
   1909 }
   1910 #endif
   1911 
   1912 #ifdef __LITTLE_ENDIAN__
   1913 __ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   1914   uint32x4_t __ret;
   1915   __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
   1916   return __ret;
   1917 }
   1918 #else
   1919 __ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   1920   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1921   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1922   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   1923   uint32x4_t __ret;
   1924   __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
   1925   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1926   return __ret;
   1927 }
   1928 #endif
   1929 
   1930 #ifdef __LITTLE_ENDIAN__
   1931 __ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   1932   uint64x2_t __ret;
   1933   __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
   1934   return __ret;
   1935 }
   1936 #else
   1937 __ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   1938   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   1939   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   1940   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   1941   uint64x2_t __ret;
   1942   __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
   1943   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   1944   return __ret;
   1945 }
   1946 #endif
   1947 
   1948 #ifdef __LITTLE_ENDIAN__
   1949 __ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   1950   uint16x8_t __ret;
   1951   __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49);
   1952   return __ret;
   1953 }
   1954 #else
   1955 __ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   1956   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   1957   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   1958   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   1959   uint16x8_t __ret;
   1960   __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49);
   1961   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   1962   return __ret;
   1963 }
   1964 #endif
   1965 
   1966 #ifdef __LITTLE_ENDIAN__
   1967 __ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   1968   int8x16_t __ret;
   1969   __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
   1970   return __ret;
   1971 }
   1972 #else
   1973 __ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   1974   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1975   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1976   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1977   int8x16_t __ret;
   1978   __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
   1979   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   1980   return __ret;
   1981 }
   1982 #endif
   1983 
   1984 #ifdef __LITTLE_ENDIAN__
   1985 __ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   1986   float32x4_t __ret;
   1987   __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
   1988   return __ret;
   1989 }
   1990 #else
   1991 __ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   1992   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   1993   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   1994   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   1995   float32x4_t __ret;
   1996   __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
   1997   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   1998   return __ret;
   1999 }
   2000 #endif
   2001 
   2002 #ifdef __LITTLE_ENDIAN__
   2003 __ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   2004   int32x4_t __ret;
   2005   __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
   2006   return __ret;
   2007 }
   2008 #else
   2009 __ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   2010   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2011   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2012   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   2013   int32x4_t __ret;
   2014   __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
   2015   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2016   return __ret;
   2017 }
   2018 #endif
   2019 
   2020 #ifdef __LITTLE_ENDIAN__
   2021 __ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   2022   int64x2_t __ret;
   2023   __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35);
   2024   return __ret;
   2025 }
   2026 #else
   2027 __ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   2028   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2029   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2030   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   2031   int64x2_t __ret;
   2032   __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35);
   2033   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2034   return __ret;
   2035 }
   2036 #endif
   2037 
   2038 #ifdef __LITTLE_ENDIAN__
   2039 __ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   2040   int16x8_t __ret;
   2041   __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
   2042   return __ret;
   2043 }
   2044 #else
   2045 __ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   2046   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2047   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2048   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   2049   int16x8_t __ret;
   2050   __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
   2051   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2052   return __ret;
   2053 }
   2054 #endif
   2055 
   2056 #ifdef __LITTLE_ENDIAN__
   2057 __ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   2058   uint8x8_t __ret;
   2059   __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
   2060   return __ret;
   2061 }
   2062 #else
   2063 __ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   2064   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2065   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2066   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   2067   uint8x8_t __ret;
   2068   __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
   2069   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2070   return __ret;
   2071 }
   2072 #endif
   2073 
   2074 #ifdef __LITTLE_ENDIAN__
   2075 __ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   2076   uint32x2_t __ret;
   2077   __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
   2078   return __ret;
   2079 }
   2080 #else
   2081 __ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   2082   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2083   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2084   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   2085   uint32x2_t __ret;
   2086   __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 18);
   2087   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2088   return __ret;
   2089 }
   2090 #endif
   2091 
   2092 #ifdef __LITTLE_ENDIAN__
   2093 __ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
   2094   uint64x1_t __ret;
   2095   __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
   2096   return __ret;
   2097 }
   2098 #else
   2099 __ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
   2100   uint64x1_t __ret;
   2101   __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
   2102   return __ret;
   2103 }
   2104 #endif
   2105 
   2106 #ifdef __LITTLE_ENDIAN__
   2107 __ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   2108   uint16x4_t __ret;
   2109   __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 17);
   2110   return __ret;
   2111 }
   2112 #else
   2113 __ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   2114   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2115   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2116   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   2117   uint16x4_t __ret;
   2118   __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 17);
   2119   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2120   return __ret;
   2121 }
   2122 #endif
   2123 
   2124 #ifdef __LITTLE_ENDIAN__
   2125 __ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   2126   int8x8_t __ret;
   2127   __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
   2128   return __ret;
   2129 }
   2130 #else
   2131 __ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   2132   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2133   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2134   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   2135   int8x8_t __ret;
   2136   __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
   2137   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2138   return __ret;
   2139 }
   2140 #endif
   2141 
   2142 #ifdef __LITTLE_ENDIAN__
   2143 __ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   2144   float32x2_t __ret;
   2145   __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
   2146   return __ret;
   2147 }
   2148 #else
   2149 __ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   2150   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2151   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2152   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   2153   float32x2_t __ret;
   2154   __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
   2155   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2156   return __ret;
   2157 }
   2158 #endif
   2159 
   2160 #ifdef __LITTLE_ENDIAN__
   2161 __ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   2162   int32x2_t __ret;
   2163   __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
   2164   return __ret;
   2165 }
   2166 #else
   2167 __ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   2168   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2169   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2170   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   2171   int32x2_t __ret;
   2172   __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
   2173   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2174   return __ret;
   2175 }
   2176 #endif
   2177 
   2178 #ifdef __LITTLE_ENDIAN__
   2179 __ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
   2180   int64x1_t __ret;
   2181   __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
   2182   return __ret;
   2183 }
   2184 #else
   2185 __ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
   2186   int64x1_t __ret;
   2187   __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
   2188   return __ret;
   2189 }
   2190 #endif
   2191 
   2192 #ifdef __LITTLE_ENDIAN__
   2193 __ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   2194   int16x4_t __ret;
   2195   __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
   2196   return __ret;
   2197 }
   2198 #else
   2199 __ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   2200   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2201   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2202   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   2203   int16x4_t __ret;
   2204   __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 1);
   2205   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2206   return __ret;
   2207 }
   2208 #endif
   2209 
   2210 #ifdef __LITTLE_ENDIAN__
   2211 __ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
   2212   uint32x4_t __ret;
   2213   __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   2214   return __ret;
   2215 }
   2216 #else
   2217 __ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
   2218   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2219   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2220   uint32x4_t __ret;
   2221   __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   2222   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2223   return __ret;
   2224 }
   2225 #endif
   2226 
   2227 #ifdef __LITTLE_ENDIAN__
   2228 __ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
   2229   uint32x2_t __ret;
   2230   __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   2231   return __ret;
   2232 }
   2233 #else
   2234 __ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
   2235   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2236   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2237   uint32x2_t __ret;
   2238   __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   2239   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2240   return __ret;
   2241 }
   2242 #endif
   2243 
   2244 #ifdef __LITTLE_ENDIAN__
   2245 __ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
   2246   uint32x4_t __ret;
   2247   __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   2248   return __ret;
   2249 }
   2250 #else
   2251 __ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
   2252   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2253   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2254   uint32x4_t __ret;
   2255   __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   2256   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2257   return __ret;
   2258 }
   2259 #endif
   2260 
   2261 #ifdef __LITTLE_ENDIAN__
   2262 __ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
   2263   uint32x2_t __ret;
   2264   __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   2265   return __ret;
   2266 }
   2267 #else
   2268 __ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
   2269   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2270   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2271   uint32x2_t __ret;
   2272   __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   2273   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2274   return __ret;
   2275 }
   2276 #endif
   2277 
   2278 #ifdef __LITTLE_ENDIAN__
   2279 __ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
   2280   uint32x4_t __ret;
   2281   __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   2282   return __ret;
   2283 }
   2284 #else
   2285 __ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
   2286   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2287   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2288   uint32x4_t __ret;
   2289   __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   2290   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2291   return __ret;
   2292 }
   2293 #endif
   2294 
   2295 #ifdef __LITTLE_ENDIAN__
   2296 __ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
   2297   uint32x2_t __ret;
   2298   __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   2299   return __ret;
   2300 }
   2301 #else
   2302 __ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
   2303   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2304   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2305   uint32x2_t __ret;
   2306   __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   2307   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2308   return __ret;
   2309 }
   2310 #endif
   2311 
   2312 #ifdef __LITTLE_ENDIAN__
   2313 __ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
   2314   uint32x4_t __ret;
   2315   __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   2316   return __ret;
   2317 }
   2318 #else
   2319 __ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
   2320   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2321   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2322   uint32x4_t __ret;
   2323   __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   2324   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2325   return __ret;
   2326 }
   2327 #endif
   2328 
   2329 #ifdef __LITTLE_ENDIAN__
   2330 __ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
   2331   uint32x2_t __ret;
   2332   __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   2333   return __ret;
   2334 }
   2335 #else
   2336 __ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
   2337   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2338   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2339   uint32x2_t __ret;
   2340   __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   2341   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2342   return __ret;
   2343 }
   2344 #endif
   2345 
   2346 #ifdef __LITTLE_ENDIAN__
   2347 __ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
   2348   uint8x8_t __ret;
   2349   __ret = (uint8x8_t)(__p0 == __p1);
   2350   return __ret;
   2351 }
   2352 #else
   2353 __ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
   2354   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2355   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2356   uint8x8_t __ret;
   2357   __ret = (uint8x8_t)(__rev0 == __rev1);
   2358   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2359   return __ret;
   2360 }
   2361 #endif
   2362 
   2363 #ifdef __LITTLE_ENDIAN__
   2364 __ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   2365   uint8x16_t __ret;
   2366   __ret = (uint8x16_t)(__p0 == __p1);
   2367   return __ret;
   2368 }
   2369 #else
   2370 __ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   2371   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2372   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2373   uint8x16_t __ret;
   2374   __ret = (uint8x16_t)(__rev0 == __rev1);
   2375   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2376   return __ret;
   2377 }
   2378 #endif
   2379 
   2380 #ifdef __LITTLE_ENDIAN__
   2381 __ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   2382   uint8x16_t __ret;
   2383   __ret = (uint8x16_t)(__p0 == __p1);
   2384   return __ret;
   2385 }
   2386 #else
   2387 __ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   2388   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2389   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2390   uint8x16_t __ret;
   2391   __ret = (uint8x16_t)(__rev0 == __rev1);
   2392   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2393   return __ret;
   2394 }
   2395 #endif
   2396 
   2397 #ifdef __LITTLE_ENDIAN__
   2398 __ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   2399   uint32x4_t __ret;
   2400   __ret = (uint32x4_t)(__p0 == __p1);
   2401   return __ret;
   2402 }
   2403 #else
   2404 __ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   2405   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2406   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2407   uint32x4_t __ret;
   2408   __ret = (uint32x4_t)(__rev0 == __rev1);
   2409   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2410   return __ret;
   2411 }
   2412 #endif
   2413 
   2414 #ifdef __LITTLE_ENDIAN__
   2415 __ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   2416   uint16x8_t __ret;
   2417   __ret = (uint16x8_t)(__p0 == __p1);
   2418   return __ret;
   2419 }
   2420 #else
   2421 __ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   2422   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2423   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2424   uint16x8_t __ret;
   2425   __ret = (uint16x8_t)(__rev0 == __rev1);
   2426   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2427   return __ret;
   2428 }
   2429 #endif
   2430 
   2431 #ifdef __LITTLE_ENDIAN__
   2432 __ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
   2433   uint8x16_t __ret;
   2434   __ret = (uint8x16_t)(__p0 == __p1);
   2435   return __ret;
   2436 }
   2437 #else
   2438 __ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
   2439   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2440   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2441   uint8x16_t __ret;
   2442   __ret = (uint8x16_t)(__rev0 == __rev1);
   2443   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2444   return __ret;
   2445 }
   2446 #endif
   2447 
   2448 #ifdef __LITTLE_ENDIAN__
   2449 __ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
   2450   uint32x4_t __ret;
   2451   __ret = (uint32x4_t)(__p0 == __p1);
   2452   return __ret;
   2453 }
   2454 #else
   2455 __ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
   2456   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2457   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2458   uint32x4_t __ret;
   2459   __ret = (uint32x4_t)(__rev0 == __rev1);
   2460   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2461   return __ret;
   2462 }
   2463 #endif
   2464 
   2465 #ifdef __LITTLE_ENDIAN__
   2466 __ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
   2467   uint32x4_t __ret;
   2468   __ret = (uint32x4_t)(__p0 == __p1);
   2469   return __ret;
   2470 }
   2471 #else
   2472 __ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
   2473   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2474   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2475   uint32x4_t __ret;
   2476   __ret = (uint32x4_t)(__rev0 == __rev1);
   2477   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2478   return __ret;
   2479 }
   2480 #endif
   2481 
   2482 #ifdef __LITTLE_ENDIAN__
   2483 __ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
   2484   uint16x8_t __ret;
   2485   __ret = (uint16x8_t)(__p0 == __p1);
   2486   return __ret;
   2487 }
   2488 #else
   2489 __ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
   2490   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2491   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2492   uint16x8_t __ret;
   2493   __ret = (uint16x8_t)(__rev0 == __rev1);
   2494   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2495   return __ret;
   2496 }
   2497 #endif
   2498 
   2499 #ifdef __LITTLE_ENDIAN__
   2500 __ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
   2501   uint8x8_t __ret;
   2502   __ret = (uint8x8_t)(__p0 == __p1);
   2503   return __ret;
   2504 }
   2505 #else
   2506 __ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
   2507   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2508   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2509   uint8x8_t __ret;
   2510   __ret = (uint8x8_t)(__rev0 == __rev1);
   2511   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2512   return __ret;
   2513 }
   2514 #endif
   2515 
   2516 #ifdef __LITTLE_ENDIAN__
   2517 __ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
   2518   uint32x2_t __ret;
   2519   __ret = (uint32x2_t)(__p0 == __p1);
   2520   return __ret;
   2521 }
   2522 #else
   2523 __ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
   2524   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2525   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2526   uint32x2_t __ret;
   2527   __ret = (uint32x2_t)(__rev0 == __rev1);
   2528   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2529   return __ret;
   2530 }
   2531 #endif
   2532 
   2533 #ifdef __LITTLE_ENDIAN__
   2534 __ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
   2535   uint16x4_t __ret;
   2536   __ret = (uint16x4_t)(__p0 == __p1);
   2537   return __ret;
   2538 }
   2539 #else
   2540 __ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
   2541   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2542   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2543   uint16x4_t __ret;
   2544   __ret = (uint16x4_t)(__rev0 == __rev1);
   2545   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2546   return __ret;
   2547 }
   2548 #endif
   2549 
   2550 #ifdef __LITTLE_ENDIAN__
   2551 __ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
   2552   uint8x8_t __ret;
   2553   __ret = (uint8x8_t)(__p0 == __p1);
   2554   return __ret;
   2555 }
   2556 #else
   2557 __ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
   2558   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2559   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2560   uint8x8_t __ret;
   2561   __ret = (uint8x8_t)(__rev0 == __rev1);
   2562   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2563   return __ret;
   2564 }
   2565 #endif
   2566 
   2567 #ifdef __LITTLE_ENDIAN__
   2568 __ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
   2569   uint32x2_t __ret;
   2570   __ret = (uint32x2_t)(__p0 == __p1);
   2571   return __ret;
   2572 }
   2573 #else
   2574 __ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
   2575   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2576   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2577   uint32x2_t __ret;
   2578   __ret = (uint32x2_t)(__rev0 == __rev1);
   2579   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2580   return __ret;
   2581 }
   2582 #endif
   2583 
   2584 #ifdef __LITTLE_ENDIAN__
   2585 __ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
   2586   uint32x2_t __ret;
   2587   __ret = (uint32x2_t)(__p0 == __p1);
   2588   return __ret;
   2589 }
   2590 #else
   2591 __ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
   2592   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2593   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2594   uint32x2_t __ret;
   2595   __ret = (uint32x2_t)(__rev0 == __rev1);
   2596   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2597   return __ret;
   2598 }
   2599 #endif
   2600 
   2601 #ifdef __LITTLE_ENDIAN__
   2602 __ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
   2603   uint16x4_t __ret;
   2604   __ret = (uint16x4_t)(__p0 == __p1);
   2605   return __ret;
   2606 }
   2607 #else
   2608 __ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
   2609   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2610   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2611   uint16x4_t __ret;
   2612   __ret = (uint16x4_t)(__rev0 == __rev1);
   2613   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2614   return __ret;
   2615 }
   2616 #endif
   2617 
   2618 #ifdef __LITTLE_ENDIAN__
   2619 __ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   2620   uint8x16_t __ret;
   2621   __ret = (uint8x16_t)(__p0 >= __p1);
   2622   return __ret;
   2623 }
   2624 #else
   2625 __ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   2626   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2627   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2628   uint8x16_t __ret;
   2629   __ret = (uint8x16_t)(__rev0 >= __rev1);
   2630   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2631   return __ret;
   2632 }
   2633 #endif
   2634 
   2635 #ifdef __LITTLE_ENDIAN__
   2636 __ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   2637   uint32x4_t __ret;
   2638   __ret = (uint32x4_t)(__p0 >= __p1);
   2639   return __ret;
   2640 }
   2641 #else
   2642 __ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   2643   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2644   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2645   uint32x4_t __ret;
   2646   __ret = (uint32x4_t)(__rev0 >= __rev1);
   2647   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2648   return __ret;
   2649 }
   2650 #endif
   2651 
   2652 #ifdef __LITTLE_ENDIAN__
   2653 __ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   2654   uint16x8_t __ret;
   2655   __ret = (uint16x8_t)(__p0 >= __p1);
   2656   return __ret;
   2657 }
   2658 #else
   2659 __ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   2660   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2661   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2662   uint16x8_t __ret;
   2663   __ret = (uint16x8_t)(__rev0 >= __rev1);
   2664   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2665   return __ret;
   2666 }
   2667 #endif
   2668 
   2669 #ifdef __LITTLE_ENDIAN__
   2670 __ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
   2671   uint8x16_t __ret;
   2672   __ret = (uint8x16_t)(__p0 >= __p1);
   2673   return __ret;
   2674 }
   2675 #else
   2676 __ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
   2677   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2678   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2679   uint8x16_t __ret;
   2680   __ret = (uint8x16_t)(__rev0 >= __rev1);
   2681   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2682   return __ret;
   2683 }
   2684 #endif
   2685 
   2686 #ifdef __LITTLE_ENDIAN__
   2687 __ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
   2688   uint32x4_t __ret;
   2689   __ret = (uint32x4_t)(__p0 >= __p1);
   2690   return __ret;
   2691 }
   2692 #else
   2693 __ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
   2694   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2695   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2696   uint32x4_t __ret;
   2697   __ret = (uint32x4_t)(__rev0 >= __rev1);
   2698   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2699   return __ret;
   2700 }
   2701 #endif
   2702 
   2703 #ifdef __LITTLE_ENDIAN__
   2704 __ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
   2705   uint32x4_t __ret;
   2706   __ret = (uint32x4_t)(__p0 >= __p1);
   2707   return __ret;
   2708 }
   2709 #else
   2710 __ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
   2711   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2712   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2713   uint32x4_t __ret;
   2714   __ret = (uint32x4_t)(__rev0 >= __rev1);
   2715   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2716   return __ret;
   2717 }
   2718 #endif
   2719 
   2720 #ifdef __LITTLE_ENDIAN__
   2721 __ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
   2722   uint16x8_t __ret;
   2723   __ret = (uint16x8_t)(__p0 >= __p1);
   2724   return __ret;
   2725 }
   2726 #else
   2727 __ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
   2728   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2729   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2730   uint16x8_t __ret;
   2731   __ret = (uint16x8_t)(__rev0 >= __rev1);
   2732   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2733   return __ret;
   2734 }
   2735 #endif
   2736 
   2737 #ifdef __LITTLE_ENDIAN__
   2738 __ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
   2739   uint8x8_t __ret;
   2740   __ret = (uint8x8_t)(__p0 >= __p1);
   2741   return __ret;
   2742 }
   2743 #else
   2744 __ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
   2745   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2746   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2747   uint8x8_t __ret;
   2748   __ret = (uint8x8_t)(__rev0 >= __rev1);
   2749   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2750   return __ret;
   2751 }
   2752 #endif
   2753 
   2754 #ifdef __LITTLE_ENDIAN__
   2755 __ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
   2756   uint32x2_t __ret;
   2757   __ret = (uint32x2_t)(__p0 >= __p1);
   2758   return __ret;
   2759 }
   2760 #else
   2761 __ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
   2762   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2763   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2764   uint32x2_t __ret;
   2765   __ret = (uint32x2_t)(__rev0 >= __rev1);
   2766   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2767   return __ret;
   2768 }
   2769 #endif
   2770 
   2771 #ifdef __LITTLE_ENDIAN__
   2772 __ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
   2773   uint16x4_t __ret;
   2774   __ret = (uint16x4_t)(__p0 >= __p1);
   2775   return __ret;
   2776 }
   2777 #else
   2778 __ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
   2779   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2780   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2781   uint16x4_t __ret;
   2782   __ret = (uint16x4_t)(__rev0 >= __rev1);
   2783   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2784   return __ret;
   2785 }
   2786 #endif
   2787 
   2788 #ifdef __LITTLE_ENDIAN__
   2789 __ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
   2790   uint8x8_t __ret;
   2791   __ret = (uint8x8_t)(__p0 >= __p1);
   2792   return __ret;
   2793 }
   2794 #else
   2795 __ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
   2796   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2797   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2798   uint8x8_t __ret;
   2799   __ret = (uint8x8_t)(__rev0 >= __rev1);
   2800   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2801   return __ret;
   2802 }
   2803 #endif
   2804 
   2805 #ifdef __LITTLE_ENDIAN__
   2806 __ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
   2807   uint32x2_t __ret;
   2808   __ret = (uint32x2_t)(__p0 >= __p1);
   2809   return __ret;
   2810 }
   2811 #else
   2812 __ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
   2813   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2814   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2815   uint32x2_t __ret;
   2816   __ret = (uint32x2_t)(__rev0 >= __rev1);
   2817   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2818   return __ret;
   2819 }
   2820 #endif
   2821 
   2822 #ifdef __LITTLE_ENDIAN__
   2823 __ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
   2824   uint32x2_t __ret;
   2825   __ret = (uint32x2_t)(__p0 >= __p1);
   2826   return __ret;
   2827 }
   2828 #else
   2829 __ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
   2830   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   2831   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   2832   uint32x2_t __ret;
   2833   __ret = (uint32x2_t)(__rev0 >= __rev1);
   2834   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   2835   return __ret;
   2836 }
   2837 #endif
   2838 
   2839 #ifdef __LITTLE_ENDIAN__
   2840 __ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
   2841   uint16x4_t __ret;
   2842   __ret = (uint16x4_t)(__p0 >= __p1);
   2843   return __ret;
   2844 }
   2845 #else
   2846 __ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
   2847   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2848   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2849   uint16x4_t __ret;
   2850   __ret = (uint16x4_t)(__rev0 >= __rev1);
   2851   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2852   return __ret;
   2853 }
   2854 #endif
   2855 
   2856 #ifdef __LITTLE_ENDIAN__
   2857 __ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   2858   uint8x16_t __ret;
   2859   __ret = (uint8x16_t)(__p0 > __p1);
   2860   return __ret;
   2861 }
   2862 #else
   2863 __ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   2864   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2865   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2866   uint8x16_t __ret;
   2867   __ret = (uint8x16_t)(__rev0 > __rev1);
   2868   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2869   return __ret;
   2870 }
   2871 #endif
   2872 
   2873 #ifdef __LITTLE_ENDIAN__
   2874 __ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   2875   uint32x4_t __ret;
   2876   __ret = (uint32x4_t)(__p0 > __p1);
   2877   return __ret;
   2878 }
   2879 #else
   2880 __ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   2881   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2882   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2883   uint32x4_t __ret;
   2884   __ret = (uint32x4_t)(__rev0 > __rev1);
   2885   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2886   return __ret;
   2887 }
   2888 #endif
   2889 
   2890 #ifdef __LITTLE_ENDIAN__
   2891 __ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   2892   uint16x8_t __ret;
   2893   __ret = (uint16x8_t)(__p0 > __p1);
   2894   return __ret;
   2895 }
   2896 #else
   2897 __ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   2898   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2899   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2900   uint16x8_t __ret;
   2901   __ret = (uint16x8_t)(__rev0 > __rev1);
   2902   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2903   return __ret;
   2904 }
   2905 #endif
   2906 
   2907 #ifdef __LITTLE_ENDIAN__
   2908 __ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
   2909   uint8x16_t __ret;
   2910   __ret = (uint8x16_t)(__p0 > __p1);
   2911   return __ret;
   2912 }
   2913 #else
   2914 __ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
   2915   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2916   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2917   uint8x16_t __ret;
   2918   __ret = (uint8x16_t)(__rev0 > __rev1);
   2919   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   2920   return __ret;
   2921 }
   2922 #endif
   2923 
   2924 #ifdef __LITTLE_ENDIAN__
   2925 __ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
   2926   uint32x4_t __ret;
   2927   __ret = (uint32x4_t)(__p0 > __p1);
   2928   return __ret;
   2929 }
   2930 #else
   2931 __ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
   2932   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2933   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2934   uint32x4_t __ret;
   2935   __ret = (uint32x4_t)(__rev0 > __rev1);
   2936   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2937   return __ret;
   2938 }
   2939 #endif
   2940 
   2941 #ifdef __LITTLE_ENDIAN__
   2942 __ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
   2943   uint32x4_t __ret;
   2944   __ret = (uint32x4_t)(__p0 > __p1);
   2945   return __ret;
   2946 }
   2947 #else
   2948 __ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
   2949   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   2950   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   2951   uint32x4_t __ret;
   2952   __ret = (uint32x4_t)(__rev0 > __rev1);
   2953   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   2954   return __ret;
   2955 }
   2956 #endif
   2957 
   2958 #ifdef __LITTLE_ENDIAN__
   2959 __ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
   2960   uint16x8_t __ret;
   2961   __ret = (uint16x8_t)(__p0 > __p1);
   2962   return __ret;
   2963 }
   2964 #else
   2965 __ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
   2966   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2967   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2968   uint16x8_t __ret;
   2969   __ret = (uint16x8_t)(__rev0 > __rev1);
   2970   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2971   return __ret;
   2972 }
   2973 #endif
   2974 
   2975 #ifdef __LITTLE_ENDIAN__
   2976 __ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
   2977   uint8x8_t __ret;
   2978   __ret = (uint8x8_t)(__p0 > __p1);
   2979   return __ret;
   2980 }
   2981 #else
   2982 __ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
   2983   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   2984   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   2985   uint8x8_t __ret;
   2986   __ret = (uint8x8_t)(__rev0 > __rev1);
   2987   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   2988   return __ret;
   2989 }
   2990 #endif
   2991 
   2992 #ifdef __LITTLE_ENDIAN__
   2993 __ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
   2994   uint32x2_t __ret;
   2995   __ret = (uint32x2_t)(__p0 > __p1);
   2996   return __ret;
   2997 }
   2998 #else
   2999 __ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
   3000   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3001   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   3002   uint32x2_t __ret;
   3003   __ret = (uint32x2_t)(__rev0 > __rev1);
   3004   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3005   return __ret;
   3006 }
   3007 #endif
   3008 
   3009 #ifdef __LITTLE_ENDIAN__
   3010 __ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
   3011   uint16x4_t __ret;
   3012   __ret = (uint16x4_t)(__p0 > __p1);
   3013   return __ret;
   3014 }
   3015 #else
   3016 __ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
   3017   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3018   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3019   uint16x4_t __ret;
   3020   __ret = (uint16x4_t)(__rev0 > __rev1);
   3021   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3022   return __ret;
   3023 }
   3024 #endif
   3025 
   3026 #ifdef __LITTLE_ENDIAN__
   3027 __ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
   3028   uint8x8_t __ret;
   3029   __ret = (uint8x8_t)(__p0 > __p1);
   3030   return __ret;
   3031 }
   3032 #else
   3033 __ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
   3034   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3035   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3036   uint8x8_t __ret;
   3037   __ret = (uint8x8_t)(__rev0 > __rev1);
   3038   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3039   return __ret;
   3040 }
   3041 #endif
   3042 
   3043 #ifdef __LITTLE_ENDIAN__
   3044 __ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
   3045   uint32x2_t __ret;
   3046   __ret = (uint32x2_t)(__p0 > __p1);
   3047   return __ret;
   3048 }
   3049 #else
   3050 __ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
   3051   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3052   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   3053   uint32x2_t __ret;
   3054   __ret = (uint32x2_t)(__rev0 > __rev1);
   3055   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3056   return __ret;
   3057 }
   3058 #endif
   3059 
   3060 #ifdef __LITTLE_ENDIAN__
   3061 __ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
   3062   uint32x2_t __ret;
   3063   __ret = (uint32x2_t)(__p0 > __p1);
   3064   return __ret;
   3065 }
   3066 #else
   3067 __ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
   3068   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3069   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   3070   uint32x2_t __ret;
   3071   __ret = (uint32x2_t)(__rev0 > __rev1);
   3072   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3073   return __ret;
   3074 }
   3075 #endif
   3076 
   3077 #ifdef __LITTLE_ENDIAN__
   3078 __ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
   3079   uint16x4_t __ret;
   3080   __ret = (uint16x4_t)(__p0 > __p1);
   3081   return __ret;
   3082 }
   3083 #else
   3084 __ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
   3085   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3086   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3087   uint16x4_t __ret;
   3088   __ret = (uint16x4_t)(__rev0 > __rev1);
   3089   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3090   return __ret;
   3091 }
   3092 #endif
   3093 
   3094 #ifdef __LITTLE_ENDIAN__
   3095 __ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   3096   uint8x16_t __ret;
   3097   __ret = (uint8x16_t)(__p0 <= __p1);
   3098   return __ret;
   3099 }
   3100 #else
   3101 __ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   3102   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3103   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3104   uint8x16_t __ret;
   3105   __ret = (uint8x16_t)(__rev0 <= __rev1);
   3106   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3107   return __ret;
   3108 }
   3109 #endif
   3110 
   3111 #ifdef __LITTLE_ENDIAN__
   3112 __ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   3113   uint32x4_t __ret;
   3114   __ret = (uint32x4_t)(__p0 <= __p1);
   3115   return __ret;
   3116 }
   3117 #else
   3118 __ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   3119   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3120   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3121   uint32x4_t __ret;
   3122   __ret = (uint32x4_t)(__rev0 <= __rev1);
   3123   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3124   return __ret;
   3125 }
   3126 #endif
   3127 
   3128 #ifdef __LITTLE_ENDIAN__
   3129 __ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   3130   uint16x8_t __ret;
   3131   __ret = (uint16x8_t)(__p0 <= __p1);
   3132   return __ret;
   3133 }
   3134 #else
   3135 __ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   3136   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3137   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3138   uint16x8_t __ret;
   3139   __ret = (uint16x8_t)(__rev0 <= __rev1);
   3140   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3141   return __ret;
   3142 }
   3143 #endif
   3144 
   3145 #ifdef __LITTLE_ENDIAN__
   3146 __ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
   3147   uint8x16_t __ret;
   3148   __ret = (uint8x16_t)(__p0 <= __p1);
   3149   return __ret;
   3150 }
   3151 #else
   3152 __ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
   3153   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3154   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3155   uint8x16_t __ret;
   3156   __ret = (uint8x16_t)(__rev0 <= __rev1);
   3157   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3158   return __ret;
   3159 }
   3160 #endif
   3161 
   3162 #ifdef __LITTLE_ENDIAN__
   3163 __ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
   3164   uint32x4_t __ret;
   3165   __ret = (uint32x4_t)(__p0 <= __p1);
   3166   return __ret;
   3167 }
   3168 #else
   3169 __ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
   3170   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3171   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3172   uint32x4_t __ret;
   3173   __ret = (uint32x4_t)(__rev0 <= __rev1);
   3174   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3175   return __ret;
   3176 }
   3177 #endif
   3178 
   3179 #ifdef __LITTLE_ENDIAN__
   3180 __ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
   3181   uint32x4_t __ret;
   3182   __ret = (uint32x4_t)(__p0 <= __p1);
   3183   return __ret;
   3184 }
   3185 #else
   3186 __ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
   3187   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3188   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3189   uint32x4_t __ret;
   3190   __ret = (uint32x4_t)(__rev0 <= __rev1);
   3191   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3192   return __ret;
   3193 }
   3194 #endif
   3195 
   3196 #ifdef __LITTLE_ENDIAN__
   3197 __ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
   3198   uint16x8_t __ret;
   3199   __ret = (uint16x8_t)(__p0 <= __p1);
   3200   return __ret;
   3201 }
   3202 #else
   3203 __ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
   3204   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3205   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3206   uint16x8_t __ret;
   3207   __ret = (uint16x8_t)(__rev0 <= __rev1);
   3208   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3209   return __ret;
   3210 }
   3211 #endif
   3212 
   3213 #ifdef __LITTLE_ENDIAN__
   3214 __ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
   3215   uint8x8_t __ret;
   3216   __ret = (uint8x8_t)(__p0 <= __p1);
   3217   return __ret;
   3218 }
   3219 #else
   3220 __ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
   3221   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3222   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3223   uint8x8_t __ret;
   3224   __ret = (uint8x8_t)(__rev0 <= __rev1);
   3225   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3226   return __ret;
   3227 }
   3228 #endif
   3229 
   3230 #ifdef __LITTLE_ENDIAN__
   3231 __ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
   3232   uint32x2_t __ret;
   3233   __ret = (uint32x2_t)(__p0 <= __p1);
   3234   return __ret;
   3235 }
   3236 #else
   3237 __ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
   3238   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3239   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   3240   uint32x2_t __ret;
   3241   __ret = (uint32x2_t)(__rev0 <= __rev1);
   3242   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3243   return __ret;
   3244 }
   3245 #endif
   3246 
   3247 #ifdef __LITTLE_ENDIAN__
   3248 __ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
   3249   uint16x4_t __ret;
   3250   __ret = (uint16x4_t)(__p0 <= __p1);
   3251   return __ret;
   3252 }
   3253 #else
   3254 __ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
   3255   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3256   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3257   uint16x4_t __ret;
   3258   __ret = (uint16x4_t)(__rev0 <= __rev1);
   3259   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3260   return __ret;
   3261 }
   3262 #endif
   3263 
   3264 #ifdef __LITTLE_ENDIAN__
   3265 __ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
   3266   uint8x8_t __ret;
   3267   __ret = (uint8x8_t)(__p0 <= __p1);
   3268   return __ret;
   3269 }
   3270 #else
   3271 __ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
   3272   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3273   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3274   uint8x8_t __ret;
   3275   __ret = (uint8x8_t)(__rev0 <= __rev1);
   3276   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3277   return __ret;
   3278 }
   3279 #endif
   3280 
   3281 #ifdef __LITTLE_ENDIAN__
   3282 __ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
   3283   uint32x2_t __ret;
   3284   __ret = (uint32x2_t)(__p0 <= __p1);
   3285   return __ret;
   3286 }
   3287 #else
   3288 __ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
   3289   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3290   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   3291   uint32x2_t __ret;
   3292   __ret = (uint32x2_t)(__rev0 <= __rev1);
   3293   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3294   return __ret;
   3295 }
   3296 #endif
   3297 
   3298 #ifdef __LITTLE_ENDIAN__
   3299 __ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
   3300   uint32x2_t __ret;
   3301   __ret = (uint32x2_t)(__p0 <= __p1);
   3302   return __ret;
   3303 }
   3304 #else
   3305 __ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
   3306   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3307   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   3308   uint32x2_t __ret;
   3309   __ret = (uint32x2_t)(__rev0 <= __rev1);
   3310   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3311   return __ret;
   3312 }
   3313 #endif
   3314 
   3315 #ifdef __LITTLE_ENDIAN__
   3316 __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
   3317   uint16x4_t __ret;
   3318   __ret = (uint16x4_t)(__p0 <= __p1);
   3319   return __ret;
   3320 }
   3321 #else
   3322 __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
   3323   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3324   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3325   uint16x4_t __ret;
   3326   __ret = (uint16x4_t)(__rev0 <= __rev1);
   3327   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3328   return __ret;
   3329 }
   3330 #endif
   3331 
   3332 #ifdef __LITTLE_ENDIAN__
   3333 __ai int8x16_t vclsq_s8(int8x16_t __p0) {
   3334   int8x16_t __ret;
   3335   __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32);
   3336   return __ret;
   3337 }
   3338 #else
   3339 __ai int8x16_t vclsq_s8(int8x16_t __p0) {
   3340   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3341   int8x16_t __ret;
   3342   __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32);
   3343   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3344   return __ret;
   3345 }
   3346 #endif
   3347 
   3348 #ifdef __LITTLE_ENDIAN__
   3349 __ai int32x4_t vclsq_s32(int32x4_t __p0) {
   3350   int32x4_t __ret;
   3351   __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34);
   3352   return __ret;
   3353 }
   3354 #else
   3355 __ai int32x4_t vclsq_s32(int32x4_t __p0) {
   3356   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3357   int32x4_t __ret;
   3358   __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34);
   3359   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3360   return __ret;
   3361 }
   3362 #endif
   3363 
   3364 #ifdef __LITTLE_ENDIAN__
   3365 __ai int16x8_t vclsq_s16(int16x8_t __p0) {
   3366   int16x8_t __ret;
   3367   __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33);
   3368   return __ret;
   3369 }
   3370 #else
   3371 __ai int16x8_t vclsq_s16(int16x8_t __p0) {
   3372   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3373   int16x8_t __ret;
   3374   __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33);
   3375   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3376   return __ret;
   3377 }
   3378 #endif
   3379 
   3380 #ifdef __LITTLE_ENDIAN__
   3381 __ai int8x8_t vcls_s8(int8x8_t __p0) {
   3382   int8x8_t __ret;
   3383   __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0);
   3384   return __ret;
   3385 }
   3386 #else
   3387 __ai int8x8_t vcls_s8(int8x8_t __p0) {
   3388   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3389   int8x8_t __ret;
   3390   __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0);
   3391   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3392   return __ret;
   3393 }
   3394 #endif
   3395 
   3396 #ifdef __LITTLE_ENDIAN__
   3397 __ai int32x2_t vcls_s32(int32x2_t __p0) {
   3398   int32x2_t __ret;
   3399   __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2);
   3400   return __ret;
   3401 }
   3402 #else
   3403 __ai int32x2_t vcls_s32(int32x2_t __p0) {
   3404   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3405   int32x2_t __ret;
   3406   __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2);
   3407   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3408   return __ret;
   3409 }
   3410 #endif
   3411 
   3412 #ifdef __LITTLE_ENDIAN__
   3413 __ai int16x4_t vcls_s16(int16x4_t __p0) {
   3414   int16x4_t __ret;
   3415   __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1);
   3416   return __ret;
   3417 }
   3418 #else
   3419 __ai int16x4_t vcls_s16(int16x4_t __p0) {
   3420   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3421   int16x4_t __ret;
   3422   __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1);
   3423   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3424   return __ret;
   3425 }
   3426 #endif
   3427 
   3428 #ifdef __LITTLE_ENDIAN__
   3429 __ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   3430   uint8x16_t __ret;
   3431   __ret = (uint8x16_t)(__p0 < __p1);
   3432   return __ret;
   3433 }
   3434 #else
   3435 __ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   3436   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3437   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3438   uint8x16_t __ret;
   3439   __ret = (uint8x16_t)(__rev0 < __rev1);
   3440   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3441   return __ret;
   3442 }
   3443 #endif
   3444 
   3445 #ifdef __LITTLE_ENDIAN__
   3446 __ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   3447   uint32x4_t __ret;
   3448   __ret = (uint32x4_t)(__p0 < __p1);
   3449   return __ret;
   3450 }
   3451 #else
   3452 __ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   3453   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3454   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3455   uint32x4_t __ret;
   3456   __ret = (uint32x4_t)(__rev0 < __rev1);
   3457   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3458   return __ret;
   3459 }
   3460 #endif
   3461 
   3462 #ifdef __LITTLE_ENDIAN__
   3463 __ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   3464   uint16x8_t __ret;
   3465   __ret = (uint16x8_t)(__p0 < __p1);
   3466   return __ret;
   3467 }
   3468 #else
   3469 __ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   3470   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3471   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3472   uint16x8_t __ret;
   3473   __ret = (uint16x8_t)(__rev0 < __rev1);
   3474   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3475   return __ret;
   3476 }
   3477 #endif
   3478 
   3479 #ifdef __LITTLE_ENDIAN__
   3480 __ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
   3481   uint8x16_t __ret;
   3482   __ret = (uint8x16_t)(__p0 < __p1);
   3483   return __ret;
   3484 }
   3485 #else
   3486 __ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
   3487   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3488   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3489   uint8x16_t __ret;
   3490   __ret = (uint8x16_t)(__rev0 < __rev1);
   3491   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3492   return __ret;
   3493 }
   3494 #endif
   3495 
   3496 #ifdef __LITTLE_ENDIAN__
   3497 __ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
   3498   uint32x4_t __ret;
   3499   __ret = (uint32x4_t)(__p0 < __p1);
   3500   return __ret;
   3501 }
   3502 #else
   3503 __ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
   3504   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3505   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3506   uint32x4_t __ret;
   3507   __ret = (uint32x4_t)(__rev0 < __rev1);
   3508   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3509   return __ret;
   3510 }
   3511 #endif
   3512 
   3513 #ifdef __LITTLE_ENDIAN__
   3514 __ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
   3515   uint32x4_t __ret;
   3516   __ret = (uint32x4_t)(__p0 < __p1);
   3517   return __ret;
   3518 }
   3519 #else
   3520 __ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
   3521   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3522   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3523   uint32x4_t __ret;
   3524   __ret = (uint32x4_t)(__rev0 < __rev1);
   3525   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3526   return __ret;
   3527 }
   3528 #endif
   3529 
   3530 #ifdef __LITTLE_ENDIAN__
   3531 __ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
   3532   uint16x8_t __ret;
   3533   __ret = (uint16x8_t)(__p0 < __p1);
   3534   return __ret;
   3535 }
   3536 #else
   3537 __ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
   3538   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3539   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3540   uint16x8_t __ret;
   3541   __ret = (uint16x8_t)(__rev0 < __rev1);
   3542   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3543   return __ret;
   3544 }
   3545 #endif
   3546 
   3547 #ifdef __LITTLE_ENDIAN__
   3548 __ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
   3549   uint8x8_t __ret;
   3550   __ret = (uint8x8_t)(__p0 < __p1);
   3551   return __ret;
   3552 }
   3553 #else
   3554 __ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
   3555   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3556   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3557   uint8x8_t __ret;
   3558   __ret = (uint8x8_t)(__rev0 < __rev1);
   3559   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3560   return __ret;
   3561 }
   3562 #endif
   3563 
   3564 #ifdef __LITTLE_ENDIAN__
   3565 __ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
   3566   uint32x2_t __ret;
   3567   __ret = (uint32x2_t)(__p0 < __p1);
   3568   return __ret;
   3569 }
   3570 #else
   3571 __ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
   3572   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3573   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   3574   uint32x2_t __ret;
   3575   __ret = (uint32x2_t)(__rev0 < __rev1);
   3576   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3577   return __ret;
   3578 }
   3579 #endif
   3580 
   3581 #ifdef __LITTLE_ENDIAN__
   3582 __ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
   3583   uint16x4_t __ret;
   3584   __ret = (uint16x4_t)(__p0 < __p1);
   3585   return __ret;
   3586 }
   3587 #else
   3588 __ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
   3589   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3590   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3591   uint16x4_t __ret;
   3592   __ret = (uint16x4_t)(__rev0 < __rev1);
   3593   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3594   return __ret;
   3595 }
   3596 #endif
   3597 
   3598 #ifdef __LITTLE_ENDIAN__
   3599 __ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
   3600   uint8x8_t __ret;
   3601   __ret = (uint8x8_t)(__p0 < __p1);
   3602   return __ret;
   3603 }
   3604 #else
   3605 __ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
   3606   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3607   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3608   uint8x8_t __ret;
   3609   __ret = (uint8x8_t)(__rev0 < __rev1);
   3610   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3611   return __ret;
   3612 }
   3613 #endif
   3614 
   3615 #ifdef __LITTLE_ENDIAN__
   3616 __ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
   3617   uint32x2_t __ret;
   3618   __ret = (uint32x2_t)(__p0 < __p1);
   3619   return __ret;
   3620 }
   3621 #else
   3622 __ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
   3623   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3624   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   3625   uint32x2_t __ret;
   3626   __ret = (uint32x2_t)(__rev0 < __rev1);
   3627   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3628   return __ret;
   3629 }
   3630 #endif
   3631 
   3632 #ifdef __LITTLE_ENDIAN__
   3633 __ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
   3634   uint32x2_t __ret;
   3635   __ret = (uint32x2_t)(__p0 < __p1);
   3636   return __ret;
   3637 }
   3638 #else
   3639 __ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
   3640   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3641   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   3642   uint32x2_t __ret;
   3643   __ret = (uint32x2_t)(__rev0 < __rev1);
   3644   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3645   return __ret;
   3646 }
   3647 #endif
   3648 
   3649 #ifdef __LITTLE_ENDIAN__
   3650 __ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
   3651   uint16x4_t __ret;
   3652   __ret = (uint16x4_t)(__p0 < __p1);
   3653   return __ret;
   3654 }
   3655 #else
   3656 __ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
   3657   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3658   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3659   uint16x4_t __ret;
   3660   __ret = (uint16x4_t)(__rev0 < __rev1);
   3661   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3662   return __ret;
   3663 }
   3664 #endif
   3665 
   3666 #ifdef __LITTLE_ENDIAN__
   3667 __ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
   3668   uint8x16_t __ret;
   3669   __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 48);
   3670   return __ret;
   3671 }
   3672 #else
   3673 __ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
   3674   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3675   uint8x16_t __ret;
   3676   __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 48);
   3677   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3678   return __ret;
   3679 }
   3680 #endif
   3681 
   3682 #ifdef __LITTLE_ENDIAN__
   3683 __ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
   3684   uint32x4_t __ret;
   3685   __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 50);
   3686   return __ret;
   3687 }
   3688 #else
   3689 __ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
   3690   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3691   uint32x4_t __ret;
   3692   __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 50);
   3693   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3694   return __ret;
   3695 }
   3696 #endif
   3697 
   3698 #ifdef __LITTLE_ENDIAN__
   3699 __ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
   3700   uint16x8_t __ret;
   3701   __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 49);
   3702   return __ret;
   3703 }
   3704 #else
   3705 __ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
   3706   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3707   uint16x8_t __ret;
   3708   __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 49);
   3709   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3710   return __ret;
   3711 }
   3712 #endif
   3713 
   3714 #ifdef __LITTLE_ENDIAN__
   3715 __ai int8x16_t vclzq_s8(int8x16_t __p0) {
   3716   int8x16_t __ret;
   3717   __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 32);
   3718   return __ret;
   3719 }
   3720 #else
   3721 __ai int8x16_t vclzq_s8(int8x16_t __p0) {
   3722   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3723   int8x16_t __ret;
   3724   __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 32);
   3725   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3726   return __ret;
   3727 }
   3728 #endif
   3729 
   3730 #ifdef __LITTLE_ENDIAN__
   3731 __ai int32x4_t vclzq_s32(int32x4_t __p0) {
   3732   int32x4_t __ret;
   3733   __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 34);
   3734   return __ret;
   3735 }
   3736 #else
   3737 __ai int32x4_t vclzq_s32(int32x4_t __p0) {
   3738   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3739   int32x4_t __ret;
   3740   __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 34);
   3741   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3742   return __ret;
   3743 }
   3744 #endif
   3745 
   3746 #ifdef __LITTLE_ENDIAN__
   3747 __ai int16x8_t vclzq_s16(int16x8_t __p0) {
   3748   int16x8_t __ret;
   3749   __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 33);
   3750   return __ret;
   3751 }
   3752 #else
   3753 __ai int16x8_t vclzq_s16(int16x8_t __p0) {
   3754   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3755   int16x8_t __ret;
   3756   __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 33);
   3757   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3758   return __ret;
   3759 }
   3760 #endif
   3761 
   3762 #ifdef __LITTLE_ENDIAN__
   3763 __ai uint8x8_t vclz_u8(uint8x8_t __p0) {
   3764   uint8x8_t __ret;
   3765   __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 16);
   3766   return __ret;
   3767 }
   3768 #else
   3769 __ai uint8x8_t vclz_u8(uint8x8_t __p0) {
   3770   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3771   uint8x8_t __ret;
   3772   __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 16);
   3773   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3774   return __ret;
   3775 }
   3776 #endif
   3777 
   3778 #ifdef __LITTLE_ENDIAN__
   3779 __ai uint32x2_t vclz_u32(uint32x2_t __p0) {
   3780   uint32x2_t __ret;
   3781   __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 18);
   3782   return __ret;
   3783 }
   3784 #else
   3785 __ai uint32x2_t vclz_u32(uint32x2_t __p0) {
   3786   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3787   uint32x2_t __ret;
   3788   __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 18);
   3789   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3790   return __ret;
   3791 }
   3792 #endif
   3793 
   3794 #ifdef __LITTLE_ENDIAN__
   3795 __ai uint16x4_t vclz_u16(uint16x4_t __p0) {
   3796   uint16x4_t __ret;
   3797   __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 17);
   3798   return __ret;
   3799 }
   3800 #else
   3801 __ai uint16x4_t vclz_u16(uint16x4_t __p0) {
   3802   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3803   uint16x4_t __ret;
   3804   __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 17);
   3805   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3806   return __ret;
   3807 }
   3808 #endif
   3809 
   3810 #ifdef __LITTLE_ENDIAN__
   3811 __ai int8x8_t vclz_s8(int8x8_t __p0) {
   3812   int8x8_t __ret;
   3813   __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 0);
   3814   return __ret;
   3815 }
   3816 #else
   3817 __ai int8x8_t vclz_s8(int8x8_t __p0) {
   3818   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3819   int8x8_t __ret;
   3820   __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 0);
   3821   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3822   return __ret;
   3823 }
   3824 #endif
   3825 
   3826 #ifdef __LITTLE_ENDIAN__
   3827 __ai int32x2_t vclz_s32(int32x2_t __p0) {
   3828   int32x2_t __ret;
   3829   __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 2);
   3830   return __ret;
   3831 }
   3832 #else
   3833 __ai int32x2_t vclz_s32(int32x2_t __p0) {
   3834   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   3835   int32x2_t __ret;
   3836   __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 2);
   3837   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   3838   return __ret;
   3839 }
   3840 #endif
   3841 
   3842 #ifdef __LITTLE_ENDIAN__
   3843 __ai int16x4_t vclz_s16(int16x4_t __p0) {
   3844   int16x4_t __ret;
   3845   __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 1);
   3846   return __ret;
   3847 }
   3848 #else
   3849 __ai int16x4_t vclz_s16(int16x4_t __p0) {
   3850   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3851   int16x4_t __ret;
   3852   __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 1);
   3853   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   3854   return __ret;
   3855 }
   3856 #endif
   3857 
   3858 #ifdef __LITTLE_ENDIAN__
   3859 __ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
   3860   poly8x8_t __ret;
   3861   __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 4);
   3862   return __ret;
   3863 }
   3864 #else
   3865 __ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
   3866   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3867   poly8x8_t __ret;
   3868   __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 4);
   3869   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3870   return __ret;
   3871 }
   3872 #endif
   3873 
   3874 #ifdef __LITTLE_ENDIAN__
   3875 __ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
   3876   poly8x16_t __ret;
   3877   __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 36);
   3878   return __ret;
   3879 }
   3880 #else
   3881 __ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
   3882   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3883   poly8x16_t __ret;
   3884   __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 36);
   3885   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3886   return __ret;
   3887 }
   3888 #endif
   3889 
   3890 #ifdef __LITTLE_ENDIAN__
   3891 __ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
   3892   uint8x16_t __ret;
   3893   __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 48);
   3894   return __ret;
   3895 }
   3896 #else
   3897 __ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
   3898   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3899   uint8x16_t __ret;
   3900   __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 48);
   3901   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3902   return __ret;
   3903 }
   3904 #endif
   3905 
   3906 #ifdef __LITTLE_ENDIAN__
   3907 __ai int8x16_t vcntq_s8(int8x16_t __p0) {
   3908   int8x16_t __ret;
   3909   __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 32);
   3910   return __ret;
   3911 }
   3912 #else
   3913 __ai int8x16_t vcntq_s8(int8x16_t __p0) {
   3914   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3915   int8x16_t __ret;
   3916   __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 32);
   3917   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3918   return __ret;
   3919 }
   3920 #endif
   3921 
   3922 #ifdef __LITTLE_ENDIAN__
   3923 __ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
   3924   uint8x8_t __ret;
   3925   __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 16);
   3926   return __ret;
   3927 }
   3928 #else
   3929 __ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
   3930   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3931   uint8x8_t __ret;
   3932   __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 16);
   3933   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3934   return __ret;
   3935 }
   3936 #endif
   3937 
   3938 #ifdef __LITTLE_ENDIAN__
   3939 __ai int8x8_t vcnt_s8(int8x8_t __p0) {
   3940   int8x8_t __ret;
   3941   __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 0);
   3942   return __ret;
   3943 }
   3944 #else
   3945 __ai int8x8_t vcnt_s8(int8x8_t __p0) {
   3946   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3947   int8x8_t __ret;
   3948   __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 0);
   3949   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3950   return __ret;
   3951 }
   3952 #endif
   3953 
   3954 #ifdef __LITTLE_ENDIAN__
   3955 __ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
   3956   poly8x16_t __ret;
   3957   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
   3958   return __ret;
   3959 }
   3960 #else
   3961 __ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
   3962   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3963   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3964   poly8x16_t __ret;
   3965   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
   3966   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   3967   return __ret;
   3968 }
   3969 #endif
   3970 
   3971 #ifdef __LITTLE_ENDIAN__
   3972 __ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
   3973   poly16x8_t __ret;
   3974   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
   3975   return __ret;
   3976 }
   3977 #else
   3978 __ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
   3979   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   3980   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   3981   poly16x8_t __ret;
   3982   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
   3983   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   3984   return __ret;
   3985 }
   3986 #endif
   3987 
   3988 #ifdef __LITTLE_ENDIAN__
   3989 __ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
   3990   uint8x16_t __ret;
   3991   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
   3992   return __ret;
   3993 }
   3994 #else
   3995 __ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
   3996   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   3997   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   3998   uint8x16_t __ret;
   3999   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
   4000   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   4001   return __ret;
   4002 }
   4003 __ai uint8x16_t __noswap_vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
   4004   uint8x16_t __ret;
   4005   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
   4006   return __ret;
   4007 }
   4008 #endif
   4009 
   4010 #ifdef __LITTLE_ENDIAN__
   4011 __ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
   4012   uint32x4_t __ret;
   4013   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
   4014   return __ret;
   4015 }
   4016 #else
   4017 __ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
   4018   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   4019   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   4020   uint32x4_t __ret;
   4021   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
   4022   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   4023   return __ret;
   4024 }
   4025 __ai uint32x4_t __noswap_vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
   4026   uint32x4_t __ret;
   4027   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
   4028   return __ret;
   4029 }
   4030 #endif
   4031 
   4032 #ifdef __LITTLE_ENDIAN__
   4033 __ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
   4034   uint64x2_t __ret;
   4035   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
   4036   return __ret;
   4037 }
   4038 #else
   4039 __ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
   4040   uint64x2_t __ret;
   4041   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
   4042   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   4043   return __ret;
   4044 }
   4045 #endif
   4046 
   4047 #ifdef __LITTLE_ENDIAN__
   4048 __ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
   4049   uint16x8_t __ret;
   4050   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
   4051   return __ret;
   4052 }
   4053 #else
   4054 __ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
   4055   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   4056   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   4057   uint16x8_t __ret;
   4058   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
   4059   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   4060   return __ret;
   4061 }
   4062 __ai uint16x8_t __noswap_vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
   4063   uint16x8_t __ret;
   4064   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
   4065   return __ret;
   4066 }
   4067 #endif
   4068 
   4069 #ifdef __LITTLE_ENDIAN__
   4070 __ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
   4071   int8x16_t __ret;
   4072   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
   4073   return __ret;
   4074 }
   4075 #else
   4076 __ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
   4077   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   4078   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   4079   int8x16_t __ret;
   4080   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
   4081   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   4082   return __ret;
   4083 }
   4084 __ai int8x16_t __noswap_vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
   4085   int8x16_t __ret;
   4086   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
   4087   return __ret;
   4088 }
   4089 #endif
   4090 
   4091 #ifdef __LITTLE_ENDIAN__
   4092 __ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
   4093   float32x4_t __ret;
   4094   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
   4095   return __ret;
   4096 }
   4097 #else
   4098 __ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
   4099   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   4100   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   4101   float32x4_t __ret;
   4102   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
   4103   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   4104   return __ret;
   4105 }
   4106 __ai float32x4_t __noswap_vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
   4107   float32x4_t __ret;
   4108   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
   4109   return __ret;
   4110 }
   4111 #endif
   4112 
   4113 #ifdef __LITTLE_ENDIAN__
   4114 __ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
   4115   float16x8_t __ret;
   4116   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
   4117   return __ret;
   4118 }
   4119 #else
   4120 __ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
   4121   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   4122   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   4123   float16x8_t __ret;
   4124   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
   4125   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   4126   return __ret;
   4127 }
   4128 __ai float16x8_t __noswap_vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
   4129   float16x8_t __ret;
   4130   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
   4131   return __ret;
   4132 }
   4133 #endif
   4134 
   4135 #ifdef __LITTLE_ENDIAN__
   4136 __ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
   4137   int32x4_t __ret;
   4138   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
   4139   return __ret;
   4140 }
   4141 #else
   4142 __ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
   4143   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   4144   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   4145   int32x4_t __ret;
   4146   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
   4147   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   4148   return __ret;
   4149 }
   4150 __ai int32x4_t __noswap_vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
   4151   int32x4_t __ret;
   4152   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
   4153   return __ret;
   4154 }
   4155 #endif
   4156 
   4157 #ifdef __LITTLE_ENDIAN__
   4158 __ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
   4159   int64x2_t __ret;
   4160   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
   4161   return __ret;
   4162 }
   4163 #else
   4164 __ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
   4165   int64x2_t __ret;
   4166   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
   4167   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   4168   return __ret;
   4169 }
   4170 #endif
   4171 
   4172 #ifdef __LITTLE_ENDIAN__
   4173 __ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
   4174   int16x8_t __ret;
   4175   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
   4176   return __ret;
   4177 }
   4178 #else
   4179 __ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
   4180   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   4181   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   4182   int16x8_t __ret;
   4183   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
   4184   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   4185   return __ret;
   4186 }
   4187 __ai int16x8_t __noswap_vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
   4188   int16x8_t __ret;
   4189   __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
   4190   return __ret;
   4191 }
   4192 #endif
   4193 
   4194 #ifdef __LITTLE_ENDIAN__
   4195 __ai poly8x8_t vcreate_p8(uint64_t __p0) {
   4196   poly8x8_t __ret;
   4197   __ret = (poly8x8_t)(__p0);
   4198   return __ret;
   4199 }
   4200 #else
   4201 __ai poly8x8_t vcreate_p8(uint64_t __p0) {
   4202   poly8x8_t __ret;
   4203   __ret = (poly8x8_t)(__p0);
   4204   return __ret;
   4205 }
   4206 #endif
   4207 
   4208 #ifdef __LITTLE_ENDIAN__
   4209 __ai poly16x4_t vcreate_p16(uint64_t __p0) {
   4210   poly16x4_t __ret;
   4211   __ret = (poly16x4_t)(__p0);
   4212   return __ret;
   4213 }
   4214 #else
   4215 __ai poly16x4_t vcreate_p16(uint64_t __p0) {
   4216   poly16x4_t __ret;
   4217   __ret = (poly16x4_t)(__p0);
   4218   return __ret;
   4219 }
   4220 #endif
   4221 
   4222 #ifdef __LITTLE_ENDIAN__
   4223 __ai uint8x8_t vcreate_u8(uint64_t __p0) {
   4224   uint8x8_t __ret;
   4225   __ret = (uint8x8_t)(__p0);
   4226   return __ret;
   4227 }
   4228 #else
   4229 __ai uint8x8_t vcreate_u8(uint64_t __p0) {
   4230   uint8x8_t __ret;
   4231   __ret = (uint8x8_t)(__p0);
   4232   return __ret;
   4233 }
   4234 #endif
   4235 
   4236 #ifdef __LITTLE_ENDIAN__
   4237 __ai uint32x2_t vcreate_u32(uint64_t __p0) {
   4238   uint32x2_t __ret;
   4239   __ret = (uint32x2_t)(__p0);
   4240   return __ret;
   4241 }
   4242 #else
   4243 __ai uint32x2_t vcreate_u32(uint64_t __p0) {
   4244   uint32x2_t __ret;
   4245   __ret = (uint32x2_t)(__p0);
   4246   return __ret;
   4247 }
   4248 #endif
   4249 
   4250 #ifdef __LITTLE_ENDIAN__
   4251 __ai uint64x1_t vcreate_u64(uint64_t __p0) {
   4252   uint64x1_t __ret;
   4253   __ret = (uint64x1_t)(__p0);
   4254   return __ret;
   4255 }
   4256 #else
   4257 __ai uint64x1_t vcreate_u64(uint64_t __p0) {
   4258   uint64x1_t __ret;
   4259   __ret = (uint64x1_t)(__p0);
   4260   return __ret;
   4261 }
   4262 #endif
   4263 
   4264 #ifdef __LITTLE_ENDIAN__
   4265 __ai uint16x4_t vcreate_u16(uint64_t __p0) {
   4266   uint16x4_t __ret;
   4267   __ret = (uint16x4_t)(__p0);
   4268   return __ret;
   4269 }
   4270 #else
   4271 __ai uint16x4_t vcreate_u16(uint64_t __p0) {
   4272   uint16x4_t __ret;
   4273   __ret = (uint16x4_t)(__p0);
   4274   return __ret;
   4275 }
   4276 #endif
   4277 
   4278 #ifdef __LITTLE_ENDIAN__
   4279 __ai int8x8_t vcreate_s8(uint64_t __p0) {
   4280   int8x8_t __ret;
   4281   __ret = (int8x8_t)(__p0);
   4282   return __ret;
   4283 }
   4284 #else
   4285 __ai int8x8_t vcreate_s8(uint64_t __p0) {
   4286   int8x8_t __ret;
   4287   __ret = (int8x8_t)(__p0);
   4288   return __ret;
   4289 }
   4290 #endif
   4291 
   4292 #ifdef __LITTLE_ENDIAN__
   4293 __ai float32x2_t vcreate_f32(uint64_t __p0) {
   4294   float32x2_t __ret;
   4295   __ret = (float32x2_t)(__p0);
   4296   return __ret;
   4297 }
   4298 #else
   4299 __ai float32x2_t vcreate_f32(uint64_t __p0) {
   4300   float32x2_t __ret;
   4301   __ret = (float32x2_t)(__p0);
   4302   return __ret;
   4303 }
   4304 #endif
   4305 
   4306 #ifdef __LITTLE_ENDIAN__
   4307 __ai float16x4_t vcreate_f16(uint64_t __p0) {
   4308   float16x4_t __ret;
   4309   __ret = (float16x4_t)(__p0);
   4310   return __ret;
   4311 }
   4312 #else
   4313 __ai float16x4_t vcreate_f16(uint64_t __p0) {
   4314   float16x4_t __ret;
   4315   __ret = (float16x4_t)(__p0);
   4316   return __ret;
   4317 }
   4318 #endif
   4319 
   4320 #ifdef __LITTLE_ENDIAN__
   4321 __ai int32x2_t vcreate_s32(uint64_t __p0) {
   4322   int32x2_t __ret;
   4323   __ret = (int32x2_t)(__p0);
   4324   return __ret;
   4325 }
   4326 #else
   4327 __ai int32x2_t vcreate_s32(uint64_t __p0) {
   4328   int32x2_t __ret;
   4329   __ret = (int32x2_t)(__p0);
   4330   return __ret;
   4331 }
   4332 #endif
   4333 
   4334 #ifdef __LITTLE_ENDIAN__
   4335 __ai int64x1_t vcreate_s64(uint64_t __p0) {
   4336   int64x1_t __ret;
   4337   __ret = (int64x1_t)(__p0);
   4338   return __ret;
   4339 }
   4340 #else
   4341 __ai int64x1_t vcreate_s64(uint64_t __p0) {
   4342   int64x1_t __ret;
   4343   __ret = (int64x1_t)(__p0);
   4344   return __ret;
   4345 }
   4346 #endif
   4347 
   4348 #ifdef __LITTLE_ENDIAN__
   4349 __ai int16x4_t vcreate_s16(uint64_t __p0) {
   4350   int16x4_t __ret;
   4351   __ret = (int16x4_t)(__p0);
   4352   return __ret;
   4353 }
   4354 #else
   4355 __ai int16x4_t vcreate_s16(uint64_t __p0) {
   4356   int16x4_t __ret;
   4357   __ret = (int16x4_t)(__p0);
   4358   return __ret;
   4359 }
   4360 #endif
   4361 
   4362 #ifdef __LITTLE_ENDIAN__
   4363 __ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
   4364   float32x4_t __ret;
   4365   __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 50);
   4366   return __ret;
   4367 }
   4368 #else
   4369 __ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
   4370   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   4371   float32x4_t __ret;
   4372   __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 50);
   4373   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   4374   return __ret;
   4375 }
   4376 #endif
   4377 
   4378 #ifdef __LITTLE_ENDIAN__
   4379 __ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
   4380   float32x4_t __ret;
   4381   __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 34);
   4382   return __ret;
   4383 }
   4384 #else
   4385 __ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
   4386   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   4387   float32x4_t __ret;
   4388   __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 34);
   4389   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   4390   return __ret;
   4391 }
   4392 #endif
   4393 
   4394 #ifdef __LITTLE_ENDIAN__
   4395 __ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
   4396   float32x2_t __ret;
   4397   __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 18);
   4398   return __ret;
   4399 }
   4400 #else
   4401 __ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
   4402   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   4403   float32x2_t __ret;
   4404   __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 18);
   4405   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   4406   return __ret;
   4407 }
   4408 #endif
   4409 
   4410 #ifdef __LITTLE_ENDIAN__
   4411 __ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
   4412   float32x2_t __ret;
   4413   __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 2);
   4414   return __ret;
   4415 }
   4416 #else
   4417 __ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
   4418   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   4419   float32x2_t __ret;
   4420   __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 2);
   4421   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   4422   return __ret;
   4423 }
   4424 #endif
   4425 
   4426 #ifdef __LITTLE_ENDIAN__
   4427 #define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
   4428   uint32x4_t __s0 = __p0; \
   4429   float32x4_t __ret; \
   4430   __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 50); \
   4431   __ret; \
   4432 })
   4433 #else
   4434 #define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
   4435   uint32x4_t __s0 = __p0; \
   4436   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   4437   float32x4_t __ret; \
   4438   __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 50); \
   4439   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   4440   __ret; \
   4441 })
   4442 #endif
   4443 
   4444 #ifdef __LITTLE_ENDIAN__
   4445 #define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
   4446   int32x4_t __s0 = __p0; \
   4447   float32x4_t __ret; \
   4448   __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 34); \
   4449   __ret; \
   4450 })
   4451 #else
   4452 #define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
   4453   int32x4_t __s0 = __p0; \
   4454   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   4455   float32x4_t __ret; \
   4456   __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 34); \
   4457   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   4458   __ret; \
   4459 })
   4460 #endif
   4461 
   4462 #ifdef __LITTLE_ENDIAN__
   4463 #define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
   4464   uint32x2_t __s0 = __p0; \
   4465   float32x2_t __ret; \
   4466   __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 18); \
   4467   __ret; \
   4468 })
   4469 #else
   4470 #define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
   4471   uint32x2_t __s0 = __p0; \
   4472   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4473   float32x2_t __ret; \
   4474   __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 18); \
   4475   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   4476   __ret; \
   4477 })
   4478 #endif
   4479 
   4480 #ifdef __LITTLE_ENDIAN__
   4481 #define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
   4482   int32x2_t __s0 = __p0; \
   4483   float32x2_t __ret; \
   4484   __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 2); \
   4485   __ret; \
   4486 })
   4487 #else
   4488 #define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
   4489   int32x2_t __s0 = __p0; \
   4490   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4491   float32x2_t __ret; \
   4492   __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 2); \
   4493   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   4494   __ret; \
   4495 })
   4496 #endif
   4497 
   4498 #ifdef __LITTLE_ENDIAN__
   4499 #define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
   4500   float32x4_t __s0 = __p0; \
   4501   int32x4_t __ret; \
   4502   __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__s0, __p1, 34); \
   4503   __ret; \
   4504 })
   4505 #else
   4506 #define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
   4507   float32x4_t __s0 = __p0; \
   4508   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   4509   int32x4_t __ret; \
   4510   __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__rev0, __p1, 34); \
   4511   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   4512   __ret; \
   4513 })
   4514 #endif
   4515 
   4516 #ifdef __LITTLE_ENDIAN__
   4517 #define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
   4518   float32x2_t __s0 = __p0; \
   4519   int32x2_t __ret; \
   4520   __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__s0, __p1, 2); \
   4521   __ret; \
   4522 })
   4523 #else
   4524 #define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
   4525   float32x2_t __s0 = __p0; \
   4526   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4527   int32x2_t __ret; \
   4528   __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__rev0, __p1, 2); \
   4529   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   4530   __ret; \
   4531 })
   4532 #endif
   4533 
   4534 #ifdef __LITTLE_ENDIAN__
   4535 #define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
   4536   float32x4_t __s0 = __p0; \
   4537   uint32x4_t __ret; \
   4538   __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__s0, __p1, 50); \
   4539   __ret; \
   4540 })
   4541 #else
   4542 #define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
   4543   float32x4_t __s0 = __p0; \
   4544   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   4545   uint32x4_t __ret; \
   4546   __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__rev0, __p1, 50); \
   4547   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   4548   __ret; \
   4549 })
   4550 #endif
   4551 
   4552 #ifdef __LITTLE_ENDIAN__
   4553 #define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
   4554   float32x2_t __s0 = __p0; \
   4555   uint32x2_t __ret; \
   4556   __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__s0, __p1, 18); \
   4557   __ret; \
   4558 })
   4559 #else
   4560 #define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
   4561   float32x2_t __s0 = __p0; \
   4562   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4563   uint32x2_t __ret; \
   4564   __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__rev0, __p1, 18); \
   4565   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   4566   __ret; \
   4567 })
   4568 #endif
   4569 
   4570 #ifdef __LITTLE_ENDIAN__
   4571 __ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
   4572   int32x4_t __ret;
   4573   __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__p0, 34);
   4574   return __ret;
   4575 }
   4576 #else
   4577 __ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
   4578   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   4579   int32x4_t __ret;
   4580   __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__rev0, 34);
   4581   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   4582   return __ret;
   4583 }
   4584 #endif
   4585 
   4586 #ifdef __LITTLE_ENDIAN__
   4587 __ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
   4588   int32x2_t __ret;
   4589   __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__p0, 2);
   4590   return __ret;
   4591 }
   4592 #else
   4593 __ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
   4594   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   4595   int32x2_t __ret;
   4596   __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__rev0, 2);
   4597   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   4598   return __ret;
   4599 }
   4600 #endif
   4601 
   4602 #ifdef __LITTLE_ENDIAN__
   4603 __ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
   4604   uint32x4_t __ret;
   4605   __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__p0, 50);
   4606   return __ret;
   4607 }
   4608 #else
   4609 __ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
   4610   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   4611   uint32x4_t __ret;
   4612   __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__rev0, 50);
   4613   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   4614   return __ret;
   4615 }
   4616 #endif
   4617 
   4618 #ifdef __LITTLE_ENDIAN__
   4619 __ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
   4620   uint32x2_t __ret;
   4621   __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__p0, 18);
   4622   return __ret;
   4623 }
   4624 #else
   4625 __ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
   4626   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   4627   uint32x2_t __ret;
   4628   __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__rev0, 18);
   4629   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   4630   return __ret;
   4631 }
   4632 #endif
   4633 
   4634 #ifdef __LITTLE_ENDIAN__
   4635 #define vdup_lane_p8(__p0, __p1) __extension__ ({ \
   4636   poly8x8_t __s0 = __p0; \
   4637   poly8x8_t __ret; \
   4638   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4639   __ret; \
   4640 })
   4641 #else
   4642 #define vdup_lane_p8(__p0, __p1) __extension__ ({ \
   4643   poly8x8_t __s0 = __p0; \
   4644   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   4645   poly8x8_t __ret; \
   4646   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4647   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   4648   __ret; \
   4649 })
   4650 #endif
   4651 
   4652 #ifdef __LITTLE_ENDIAN__
   4653 #define vdup_lane_p16(__p0, __p1) __extension__ ({ \
   4654   poly16x4_t __s0 = __p0; \
   4655   poly16x4_t __ret; \
   4656   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   4657   __ret; \
   4658 })
   4659 #else
   4660 #define vdup_lane_p16(__p0, __p1) __extension__ ({ \
   4661   poly16x4_t __s0 = __p0; \
   4662   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   4663   poly16x4_t __ret; \
   4664   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   4665   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   4666   __ret; \
   4667 })
   4668 #endif
   4669 
   4670 #ifdef __LITTLE_ENDIAN__
   4671 #define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
   4672   poly8x8_t __s0 = __p0; \
   4673   poly8x16_t __ret; \
   4674   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4675   __ret; \
   4676 })
   4677 #else
   4678 #define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
   4679   poly8x8_t __s0 = __p0; \
   4680   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   4681   poly8x16_t __ret; \
   4682   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4683   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   4684   __ret; \
   4685 })
   4686 #endif
   4687 
   4688 #ifdef __LITTLE_ENDIAN__
   4689 #define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
   4690   poly16x4_t __s0 = __p0; \
   4691   poly16x8_t __ret; \
   4692   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4693   __ret; \
   4694 })
   4695 #else
   4696 #define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
   4697   poly16x4_t __s0 = __p0; \
   4698   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   4699   poly16x8_t __ret; \
   4700   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4701   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   4702   __ret; \
   4703 })
   4704 #endif
   4705 
   4706 #ifdef __LITTLE_ENDIAN__
   4707 #define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
   4708   uint8x8_t __s0 = __p0; \
   4709   uint8x16_t __ret; \
   4710   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4711   __ret; \
   4712 })
   4713 #else
   4714 #define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
   4715   uint8x8_t __s0 = __p0; \
   4716   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   4717   uint8x16_t __ret; \
   4718   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4719   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   4720   __ret; \
   4721 })
   4722 #endif
   4723 
   4724 #ifdef __LITTLE_ENDIAN__
   4725 #define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
   4726   uint32x2_t __s0 = __p0; \
   4727   uint32x4_t __ret; \
   4728   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   4729   __ret; \
   4730 })
   4731 #else
   4732 #define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
   4733   uint32x2_t __s0 = __p0; \
   4734   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4735   uint32x4_t __ret; \
   4736   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   4737   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   4738   __ret; \
   4739 })
   4740 #endif
   4741 
   4742 #ifdef __LITTLE_ENDIAN__
   4743 #define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
   4744   uint64x1_t __s0 = __p0; \
   4745   uint64x2_t __ret; \
   4746   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   4747   __ret; \
   4748 })
   4749 #else
   4750 #define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
   4751   uint64x1_t __s0 = __p0; \
   4752   uint64x2_t __ret; \
   4753   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   4754   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   4755   __ret; \
   4756 })
   4757 #endif
   4758 
   4759 #ifdef __LITTLE_ENDIAN__
   4760 #define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
   4761   uint16x4_t __s0 = __p0; \
   4762   uint16x8_t __ret; \
   4763   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4764   __ret; \
   4765 })
   4766 #else
   4767 #define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
   4768   uint16x4_t __s0 = __p0; \
   4769   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   4770   uint16x8_t __ret; \
   4771   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4772   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   4773   __ret; \
   4774 })
   4775 #endif
   4776 
   4777 #ifdef __LITTLE_ENDIAN__
   4778 #define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
   4779   int8x8_t __s0 = __p0; \
   4780   int8x16_t __ret; \
   4781   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4782   __ret; \
   4783 })
   4784 #else
   4785 #define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
   4786   int8x8_t __s0 = __p0; \
   4787   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   4788   int8x16_t __ret; \
   4789   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4790   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   4791   __ret; \
   4792 })
   4793 #endif
   4794 
   4795 #ifdef __LITTLE_ENDIAN__
   4796 #define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
   4797   float32x2_t __s0 = __p0; \
   4798   float32x4_t __ret; \
   4799   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   4800   __ret; \
   4801 })
   4802 #else
   4803 #define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
   4804   float32x2_t __s0 = __p0; \
   4805   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4806   float32x4_t __ret; \
   4807   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   4808   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   4809   __ret; \
   4810 })
   4811 #endif
   4812 
   4813 #ifdef __LITTLE_ENDIAN__
   4814 #define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
   4815   int32x2_t __s0 = __p0; \
   4816   int32x4_t __ret; \
   4817   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   4818   __ret; \
   4819 })
   4820 #else
   4821 #define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
   4822   int32x2_t __s0 = __p0; \
   4823   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4824   int32x4_t __ret; \
   4825   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   4826   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   4827   __ret; \
   4828 })
   4829 #endif
   4830 
   4831 #ifdef __LITTLE_ENDIAN__
   4832 #define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
   4833   int64x1_t __s0 = __p0; \
   4834   int64x2_t __ret; \
   4835   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   4836   __ret; \
   4837 })
   4838 #else
   4839 #define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
   4840   int64x1_t __s0 = __p0; \
   4841   int64x2_t __ret; \
   4842   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   4843   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   4844   __ret; \
   4845 })
   4846 #endif
   4847 
   4848 #ifdef __LITTLE_ENDIAN__
   4849 #define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
   4850   int16x4_t __s0 = __p0; \
   4851   int16x8_t __ret; \
   4852   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4853   __ret; \
   4854 })
   4855 #else
   4856 #define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
   4857   int16x4_t __s0 = __p0; \
   4858   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   4859   int16x8_t __ret; \
   4860   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4861   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   4862   __ret; \
   4863 })
   4864 #endif
   4865 
   4866 #ifdef __LITTLE_ENDIAN__
   4867 #define vdup_lane_u8(__p0, __p1) __extension__ ({ \
   4868   uint8x8_t __s0 = __p0; \
   4869   uint8x8_t __ret; \
   4870   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4871   __ret; \
   4872 })
   4873 #else
   4874 #define vdup_lane_u8(__p0, __p1) __extension__ ({ \
   4875   uint8x8_t __s0 = __p0; \
   4876   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   4877   uint8x8_t __ret; \
   4878   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4879   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   4880   __ret; \
   4881 })
   4882 #endif
   4883 
   4884 #ifdef __LITTLE_ENDIAN__
   4885 #define vdup_lane_u32(__p0, __p1) __extension__ ({ \
   4886   uint32x2_t __s0 = __p0; \
   4887   uint32x2_t __ret; \
   4888   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   4889   __ret; \
   4890 })
   4891 #else
   4892 #define vdup_lane_u32(__p0, __p1) __extension__ ({ \
   4893   uint32x2_t __s0 = __p0; \
   4894   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4895   uint32x2_t __ret; \
   4896   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   4897   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   4898   __ret; \
   4899 })
   4900 #endif
   4901 
   4902 #ifdef __LITTLE_ENDIAN__
   4903 #define vdup_lane_u64(__p0, __p1) __extension__ ({ \
   4904   uint64x1_t __s0 = __p0; \
   4905   uint64x1_t __ret; \
   4906   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   4907   __ret; \
   4908 })
   4909 #else
   4910 #define vdup_lane_u64(__p0, __p1) __extension__ ({ \
   4911   uint64x1_t __s0 = __p0; \
   4912   uint64x1_t __ret; \
   4913   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   4914   __ret; \
   4915 })
   4916 #endif
   4917 
   4918 #ifdef __LITTLE_ENDIAN__
   4919 #define vdup_lane_u16(__p0, __p1) __extension__ ({ \
   4920   uint16x4_t __s0 = __p0; \
   4921   uint16x4_t __ret; \
   4922   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   4923   __ret; \
   4924 })
   4925 #else
   4926 #define vdup_lane_u16(__p0, __p1) __extension__ ({ \
   4927   uint16x4_t __s0 = __p0; \
   4928   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   4929   uint16x4_t __ret; \
   4930   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   4931   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   4932   __ret; \
   4933 })
   4934 #endif
   4935 
   4936 #ifdef __LITTLE_ENDIAN__
   4937 #define vdup_lane_s8(__p0, __p1) __extension__ ({ \
   4938   int8x8_t __s0 = __p0; \
   4939   int8x8_t __ret; \
   4940   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4941   __ret; \
   4942 })
   4943 #else
   4944 #define vdup_lane_s8(__p0, __p1) __extension__ ({ \
   4945   int8x8_t __s0 = __p0; \
   4946   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   4947   int8x8_t __ret; \
   4948   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   4949   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   4950   __ret; \
   4951 })
   4952 #endif
   4953 
   4954 #ifdef __LITTLE_ENDIAN__
   4955 #define vdup_lane_f32(__p0, __p1) __extension__ ({ \
   4956   float32x2_t __s0 = __p0; \
   4957   float32x2_t __ret; \
   4958   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   4959   __ret; \
   4960 })
   4961 #else
   4962 #define vdup_lane_f32(__p0, __p1) __extension__ ({ \
   4963   float32x2_t __s0 = __p0; \
   4964   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4965   float32x2_t __ret; \
   4966   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   4967   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   4968   __ret; \
   4969 })
   4970 #endif
   4971 
   4972 #ifdef __LITTLE_ENDIAN__
   4973 #define vdup_lane_s32(__p0, __p1) __extension__ ({ \
   4974   int32x2_t __s0 = __p0; \
   4975   int32x2_t __ret; \
   4976   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   4977   __ret; \
   4978 })
   4979 #else
   4980 #define vdup_lane_s32(__p0, __p1) __extension__ ({ \
   4981   int32x2_t __s0 = __p0; \
   4982   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   4983   int32x2_t __ret; \
   4984   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   4985   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   4986   __ret; \
   4987 })
   4988 #endif
   4989 
   4990 #ifdef __LITTLE_ENDIAN__
   4991 #define vdup_lane_s64(__p0, __p1) __extension__ ({ \
   4992   int64x1_t __s0 = __p0; \
   4993   int64x1_t __ret; \
   4994   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   4995   __ret; \
   4996 })
   4997 #else
   4998 #define vdup_lane_s64(__p0, __p1) __extension__ ({ \
   4999   int64x1_t __s0 = __p0; \
   5000   int64x1_t __ret; \
   5001   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   5002   __ret; \
   5003 })
   5004 #endif
   5005 
   5006 #ifdef __LITTLE_ENDIAN__
   5007 #define vdup_lane_s16(__p0, __p1) __extension__ ({ \
   5008   int16x4_t __s0 = __p0; \
   5009   int16x4_t __ret; \
   5010   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   5011   __ret; \
   5012 })
   5013 #else
   5014 #define vdup_lane_s16(__p0, __p1) __extension__ ({ \
   5015   int16x4_t __s0 = __p0; \
   5016   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   5017   int16x4_t __ret; \
   5018   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   5019   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   5020   __ret; \
   5021 })
   5022 #endif
   5023 
   5024 #ifdef __LITTLE_ENDIAN__
   5025 __ai poly8x8_t vdup_n_p8(poly8_t __p0) {
   5026   poly8x8_t __ret;
   5027   __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5028   return __ret;
   5029 }
   5030 #else
   5031 __ai poly8x8_t vdup_n_p8(poly8_t __p0) {
   5032   poly8x8_t __ret;
   5033   __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5034   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5035   return __ret;
   5036 }
   5037 #endif
   5038 
   5039 #ifdef __LITTLE_ENDIAN__
   5040 __ai poly16x4_t vdup_n_p16(poly16_t __p0) {
   5041   poly16x4_t __ret;
   5042   __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
   5043   return __ret;
   5044 }
   5045 #else
   5046 __ai poly16x4_t vdup_n_p16(poly16_t __p0) {
   5047   poly16x4_t __ret;
   5048   __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
   5049   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5050   return __ret;
   5051 }
   5052 #endif
   5053 
   5054 #ifdef __LITTLE_ENDIAN__
   5055 __ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
   5056   poly8x16_t __ret;
   5057   __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5058   return __ret;
   5059 }
   5060 #else
   5061 __ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
   5062   poly8x16_t __ret;
   5063   __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5064   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   5065   return __ret;
   5066 }
   5067 #endif
   5068 
   5069 #ifdef __LITTLE_ENDIAN__
   5070 __ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
   5071   poly16x8_t __ret;
   5072   __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5073   return __ret;
   5074 }
   5075 #else
   5076 __ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
   5077   poly16x8_t __ret;
   5078   __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5079   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5080   return __ret;
   5081 }
   5082 #endif
   5083 
   5084 #ifdef __LITTLE_ENDIAN__
   5085 __ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
   5086   uint8x16_t __ret;
   5087   __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5088   return __ret;
   5089 }
   5090 #else
   5091 __ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
   5092   uint8x16_t __ret;
   5093   __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5094   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   5095   return __ret;
   5096 }
   5097 #endif
   5098 
   5099 #ifdef __LITTLE_ENDIAN__
   5100 __ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
   5101   uint32x4_t __ret;
   5102   __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
   5103   return __ret;
   5104 }
   5105 #else
   5106 __ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
   5107   uint32x4_t __ret;
   5108   __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
   5109   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5110   return __ret;
   5111 }
   5112 #endif
   5113 
   5114 #ifdef __LITTLE_ENDIAN__
   5115 __ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
   5116   uint64x2_t __ret;
   5117   __ret = (uint64x2_t) {__p0, __p0};
   5118   return __ret;
   5119 }
   5120 #else
   5121 __ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
   5122   uint64x2_t __ret;
   5123   __ret = (uint64x2_t) {__p0, __p0};
   5124   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   5125   return __ret;
   5126 }
   5127 #endif
   5128 
   5129 #ifdef __LITTLE_ENDIAN__
   5130 __ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
   5131   uint16x8_t __ret;
   5132   __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5133   return __ret;
   5134 }
   5135 #else
   5136 __ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
   5137   uint16x8_t __ret;
   5138   __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5139   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5140   return __ret;
   5141 }
   5142 #endif
   5143 
   5144 #ifdef __LITTLE_ENDIAN__
   5145 __ai int8x16_t vdupq_n_s8(int8_t __p0) {
   5146   int8x16_t __ret;
   5147   __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5148   return __ret;
   5149 }
   5150 #else
   5151 __ai int8x16_t vdupq_n_s8(int8_t __p0) {
   5152   int8x16_t __ret;
   5153   __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5154   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   5155   return __ret;
   5156 }
   5157 #endif
   5158 
   5159 #ifdef __LITTLE_ENDIAN__
   5160 __ai float32x4_t vdupq_n_f32(float32_t __p0) {
   5161   float32x4_t __ret;
   5162   __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
   5163   return __ret;
   5164 }
   5165 #else
   5166 __ai float32x4_t vdupq_n_f32(float32_t __p0) {
   5167   float32x4_t __ret;
   5168   __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
   5169   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5170   return __ret;
   5171 }
   5172 #endif
   5173 
   5174 #ifdef __LITTLE_ENDIAN__
   5175 #define vdupq_n_f16(__p0) __extension__ ({ \
   5176   float16_t __s0 = __p0; \
   5177   float16x8_t __ret; \
   5178   __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
   5179   __ret; \
   5180 })
   5181 #else
   5182 #define vdupq_n_f16(__p0) __extension__ ({ \
   5183   float16_t __s0 = __p0; \
   5184   float16x8_t __ret; \
   5185   __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
   5186   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   5187   __ret; \
   5188 })
   5189 #endif
   5190 
   5191 #ifdef __LITTLE_ENDIAN__
   5192 __ai int32x4_t vdupq_n_s32(int32_t __p0) {
   5193   int32x4_t __ret;
   5194   __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
   5195   return __ret;
   5196 }
   5197 #else
   5198 __ai int32x4_t vdupq_n_s32(int32_t __p0) {
   5199   int32x4_t __ret;
   5200   __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
   5201   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5202   return __ret;
   5203 }
   5204 #endif
   5205 
   5206 #ifdef __LITTLE_ENDIAN__
   5207 __ai int64x2_t vdupq_n_s64(int64_t __p0) {
   5208   int64x2_t __ret;
   5209   __ret = (int64x2_t) {__p0, __p0};
   5210   return __ret;
   5211 }
   5212 #else
   5213 __ai int64x2_t vdupq_n_s64(int64_t __p0) {
   5214   int64x2_t __ret;
   5215   __ret = (int64x2_t) {__p0, __p0};
   5216   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   5217   return __ret;
   5218 }
   5219 #endif
   5220 
   5221 #ifdef __LITTLE_ENDIAN__
   5222 __ai int16x8_t vdupq_n_s16(int16_t __p0) {
   5223   int16x8_t __ret;
   5224   __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5225   return __ret;
   5226 }
   5227 #else
   5228 __ai int16x8_t vdupq_n_s16(int16_t __p0) {
   5229   int16x8_t __ret;
   5230   __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5231   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5232   return __ret;
   5233 }
   5234 #endif
   5235 
   5236 #ifdef __LITTLE_ENDIAN__
   5237 __ai uint8x8_t vdup_n_u8(uint8_t __p0) {
   5238   uint8x8_t __ret;
   5239   __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5240   return __ret;
   5241 }
   5242 #else
   5243 __ai uint8x8_t vdup_n_u8(uint8_t __p0) {
   5244   uint8x8_t __ret;
   5245   __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5246   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5247   return __ret;
   5248 }
   5249 #endif
   5250 
   5251 #ifdef __LITTLE_ENDIAN__
   5252 __ai uint32x2_t vdup_n_u32(uint32_t __p0) {
   5253   uint32x2_t __ret;
   5254   __ret = (uint32x2_t) {__p0, __p0};
   5255   return __ret;
   5256 }
   5257 #else
   5258 __ai uint32x2_t vdup_n_u32(uint32_t __p0) {
   5259   uint32x2_t __ret;
   5260   __ret = (uint32x2_t) {__p0, __p0};
   5261   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   5262   return __ret;
   5263 }
   5264 #endif
   5265 
   5266 #ifdef __LITTLE_ENDIAN__
   5267 __ai uint64x1_t vdup_n_u64(uint64_t __p0) {
   5268   uint64x1_t __ret;
   5269   __ret = (uint64x1_t) {__p0};
   5270   return __ret;
   5271 }
   5272 #else
   5273 __ai uint64x1_t vdup_n_u64(uint64_t __p0) {
   5274   uint64x1_t __ret;
   5275   __ret = (uint64x1_t) {__p0};
   5276   return __ret;
   5277 }
   5278 #endif
   5279 
   5280 #ifdef __LITTLE_ENDIAN__
   5281 __ai uint16x4_t vdup_n_u16(uint16_t __p0) {
   5282   uint16x4_t __ret;
   5283   __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
   5284   return __ret;
   5285 }
   5286 #else
   5287 __ai uint16x4_t vdup_n_u16(uint16_t __p0) {
   5288   uint16x4_t __ret;
   5289   __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
   5290   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5291   return __ret;
   5292 }
   5293 #endif
   5294 
   5295 #ifdef __LITTLE_ENDIAN__
   5296 __ai int8x8_t vdup_n_s8(int8_t __p0) {
   5297   int8x8_t __ret;
   5298   __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5299   return __ret;
   5300 }
   5301 #else
   5302 __ai int8x8_t vdup_n_s8(int8_t __p0) {
   5303   int8x8_t __ret;
   5304   __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   5305   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5306   return __ret;
   5307 }
   5308 #endif
   5309 
   5310 #ifdef __LITTLE_ENDIAN__
   5311 __ai float32x2_t vdup_n_f32(float32_t __p0) {
   5312   float32x2_t __ret;
   5313   __ret = (float32x2_t) {__p0, __p0};
   5314   return __ret;
   5315 }
   5316 #else
   5317 __ai float32x2_t vdup_n_f32(float32_t __p0) {
   5318   float32x2_t __ret;
   5319   __ret = (float32x2_t) {__p0, __p0};
   5320   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   5321   return __ret;
   5322 }
   5323 #endif
   5324 
   5325 #ifdef __LITTLE_ENDIAN__
   5326 #define vdup_n_f16(__p0) __extension__ ({ \
   5327   float16_t __s0 = __p0; \
   5328   float16x4_t __ret; \
   5329   __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
   5330   __ret; \
   5331 })
   5332 #else
   5333 #define vdup_n_f16(__p0) __extension__ ({ \
   5334   float16_t __s0 = __p0; \
   5335   float16x4_t __ret; \
   5336   __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
   5337   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   5338   __ret; \
   5339 })
   5340 #endif
   5341 
   5342 #ifdef __LITTLE_ENDIAN__
   5343 __ai int32x2_t vdup_n_s32(int32_t __p0) {
   5344   int32x2_t __ret;
   5345   __ret = (int32x2_t) {__p0, __p0};
   5346   return __ret;
   5347 }
   5348 #else
   5349 __ai int32x2_t vdup_n_s32(int32_t __p0) {
   5350   int32x2_t __ret;
   5351   __ret = (int32x2_t) {__p0, __p0};
   5352   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   5353   return __ret;
   5354 }
   5355 #endif
   5356 
   5357 #ifdef __LITTLE_ENDIAN__
   5358 __ai int64x1_t vdup_n_s64(int64_t __p0) {
   5359   int64x1_t __ret;
   5360   __ret = (int64x1_t) {__p0};
   5361   return __ret;
   5362 }
   5363 #else
   5364 __ai int64x1_t vdup_n_s64(int64_t __p0) {
   5365   int64x1_t __ret;
   5366   __ret = (int64x1_t) {__p0};
   5367   return __ret;
   5368 }
   5369 #endif
   5370 
   5371 #ifdef __LITTLE_ENDIAN__
   5372 __ai int16x4_t vdup_n_s16(int16_t __p0) {
   5373   int16x4_t __ret;
   5374   __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
   5375   return __ret;
   5376 }
   5377 #else
   5378 __ai int16x4_t vdup_n_s16(int16_t __p0) {
   5379   int16x4_t __ret;
   5380   __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
   5381   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5382   return __ret;
   5383 }
   5384 #endif
   5385 
   5386 #ifdef __LITTLE_ENDIAN__
   5387 __ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   5388   uint8x16_t __ret;
   5389   __ret = __p0 ^ __p1;
   5390   return __ret;
   5391 }
   5392 #else
   5393 __ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   5394   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   5395   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   5396   uint8x16_t __ret;
   5397   __ret = __rev0 ^ __rev1;
   5398   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   5399   return __ret;
   5400 }
   5401 #endif
   5402 
   5403 #ifdef __LITTLE_ENDIAN__
   5404 __ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   5405   uint32x4_t __ret;
   5406   __ret = __p0 ^ __p1;
   5407   return __ret;
   5408 }
   5409 #else
   5410 __ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   5411   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   5412   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   5413   uint32x4_t __ret;
   5414   __ret = __rev0 ^ __rev1;
   5415   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5416   return __ret;
   5417 }
   5418 #endif
   5419 
   5420 #ifdef __LITTLE_ENDIAN__
   5421 __ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   5422   uint64x2_t __ret;
   5423   __ret = __p0 ^ __p1;
   5424   return __ret;
   5425 }
   5426 #else
   5427 __ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   5428   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   5429   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   5430   uint64x2_t __ret;
   5431   __ret = __rev0 ^ __rev1;
   5432   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   5433   return __ret;
   5434 }
   5435 #endif
   5436 
   5437 #ifdef __LITTLE_ENDIAN__
   5438 __ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   5439   uint16x8_t __ret;
   5440   __ret = __p0 ^ __p1;
   5441   return __ret;
   5442 }
   5443 #else
   5444 __ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   5445   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   5446   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   5447   uint16x8_t __ret;
   5448   __ret = __rev0 ^ __rev1;
   5449   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5450   return __ret;
   5451 }
   5452 #endif
   5453 
   5454 #ifdef __LITTLE_ENDIAN__
   5455 __ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
   5456   int8x16_t __ret;
   5457   __ret = __p0 ^ __p1;
   5458   return __ret;
   5459 }
   5460 #else
   5461 __ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
   5462   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   5463   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   5464   int8x16_t __ret;
   5465   __ret = __rev0 ^ __rev1;
   5466   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   5467   return __ret;
   5468 }
   5469 #endif
   5470 
   5471 #ifdef __LITTLE_ENDIAN__
   5472 __ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
   5473   int32x4_t __ret;
   5474   __ret = __p0 ^ __p1;
   5475   return __ret;
   5476 }
   5477 #else
   5478 __ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
   5479   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   5480   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   5481   int32x4_t __ret;
   5482   __ret = __rev0 ^ __rev1;
   5483   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5484   return __ret;
   5485 }
   5486 #endif
   5487 
   5488 #ifdef __LITTLE_ENDIAN__
   5489 __ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
   5490   int64x2_t __ret;
   5491   __ret = __p0 ^ __p1;
   5492   return __ret;
   5493 }
   5494 #else
   5495 __ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
   5496   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   5497   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   5498   int64x2_t __ret;
   5499   __ret = __rev0 ^ __rev1;
   5500   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   5501   return __ret;
   5502 }
   5503 #endif
   5504 
   5505 #ifdef __LITTLE_ENDIAN__
   5506 __ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
   5507   int16x8_t __ret;
   5508   __ret = __p0 ^ __p1;
   5509   return __ret;
   5510 }
   5511 #else
   5512 __ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
   5513   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   5514   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   5515   int16x8_t __ret;
   5516   __ret = __rev0 ^ __rev1;
   5517   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5518   return __ret;
   5519 }
   5520 #endif
   5521 
   5522 #ifdef __LITTLE_ENDIAN__
   5523 __ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
   5524   uint8x8_t __ret;
   5525   __ret = __p0 ^ __p1;
   5526   return __ret;
   5527 }
   5528 #else
   5529 __ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
   5530   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   5531   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   5532   uint8x8_t __ret;
   5533   __ret = __rev0 ^ __rev1;
   5534   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5535   return __ret;
   5536 }
   5537 #endif
   5538 
   5539 #ifdef __LITTLE_ENDIAN__
   5540 __ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
   5541   uint32x2_t __ret;
   5542   __ret = __p0 ^ __p1;
   5543   return __ret;
   5544 }
   5545 #else
   5546 __ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
   5547   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   5548   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   5549   uint32x2_t __ret;
   5550   __ret = __rev0 ^ __rev1;
   5551   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   5552   return __ret;
   5553 }
   5554 #endif
   5555 
   5556 #ifdef __LITTLE_ENDIAN__
   5557 __ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
   5558   uint64x1_t __ret;
   5559   __ret = __p0 ^ __p1;
   5560   return __ret;
   5561 }
   5562 #else
   5563 __ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
   5564   uint64x1_t __ret;
   5565   __ret = __p0 ^ __p1;
   5566   return __ret;
   5567 }
   5568 #endif
   5569 
   5570 #ifdef __LITTLE_ENDIAN__
   5571 __ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
   5572   uint16x4_t __ret;
   5573   __ret = __p0 ^ __p1;
   5574   return __ret;
   5575 }
   5576 #else
   5577 __ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
   5578   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   5579   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   5580   uint16x4_t __ret;
   5581   __ret = __rev0 ^ __rev1;
   5582   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5583   return __ret;
   5584 }
   5585 #endif
   5586 
   5587 #ifdef __LITTLE_ENDIAN__
   5588 __ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
   5589   int8x8_t __ret;
   5590   __ret = __p0 ^ __p1;
   5591   return __ret;
   5592 }
   5593 #else
   5594 __ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
   5595   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   5596   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   5597   int8x8_t __ret;
   5598   __ret = __rev0 ^ __rev1;
   5599   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   5600   return __ret;
   5601 }
   5602 #endif
   5603 
   5604 #ifdef __LITTLE_ENDIAN__
   5605 __ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
   5606   int32x2_t __ret;
   5607   __ret = __p0 ^ __p1;
   5608   return __ret;
   5609 }
   5610 #else
   5611 __ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
   5612   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   5613   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   5614   int32x2_t __ret;
   5615   __ret = __rev0 ^ __rev1;
   5616   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   5617   return __ret;
   5618 }
   5619 #endif
   5620 
   5621 #ifdef __LITTLE_ENDIAN__
   5622 __ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
   5623   int64x1_t __ret;
   5624   __ret = __p0 ^ __p1;
   5625   return __ret;
   5626 }
   5627 #else
   5628 __ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
   5629   int64x1_t __ret;
   5630   __ret = __p0 ^ __p1;
   5631   return __ret;
   5632 }
   5633 #endif
   5634 
   5635 #ifdef __LITTLE_ENDIAN__
   5636 __ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
   5637   int16x4_t __ret;
   5638   __ret = __p0 ^ __p1;
   5639   return __ret;
   5640 }
   5641 #else
   5642 __ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
   5643   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   5644   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   5645   int16x4_t __ret;
   5646   __ret = __rev0 ^ __rev1;
   5647   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   5648   return __ret;
   5649 }
   5650 #endif
   5651 
   5652 #ifdef __LITTLE_ENDIAN__
   5653 #define vext_p8(__p0, __p1, __p2) __extension__ ({ \
   5654   poly8x8_t __s0 = __p0; \
   5655   poly8x8_t __s1 = __p1; \
   5656   poly8x8_t __ret; \
   5657   __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
   5658   __ret; \
   5659 })
   5660 #else
   5661 #define vext_p8(__p0, __p1, __p2) __extension__ ({ \
   5662   poly8x8_t __s0 = __p0; \
   5663   poly8x8_t __s1 = __p1; \
   5664   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   5665   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   5666   poly8x8_t __ret; \
   5667   __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
   5668   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   5669   __ret; \
   5670 })
   5671 #endif
   5672 
   5673 #ifdef __LITTLE_ENDIAN__
   5674 #define vext_p16(__p0, __p1, __p2) __extension__ ({ \
   5675   poly16x4_t __s0 = __p0; \
   5676   poly16x4_t __s1 = __p1; \
   5677   poly16x4_t __ret; \
   5678   __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
   5679   __ret; \
   5680 })
   5681 #else
   5682 #define vext_p16(__p0, __p1, __p2) __extension__ ({ \
   5683   poly16x4_t __s0 = __p0; \
   5684   poly16x4_t __s1 = __p1; \
   5685   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   5686   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   5687   poly16x4_t __ret; \
   5688   __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
   5689   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   5690   __ret; \
   5691 })
   5692 #endif
   5693 
   5694 #ifdef __LITTLE_ENDIAN__
   5695 #define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
   5696   poly8x16_t __s0 = __p0; \
   5697   poly8x16_t __s1 = __p1; \
   5698   poly8x16_t __ret; \
   5699   __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
   5700   __ret; \
   5701 })
   5702 #else
   5703 #define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
   5704   poly8x16_t __s0 = __p0; \
   5705   poly8x16_t __s1 = __p1; \
   5706   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   5707   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   5708   poly8x16_t __ret; \
   5709   __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
   5710   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   5711   __ret; \
   5712 })
   5713 #endif
   5714 
   5715 #ifdef __LITTLE_ENDIAN__
   5716 #define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
   5717   poly16x8_t __s0 = __p0; \
   5718   poly16x8_t __s1 = __p1; \
   5719   poly16x8_t __ret; \
   5720   __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
   5721   __ret; \
   5722 })
   5723 #else
   5724 #define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
   5725   poly16x8_t __s0 = __p0; \
   5726   poly16x8_t __s1 = __p1; \
   5727   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   5728   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   5729   poly16x8_t __ret; \
   5730   __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
   5731   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   5732   __ret; \
   5733 })
   5734 #endif
   5735 
   5736 #ifdef __LITTLE_ENDIAN__
   5737 #define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
   5738   uint8x16_t __s0 = __p0; \
   5739   uint8x16_t __s1 = __p1; \
   5740   uint8x16_t __ret; \
   5741   __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
   5742   __ret; \
   5743 })
   5744 #else
   5745 #define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
   5746   uint8x16_t __s0 = __p0; \
   5747   uint8x16_t __s1 = __p1; \
   5748   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   5749   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   5750   uint8x16_t __ret; \
   5751   __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
   5752   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   5753   __ret; \
   5754 })
   5755 #endif
   5756 
   5757 #ifdef __LITTLE_ENDIAN__
   5758 #define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
   5759   uint32x4_t __s0 = __p0; \
   5760   uint32x4_t __s1 = __p1; \
   5761   uint32x4_t __ret; \
   5762   __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
   5763   __ret; \
   5764 })
   5765 #else
   5766 #define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
   5767   uint32x4_t __s0 = __p0; \
   5768   uint32x4_t __s1 = __p1; \
   5769   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   5770   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   5771   uint32x4_t __ret; \
   5772   __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
   5773   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   5774   __ret; \
   5775 })
   5776 #endif
   5777 
   5778 #ifdef __LITTLE_ENDIAN__
   5779 #define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
   5780   uint64x2_t __s0 = __p0; \
   5781   uint64x2_t __s1 = __p1; \
   5782   uint64x2_t __ret; \
   5783   __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
   5784   __ret; \
   5785 })
   5786 #else
   5787 #define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
   5788   uint64x2_t __s0 = __p0; \
   5789   uint64x2_t __s1 = __p1; \
   5790   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   5791   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   5792   uint64x2_t __ret; \
   5793   __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
   5794   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   5795   __ret; \
   5796 })
   5797 #endif
   5798 
   5799 #ifdef __LITTLE_ENDIAN__
   5800 #define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
   5801   uint16x8_t __s0 = __p0; \
   5802   uint16x8_t __s1 = __p1; \
   5803   uint16x8_t __ret; \
   5804   __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
   5805   __ret; \
   5806 })
   5807 #else
   5808 #define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
   5809   uint16x8_t __s0 = __p0; \
   5810   uint16x8_t __s1 = __p1; \
   5811   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   5812   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   5813   uint16x8_t __ret; \
   5814   __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
   5815   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   5816   __ret; \
   5817 })
   5818 #endif
   5819 
   5820 #ifdef __LITTLE_ENDIAN__
   5821 #define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
   5822   int8x16_t __s0 = __p0; \
   5823   int8x16_t __s1 = __p1; \
   5824   int8x16_t __ret; \
   5825   __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
   5826   __ret; \
   5827 })
   5828 #else
   5829 #define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
   5830   int8x16_t __s0 = __p0; \
   5831   int8x16_t __s1 = __p1; \
   5832   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   5833   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   5834   int8x16_t __ret; \
   5835   __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
   5836   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   5837   __ret; \
   5838 })
   5839 #endif
   5840 
   5841 #ifdef __LITTLE_ENDIAN__
   5842 #define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
   5843   float32x4_t __s0 = __p0; \
   5844   float32x4_t __s1 = __p1; \
   5845   float32x4_t __ret; \
   5846   __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 41); \
   5847   __ret; \
   5848 })
   5849 #else
   5850 #define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
   5851   float32x4_t __s0 = __p0; \
   5852   float32x4_t __s1 = __p1; \
   5853   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   5854   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   5855   float32x4_t __ret; \
   5856   __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 41); \
   5857   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   5858   __ret; \
   5859 })
   5860 #endif
   5861 
   5862 #ifdef __LITTLE_ENDIAN__
   5863 #define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
   5864   int32x4_t __s0 = __p0; \
   5865   int32x4_t __s1 = __p1; \
   5866   int32x4_t __ret; \
   5867   __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
   5868   __ret; \
   5869 })
   5870 #else
   5871 #define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
   5872   int32x4_t __s0 = __p0; \
   5873   int32x4_t __s1 = __p1; \
   5874   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   5875   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   5876   int32x4_t __ret; \
   5877   __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
   5878   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   5879   __ret; \
   5880 })
   5881 #endif
   5882 
   5883 #ifdef __LITTLE_ENDIAN__
   5884 #define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
   5885   int64x2_t __s0 = __p0; \
   5886   int64x2_t __s1 = __p1; \
   5887   int64x2_t __ret; \
   5888   __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
   5889   __ret; \
   5890 })
   5891 #else
   5892 #define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
   5893   int64x2_t __s0 = __p0; \
   5894   int64x2_t __s1 = __p1; \
   5895   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   5896   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   5897   int64x2_t __ret; \
   5898   __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
   5899   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   5900   __ret; \
   5901 })
   5902 #endif
   5903 
   5904 #ifdef __LITTLE_ENDIAN__
   5905 #define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
   5906   int16x8_t __s0 = __p0; \
   5907   int16x8_t __s1 = __p1; \
   5908   int16x8_t __ret; \
   5909   __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
   5910   __ret; \
   5911 })
   5912 #else
   5913 #define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
   5914   int16x8_t __s0 = __p0; \
   5915   int16x8_t __s1 = __p1; \
   5916   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   5917   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   5918   int16x8_t __ret; \
   5919   __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
   5920   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   5921   __ret; \
   5922 })
   5923 #endif
   5924 
   5925 #ifdef __LITTLE_ENDIAN__
   5926 #define vext_u8(__p0, __p1, __p2) __extension__ ({ \
   5927   uint8x8_t __s0 = __p0; \
   5928   uint8x8_t __s1 = __p1; \
   5929   uint8x8_t __ret; \
   5930   __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
   5931   __ret; \
   5932 })
   5933 #else
   5934 #define vext_u8(__p0, __p1, __p2) __extension__ ({ \
   5935   uint8x8_t __s0 = __p0; \
   5936   uint8x8_t __s1 = __p1; \
   5937   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   5938   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   5939   uint8x8_t __ret; \
   5940   __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
   5941   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   5942   __ret; \
   5943 })
   5944 #endif
   5945 
   5946 #ifdef __LITTLE_ENDIAN__
   5947 #define vext_u32(__p0, __p1, __p2) __extension__ ({ \
   5948   uint32x2_t __s0 = __p0; \
   5949   uint32x2_t __s1 = __p1; \
   5950   uint32x2_t __ret; \
   5951   __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
   5952   __ret; \
   5953 })
   5954 #else
   5955 #define vext_u32(__p0, __p1, __p2) __extension__ ({ \
   5956   uint32x2_t __s0 = __p0; \
   5957   uint32x2_t __s1 = __p1; \
   5958   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   5959   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   5960   uint32x2_t __ret; \
   5961   __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
   5962   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   5963   __ret; \
   5964 })
   5965 #endif
   5966 
   5967 #ifdef __LITTLE_ENDIAN__
   5968 #define vext_u64(__p0, __p1, __p2) __extension__ ({ \
   5969   uint64x1_t __s0 = __p0; \
   5970   uint64x1_t __s1 = __p1; \
   5971   uint64x1_t __ret; \
   5972   __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   5973   __ret; \
   5974 })
   5975 #else
   5976 #define vext_u64(__p0, __p1, __p2) __extension__ ({ \
   5977   uint64x1_t __s0 = __p0; \
   5978   uint64x1_t __s1 = __p1; \
   5979   uint64x1_t __ret; \
   5980   __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   5981   __ret; \
   5982 })
   5983 #endif
   5984 
   5985 #ifdef __LITTLE_ENDIAN__
   5986 #define vext_u16(__p0, __p1, __p2) __extension__ ({ \
   5987   uint16x4_t __s0 = __p0; \
   5988   uint16x4_t __s1 = __p1; \
   5989   uint16x4_t __ret; \
   5990   __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
   5991   __ret; \
   5992 })
   5993 #else
   5994 #define vext_u16(__p0, __p1, __p2) __extension__ ({ \
   5995   uint16x4_t __s0 = __p0; \
   5996   uint16x4_t __s1 = __p1; \
   5997   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   5998   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   5999   uint16x4_t __ret; \
   6000   __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
   6001   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   6002   __ret; \
   6003 })
   6004 #endif
   6005 
   6006 #ifdef __LITTLE_ENDIAN__
   6007 #define vext_s8(__p0, __p1, __p2) __extension__ ({ \
   6008   int8x8_t __s0 = __p0; \
   6009   int8x8_t __s1 = __p1; \
   6010   int8x8_t __ret; \
   6011   __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
   6012   __ret; \
   6013 })
   6014 #else
   6015 #define vext_s8(__p0, __p1, __p2) __extension__ ({ \
   6016   int8x8_t __s0 = __p0; \
   6017   int8x8_t __s1 = __p1; \
   6018   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   6019   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   6020   int8x8_t __ret; \
   6021   __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
   6022   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   6023   __ret; \
   6024 })
   6025 #endif
   6026 
   6027 #ifdef __LITTLE_ENDIAN__
   6028 #define vext_f32(__p0, __p1, __p2) __extension__ ({ \
   6029   float32x2_t __s0 = __p0; \
   6030   float32x2_t __s1 = __p1; \
   6031   float32x2_t __ret; \
   6032   __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 9); \
   6033   __ret; \
   6034 })
   6035 #else
   6036 #define vext_f32(__p0, __p1, __p2) __extension__ ({ \
   6037   float32x2_t __s0 = __p0; \
   6038   float32x2_t __s1 = __p1; \
   6039   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   6040   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   6041   float32x2_t __ret; \
   6042   __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 9); \
   6043   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   6044   __ret; \
   6045 })
   6046 #endif
   6047 
   6048 #ifdef __LITTLE_ENDIAN__
   6049 #define vext_s32(__p0, __p1, __p2) __extension__ ({ \
   6050   int32x2_t __s0 = __p0; \
   6051   int32x2_t __s1 = __p1; \
   6052   int32x2_t __ret; \
   6053   __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
   6054   __ret; \
   6055 })
   6056 #else
   6057 #define vext_s32(__p0, __p1, __p2) __extension__ ({ \
   6058   int32x2_t __s0 = __p0; \
   6059   int32x2_t __s1 = __p1; \
   6060   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   6061   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   6062   int32x2_t __ret; \
   6063   __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
   6064   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   6065   __ret; \
   6066 })
   6067 #endif
   6068 
   6069 #ifdef __LITTLE_ENDIAN__
   6070 #define vext_s64(__p0, __p1, __p2) __extension__ ({ \
   6071   int64x1_t __s0 = __p0; \
   6072   int64x1_t __s1 = __p1; \
   6073   int64x1_t __ret; \
   6074   __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   6075   __ret; \
   6076 })
   6077 #else
   6078 #define vext_s64(__p0, __p1, __p2) __extension__ ({ \
   6079   int64x1_t __s0 = __p0; \
   6080   int64x1_t __s1 = __p1; \
   6081   int64x1_t __ret; \
   6082   __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   6083   __ret; \
   6084 })
   6085 #endif
   6086 
   6087 #ifdef __LITTLE_ENDIAN__
   6088 #define vext_s16(__p0, __p1, __p2) __extension__ ({ \
   6089   int16x4_t __s0 = __p0; \
   6090   int16x4_t __s1 = __p1; \
   6091   int16x4_t __ret; \
   6092   __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
   6093   __ret; \
   6094 })
   6095 #else
   6096 #define vext_s16(__p0, __p1, __p2) __extension__ ({ \
   6097   int16x4_t __s0 = __p0; \
   6098   int16x4_t __s1 = __p1; \
   6099   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   6100   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   6101   int16x4_t __ret; \
   6102   __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
   6103   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   6104   __ret; \
   6105 })
   6106 #endif
   6107 
   6108 #ifdef __LITTLE_ENDIAN__
   6109 __ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
   6110   poly8x8_t __ret;
   6111   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
   6112   return __ret;
   6113 }
   6114 #else
   6115 __ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
   6116   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   6117   poly8x8_t __ret;
   6118   __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
   6119   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   6120   return __ret;
   6121 }
   6122 __ai poly8x8_t __noswap_vget_high_p8(poly8x16_t __p0) {
   6123   poly8x8_t __ret;
   6124   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
   6125   return __ret;
   6126 }
   6127 #endif
   6128 
   6129 #ifdef __LITTLE_ENDIAN__
   6130 __ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
   6131   poly16x4_t __ret;
   6132   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
   6133   return __ret;
   6134 }
   6135 #else
   6136 __ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
   6137   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   6138   poly16x4_t __ret;
   6139   __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
   6140   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   6141   return __ret;
   6142 }
   6143 #endif
   6144 
   6145 #ifdef __LITTLE_ENDIAN__
   6146 __ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
   6147   uint8x8_t __ret;
   6148   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
   6149   return __ret;
   6150 }
   6151 #else
   6152 __ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
   6153   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   6154   uint8x8_t __ret;
   6155   __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
   6156   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   6157   return __ret;
   6158 }
   6159 __ai uint8x8_t __noswap_vget_high_u8(uint8x16_t __p0) {
   6160   uint8x8_t __ret;
   6161   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
   6162   return __ret;
   6163 }
   6164 #endif
   6165 
   6166 #ifdef __LITTLE_ENDIAN__
   6167 __ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
   6168   uint32x2_t __ret;
   6169   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
   6170   return __ret;
   6171 }
   6172 #else
   6173 __ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
   6174   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   6175   uint32x2_t __ret;
   6176   __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
   6177   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   6178   return __ret;
   6179 }
   6180 __ai uint32x2_t __noswap_vget_high_u32(uint32x4_t __p0) {
   6181   uint32x2_t __ret;
   6182   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
   6183   return __ret;
   6184 }
   6185 #endif
   6186 
   6187 #ifdef __LITTLE_ENDIAN__
   6188 __ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
   6189   uint64x1_t __ret;
   6190   __ret = __builtin_shufflevector(__p0, __p0, 1);
   6191   return __ret;
   6192 }
   6193 #else
   6194 __ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
   6195   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   6196   uint64x1_t __ret;
   6197   __ret = __builtin_shufflevector(__rev0, __rev0, 1);
   6198   return __ret;
   6199 }
   6200 #endif
   6201 
   6202 #ifdef __LITTLE_ENDIAN__
   6203 __ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
   6204   uint16x4_t __ret;
   6205   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
   6206   return __ret;
   6207 }
   6208 #else
   6209 __ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
   6210   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   6211   uint16x4_t __ret;
   6212   __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
   6213   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   6214   return __ret;
   6215 }
   6216 __ai uint16x4_t __noswap_vget_high_u16(uint16x8_t __p0) {
   6217   uint16x4_t __ret;
   6218   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
   6219   return __ret;
   6220 }
   6221 #endif
   6222 
   6223 #ifdef __LITTLE_ENDIAN__
   6224 __ai int8x8_t vget_high_s8(int8x16_t __p0) {
   6225   int8x8_t __ret;
   6226   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
   6227   return __ret;
   6228 }
   6229 #else
   6230 __ai int8x8_t vget_high_s8(int8x16_t __p0) {
   6231   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   6232   int8x8_t __ret;
   6233   __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
   6234   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   6235   return __ret;
   6236 }
   6237 __ai int8x8_t __noswap_vget_high_s8(int8x16_t __p0) {
   6238   int8x8_t __ret;
   6239   __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
   6240   return __ret;
   6241 }
   6242 #endif
   6243 
   6244 #ifdef __LITTLE_ENDIAN__
   6245 __ai float32x2_t vget_high_f32(float32x4_t __p0) {
   6246   float32x2_t __ret;
   6247   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
   6248   return __ret;
   6249 }
   6250 #else
   6251 __ai float32x2_t vget_high_f32(float32x4_t __p0) {
   6252   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   6253   float32x2_t __ret;
   6254   __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
   6255   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   6256   return __ret;
   6257 }
   6258 __ai float32x2_t __noswap_vget_high_f32(float32x4_t __p0) {
   6259   float32x2_t __ret;
   6260   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
   6261   return __ret;
   6262 }
   6263 #endif
   6264 
   6265 #ifdef __LITTLE_ENDIAN__
   6266 __ai float16x4_t vget_high_f16(float16x8_t __p0) {
   6267   float16x4_t __ret;
   6268   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
   6269   return __ret;
   6270 }
   6271 #else
   6272 __ai float16x4_t vget_high_f16(float16x8_t __p0) {
   6273   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   6274   float16x4_t __ret;
   6275   __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
   6276   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   6277   return __ret;
   6278 }
   6279 __ai float16x4_t __noswap_vget_high_f16(float16x8_t __p0) {
   6280   float16x4_t __ret;
   6281   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
   6282   return __ret;
   6283 }
   6284 #endif
   6285 
   6286 #ifdef __LITTLE_ENDIAN__
   6287 __ai int32x2_t vget_high_s32(int32x4_t __p0) {
   6288   int32x2_t __ret;
   6289   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
   6290   return __ret;
   6291 }
   6292 #else
   6293 __ai int32x2_t vget_high_s32(int32x4_t __p0) {
   6294   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   6295   int32x2_t __ret;
   6296   __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
   6297   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   6298   return __ret;
   6299 }
   6300 __ai int32x2_t __noswap_vget_high_s32(int32x4_t __p0) {
   6301   int32x2_t __ret;
   6302   __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
   6303   return __ret;
   6304 }
   6305 #endif
   6306 
   6307 #ifdef __LITTLE_ENDIAN__
   6308 __ai int64x1_t vget_high_s64(int64x2_t __p0) {
   6309   int64x1_t __ret;
   6310   __ret = __builtin_shufflevector(__p0, __p0, 1);
   6311   return __ret;
   6312 }
   6313 #else
   6314 __ai int64x1_t vget_high_s64(int64x2_t __p0) {
   6315   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   6316   int64x1_t __ret;
   6317   __ret = __builtin_shufflevector(__rev0, __rev0, 1);
   6318   return __ret;
   6319 }
   6320 #endif
   6321 
   6322 #ifdef __LITTLE_ENDIAN__
   6323 __ai int16x4_t vget_high_s16(int16x8_t __p0) {
   6324   int16x4_t __ret;
   6325   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
   6326   return __ret;
   6327 }
   6328 #else
   6329 __ai int16x4_t vget_high_s16(int16x8_t __p0) {
   6330   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   6331   int16x4_t __ret;
   6332   __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
   6333   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   6334   return __ret;
   6335 }
   6336 __ai int16x4_t __noswap_vget_high_s16(int16x8_t __p0) {
   6337   int16x4_t __ret;
   6338   __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
   6339   return __ret;
   6340 }
   6341 #endif
   6342 
   6343 #ifdef __LITTLE_ENDIAN__
   6344 #define vget_lane_p8(__p0, __p1) __extension__ ({ \
   6345   poly8x8_t __s0 = __p0; \
   6346   poly8_t __ret; \
   6347   __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
   6348   __ret; \
   6349 })
   6350 #else
   6351 #define vget_lane_p8(__p0, __p1) __extension__ ({ \
   6352   poly8x8_t __s0 = __p0; \
   6353   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   6354   poly8_t __ret; \
   6355   __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
   6356   __ret; \
   6357 })
   6358 #define __noswap_vget_lane_p8(__p0, __p1) __extension__ ({ \
   6359   poly8x8_t __s0 = __p0; \
   6360   poly8_t __ret; \
   6361   __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
   6362   __ret; \
   6363 })
   6364 #endif
   6365 
   6366 #ifdef __LITTLE_ENDIAN__
   6367 #define vget_lane_p16(__p0, __p1) __extension__ ({ \
   6368   poly16x4_t __s0 = __p0; \
   6369   poly16_t __ret; \
   6370   __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
   6371   __ret; \
   6372 })
   6373 #else
   6374 #define vget_lane_p16(__p0, __p1) __extension__ ({ \
   6375   poly16x4_t __s0 = __p0; \
   6376   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   6377   poly16_t __ret; \
   6378   __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
   6379   __ret; \
   6380 })
   6381 #define __noswap_vget_lane_p16(__p0, __p1) __extension__ ({ \
   6382   poly16x4_t __s0 = __p0; \
   6383   poly16_t __ret; \
   6384   __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
   6385   __ret; \
   6386 })
   6387 #endif
   6388 
   6389 #ifdef __LITTLE_ENDIAN__
   6390 #define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
   6391   poly8x16_t __s0 = __p0; \
   6392   poly8_t __ret; \
   6393   __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
   6394   __ret; \
   6395 })
   6396 #else
   6397 #define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
   6398   poly8x16_t __s0 = __p0; \
   6399   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   6400   poly8_t __ret; \
   6401   __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
   6402   __ret; \
   6403 })
   6404 #define __noswap_vgetq_lane_p8(__p0, __p1) __extension__ ({ \
   6405   poly8x16_t __s0 = __p0; \
   6406   poly8_t __ret; \
   6407   __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
   6408   __ret; \
   6409 })
   6410 #endif
   6411 
   6412 #ifdef __LITTLE_ENDIAN__
   6413 #define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
   6414   poly16x8_t __s0 = __p0; \
   6415   poly16_t __ret; \
   6416   __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
   6417   __ret; \
   6418 })
   6419 #else
   6420 #define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
   6421   poly16x8_t __s0 = __p0; \
   6422   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   6423   poly16_t __ret; \
   6424   __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
   6425   __ret; \
   6426 })
   6427 #define __noswap_vgetq_lane_p16(__p0, __p1) __extension__ ({ \
   6428   poly16x8_t __s0 = __p0; \
   6429   poly16_t __ret; \
   6430   __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
   6431   __ret; \
   6432 })
   6433 #endif
   6434 
   6435 #ifdef __LITTLE_ENDIAN__
   6436 #define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
   6437   uint8x16_t __s0 = __p0; \
   6438   uint8_t __ret; \
   6439   __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
   6440   __ret; \
   6441 })
   6442 #else
   6443 #define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
   6444   uint8x16_t __s0 = __p0; \
   6445   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   6446   uint8_t __ret; \
   6447   __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
   6448   __ret; \
   6449 })
   6450 #define __noswap_vgetq_lane_u8(__p0, __p1) __extension__ ({ \
   6451   uint8x16_t __s0 = __p0; \
   6452   uint8_t __ret; \
   6453   __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
   6454   __ret; \
   6455 })
   6456 #endif
   6457 
   6458 #ifdef __LITTLE_ENDIAN__
   6459 #define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
   6460   uint32x4_t __s0 = __p0; \
   6461   uint32_t __ret; \
   6462   __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
   6463   __ret; \
   6464 })
   6465 #else
   6466 #define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
   6467   uint32x4_t __s0 = __p0; \
   6468   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   6469   uint32_t __ret; \
   6470   __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
   6471   __ret; \
   6472 })
   6473 #define __noswap_vgetq_lane_u32(__p0, __p1) __extension__ ({ \
   6474   uint32x4_t __s0 = __p0; \
   6475   uint32_t __ret; \
   6476   __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
   6477   __ret; \
   6478 })
   6479 #endif
   6480 
   6481 #ifdef __LITTLE_ENDIAN__
   6482 #define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
   6483   uint64x2_t __s0 = __p0; \
   6484   uint64_t __ret; \
   6485   __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
   6486   __ret; \
   6487 })
   6488 #else
   6489 #define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
   6490   uint64x2_t __s0 = __p0; \
   6491   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   6492   uint64_t __ret; \
   6493   __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
   6494   __ret; \
   6495 })
   6496 #define __noswap_vgetq_lane_u64(__p0, __p1) __extension__ ({ \
   6497   uint64x2_t __s0 = __p0; \
   6498   uint64_t __ret; \
   6499   __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
   6500   __ret; \
   6501 })
   6502 #endif
   6503 
   6504 #ifdef __LITTLE_ENDIAN__
   6505 #define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
   6506   uint16x8_t __s0 = __p0; \
   6507   uint16_t __ret; \
   6508   __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
   6509   __ret; \
   6510 })
   6511 #else
   6512 #define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
   6513   uint16x8_t __s0 = __p0; \
   6514   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   6515   uint16_t __ret; \
   6516   __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
   6517   __ret; \
   6518 })
   6519 #define __noswap_vgetq_lane_u16(__p0, __p1) __extension__ ({ \
   6520   uint16x8_t __s0 = __p0; \
   6521   uint16_t __ret; \
   6522   __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
   6523   __ret; \
   6524 })
   6525 #endif
   6526 
   6527 #ifdef __LITTLE_ENDIAN__
   6528 #define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
   6529   int8x16_t __s0 = __p0; \
   6530   int8_t __ret; \
   6531   __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
   6532   __ret; \
   6533 })
   6534 #else
   6535 #define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
   6536   int8x16_t __s0 = __p0; \
   6537   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   6538   int8_t __ret; \
   6539   __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
   6540   __ret; \
   6541 })
   6542 #define __noswap_vgetq_lane_s8(__p0, __p1) __extension__ ({ \
   6543   int8x16_t __s0 = __p0; \
   6544   int8_t __ret; \
   6545   __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
   6546   __ret; \
   6547 })
   6548 #endif
   6549 
   6550 #ifdef __LITTLE_ENDIAN__
   6551 #define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
   6552   float32x4_t __s0 = __p0; \
   6553   float32_t __ret; \
   6554   __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
   6555   __ret; \
   6556 })
   6557 #else
   6558 #define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
   6559   float32x4_t __s0 = __p0; \
   6560   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   6561   float32_t __ret; \
   6562   __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__rev0, __p1); \
   6563   __ret; \
   6564 })
   6565 #define __noswap_vgetq_lane_f32(__p0, __p1) __extension__ ({ \
   6566   float32x4_t __s0 = __p0; \
   6567   float32_t __ret; \
   6568   __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
   6569   __ret; \
   6570 })
   6571 #endif
   6572 
   6573 #ifdef __LITTLE_ENDIAN__
   6574 #define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
   6575   int32x4_t __s0 = __p0; \
   6576   int32_t __ret; \
   6577   __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
   6578   __ret; \
   6579 })
   6580 #else
   6581 #define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
   6582   int32x4_t __s0 = __p0; \
   6583   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   6584   int32_t __ret; \
   6585   __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
   6586   __ret; \
   6587 })
   6588 #define __noswap_vgetq_lane_s32(__p0, __p1) __extension__ ({ \
   6589   int32x4_t __s0 = __p0; \
   6590   int32_t __ret; \
   6591   __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
   6592   __ret; \
   6593 })
   6594 #endif
   6595 
   6596 #ifdef __LITTLE_ENDIAN__
   6597 #define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
   6598   int64x2_t __s0 = __p0; \
   6599   int64_t __ret; \
   6600   __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
   6601   __ret; \
   6602 })
   6603 #else
   6604 #define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
   6605   int64x2_t __s0 = __p0; \
   6606   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   6607   int64_t __ret; \
   6608   __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
   6609   __ret; \
   6610 })
   6611 #define __noswap_vgetq_lane_s64(__p0, __p1) __extension__ ({ \
   6612   int64x2_t __s0 = __p0; \
   6613   int64_t __ret; \
   6614   __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
   6615   __ret; \
   6616 })
   6617 #endif
   6618 
   6619 #ifdef __LITTLE_ENDIAN__
   6620 #define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
   6621   int16x8_t __s0 = __p0; \
   6622   int16_t __ret; \
   6623   __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
   6624   __ret; \
   6625 })
   6626 #else
   6627 #define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
   6628   int16x8_t __s0 = __p0; \
   6629   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   6630   int16_t __ret; \
   6631   __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
   6632   __ret; \
   6633 })
   6634 #define __noswap_vgetq_lane_s16(__p0, __p1) __extension__ ({ \
   6635   int16x8_t __s0 = __p0; \
   6636   int16_t __ret; \
   6637   __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
   6638   __ret; \
   6639 })
   6640 #endif
   6641 
   6642 #ifdef __LITTLE_ENDIAN__
   6643 #define vget_lane_u8(__p0, __p1) __extension__ ({ \
   6644   uint8x8_t __s0 = __p0; \
   6645   uint8_t __ret; \
   6646   __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
   6647   __ret; \
   6648 })
   6649 #else
   6650 #define vget_lane_u8(__p0, __p1) __extension__ ({ \
   6651   uint8x8_t __s0 = __p0; \
   6652   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   6653   uint8_t __ret; \
   6654   __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
   6655   __ret; \
   6656 })
   6657 #define __noswap_vget_lane_u8(__p0, __p1) __extension__ ({ \
   6658   uint8x8_t __s0 = __p0; \
   6659   uint8_t __ret; \
   6660   __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
   6661   __ret; \
   6662 })
   6663 #endif
   6664 
   6665 #ifdef __LITTLE_ENDIAN__
   6666 #define vget_lane_u32(__p0, __p1) __extension__ ({ \
   6667   uint32x2_t __s0 = __p0; \
   6668   uint32_t __ret; \
   6669   __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
   6670   __ret; \
   6671 })
   6672 #else
   6673 #define vget_lane_u32(__p0, __p1) __extension__ ({ \
   6674   uint32x2_t __s0 = __p0; \
   6675   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   6676   uint32_t __ret; \
   6677   __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
   6678   __ret; \
   6679 })
   6680 #define __noswap_vget_lane_u32(__p0, __p1) __extension__ ({ \
   6681   uint32x2_t __s0 = __p0; \
   6682   uint32_t __ret; \
   6683   __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
   6684   __ret; \
   6685 })
   6686 #endif
   6687 
   6688 #ifdef __LITTLE_ENDIAN__
   6689 #define vget_lane_u64(__p0, __p1) __extension__ ({ \
   6690   uint64x1_t __s0 = __p0; \
   6691   uint64_t __ret; \
   6692   __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
   6693   __ret; \
   6694 })
   6695 #else
   6696 #define vget_lane_u64(__p0, __p1) __extension__ ({ \
   6697   uint64x1_t __s0 = __p0; \
   6698   uint64_t __ret; \
   6699   __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
   6700   __ret; \
   6701 })
   6702 #define __noswap_vget_lane_u64(__p0, __p1) __extension__ ({ \
   6703   uint64x1_t __s0 = __p0; \
   6704   uint64_t __ret; \
   6705   __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
   6706   __ret; \
   6707 })
   6708 #endif
   6709 
   6710 #ifdef __LITTLE_ENDIAN__
   6711 #define vget_lane_u16(__p0, __p1) __extension__ ({ \
   6712   uint16x4_t __s0 = __p0; \
   6713   uint16_t __ret; \
   6714   __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
   6715   __ret; \
   6716 })
   6717 #else
   6718 #define vget_lane_u16(__p0, __p1) __extension__ ({ \
   6719   uint16x4_t __s0 = __p0; \
   6720   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   6721   uint16_t __ret; \
   6722   __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
   6723   __ret; \
   6724 })
   6725 #define __noswap_vget_lane_u16(__p0, __p1) __extension__ ({ \
   6726   uint16x4_t __s0 = __p0; \
   6727   uint16_t __ret; \
   6728   __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
   6729   __ret; \
   6730 })
   6731 #endif
   6732 
   6733 #ifdef __LITTLE_ENDIAN__
   6734 #define vget_lane_s8(__p0, __p1) __extension__ ({ \
   6735   int8x8_t __s0 = __p0; \
   6736   int8_t __ret; \
   6737   __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
   6738   __ret; \
   6739 })
   6740 #else
   6741 #define vget_lane_s8(__p0, __p1) __extension__ ({ \
   6742   int8x8_t __s0 = __p0; \
   6743   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   6744   int8_t __ret; \
   6745   __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
   6746   __ret; \
   6747 })
   6748 #define __noswap_vget_lane_s8(__p0, __p1) __extension__ ({ \
   6749   int8x8_t __s0 = __p0; \
   6750   int8_t __ret; \
   6751   __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
   6752   __ret; \
   6753 })
   6754 #endif
   6755 
   6756 #ifdef __LITTLE_ENDIAN__
   6757 #define vget_lane_f32(__p0, __p1) __extension__ ({ \
   6758   float32x2_t __s0 = __p0; \
   6759   float32_t __ret; \
   6760   __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
   6761   __ret; \
   6762 })
   6763 #else
   6764 #define vget_lane_f32(__p0, __p1) __extension__ ({ \
   6765   float32x2_t __s0 = __p0; \
   6766   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   6767   float32_t __ret; \
   6768   __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__rev0, __p1); \
   6769   __ret; \
   6770 })
   6771 #define __noswap_vget_lane_f32(__p0, __p1) __extension__ ({ \
   6772   float32x2_t __s0 = __p0; \
   6773   float32_t __ret; \
   6774   __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
   6775   __ret; \
   6776 })
   6777 #endif
   6778 
   6779 #ifdef __LITTLE_ENDIAN__
   6780 #define vget_lane_s32(__p0, __p1) __extension__ ({ \
   6781   int32x2_t __s0 = __p0; \
   6782   int32_t __ret; \
   6783   __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
   6784   __ret; \
   6785 })
   6786 #else
   6787 #define vget_lane_s32(__p0, __p1) __extension__ ({ \
   6788   int32x2_t __s0 = __p0; \
   6789   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   6790   int32_t __ret; \
   6791   __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
   6792   __ret; \
   6793 })
   6794 #define __noswap_vget_lane_s32(__p0, __p1) __extension__ ({ \
   6795   int32x2_t __s0 = __p0; \
   6796   int32_t __ret; \
   6797   __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
   6798   __ret; \
   6799 })
   6800 #endif
   6801 
   6802 #ifdef __LITTLE_ENDIAN__
   6803 #define vget_lane_s64(__p0, __p1) __extension__ ({ \
   6804   int64x1_t __s0 = __p0; \
   6805   int64_t __ret; \
   6806   __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
   6807   __ret; \
   6808 })
   6809 #else
   6810 #define vget_lane_s64(__p0, __p1) __extension__ ({ \
   6811   int64x1_t __s0 = __p0; \
   6812   int64_t __ret; \
   6813   __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
   6814   __ret; \
   6815 })
   6816 #define __noswap_vget_lane_s64(__p0, __p1) __extension__ ({ \
   6817   int64x1_t __s0 = __p0; \
   6818   int64_t __ret; \
   6819   __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
   6820   __ret; \
   6821 })
   6822 #endif
   6823 
   6824 #ifdef __LITTLE_ENDIAN__
   6825 #define vget_lane_s16(__p0, __p1) __extension__ ({ \
   6826   int16x4_t __s0 = __p0; \
   6827   int16_t __ret; \
   6828   __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
   6829   __ret; \
   6830 })
   6831 #else
   6832 #define vget_lane_s16(__p0, __p1) __extension__ ({ \
   6833   int16x4_t __s0 = __p0; \
   6834   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   6835   int16_t __ret; \
   6836   __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
   6837   __ret; \
   6838 })
   6839 #define __noswap_vget_lane_s16(__p0, __p1) __extension__ ({ \
   6840   int16x4_t __s0 = __p0; \
   6841   int16_t __ret; \
   6842   __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
   6843   __ret; \
   6844 })
   6845 #endif
   6846 
   6847 #ifdef __LITTLE_ENDIAN__
   6848 __ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
   6849   poly8x8_t __ret;
   6850   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
   6851   return __ret;
   6852 }
   6853 #else
   6854 __ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
   6855   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   6856   poly8x8_t __ret;
   6857   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
   6858   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   6859   return __ret;
   6860 }
   6861 #endif
   6862 
   6863 #ifdef __LITTLE_ENDIAN__
   6864 __ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
   6865   poly16x4_t __ret;
   6866   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
   6867   return __ret;
   6868 }
   6869 #else
   6870 __ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
   6871   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   6872   poly16x4_t __ret;
   6873   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
   6874   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   6875   return __ret;
   6876 }
   6877 #endif
   6878 
   6879 #ifdef __LITTLE_ENDIAN__
   6880 __ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
   6881   uint8x8_t __ret;
   6882   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
   6883   return __ret;
   6884 }
   6885 #else
   6886 __ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
   6887   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   6888   uint8x8_t __ret;
   6889   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
   6890   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   6891   return __ret;
   6892 }
   6893 #endif
   6894 
   6895 #ifdef __LITTLE_ENDIAN__
   6896 __ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
   6897   uint32x2_t __ret;
   6898   __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
   6899   return __ret;
   6900 }
   6901 #else
   6902 __ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
   6903   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   6904   uint32x2_t __ret;
   6905   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
   6906   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   6907   return __ret;
   6908 }
   6909 #endif
   6910 
   6911 #ifdef __LITTLE_ENDIAN__
   6912 __ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
   6913   uint64x1_t __ret;
   6914   __ret = __builtin_shufflevector(__p0, __p0, 0);
   6915   return __ret;
   6916 }
   6917 #else
   6918 __ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
   6919   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   6920   uint64x1_t __ret;
   6921   __ret = __builtin_shufflevector(__rev0, __rev0, 0);
   6922   return __ret;
   6923 }
   6924 #endif
   6925 
   6926 #ifdef __LITTLE_ENDIAN__
   6927 __ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
   6928   uint16x4_t __ret;
   6929   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
   6930   return __ret;
   6931 }
   6932 #else
   6933 __ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
   6934   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   6935   uint16x4_t __ret;
   6936   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
   6937   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   6938   return __ret;
   6939 }
   6940 #endif
   6941 
   6942 #ifdef __LITTLE_ENDIAN__
   6943 __ai int8x8_t vget_low_s8(int8x16_t __p0) {
   6944   int8x8_t __ret;
   6945   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
   6946   return __ret;
   6947 }
   6948 #else
   6949 __ai int8x8_t vget_low_s8(int8x16_t __p0) {
   6950   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   6951   int8x8_t __ret;
   6952   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
   6953   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   6954   return __ret;
   6955 }
   6956 #endif
   6957 
   6958 #ifdef __LITTLE_ENDIAN__
   6959 __ai float32x2_t vget_low_f32(float32x4_t __p0) {
   6960   float32x2_t __ret;
   6961   __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
   6962   return __ret;
   6963 }
   6964 #else
   6965 __ai float32x2_t vget_low_f32(float32x4_t __p0) {
   6966   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   6967   float32x2_t __ret;
   6968   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
   6969   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   6970   return __ret;
   6971 }
   6972 #endif
   6973 
   6974 #ifdef __LITTLE_ENDIAN__
   6975 __ai float16x4_t vget_low_f16(float16x8_t __p0) {
   6976   float16x4_t __ret;
   6977   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
   6978   return __ret;
   6979 }
   6980 #else
   6981 __ai float16x4_t vget_low_f16(float16x8_t __p0) {
   6982   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   6983   float16x4_t __ret;
   6984   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
   6985   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   6986   return __ret;
   6987 }
   6988 #endif
   6989 
   6990 #ifdef __LITTLE_ENDIAN__
   6991 __ai int32x2_t vget_low_s32(int32x4_t __p0) {
   6992   int32x2_t __ret;
   6993   __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
   6994   return __ret;
   6995 }
   6996 #else
   6997 __ai int32x2_t vget_low_s32(int32x4_t __p0) {
   6998   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   6999   int32x2_t __ret;
   7000   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
   7001   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   7002   return __ret;
   7003 }
   7004 #endif
   7005 
   7006 #ifdef __LITTLE_ENDIAN__
   7007 __ai int64x1_t vget_low_s64(int64x2_t __p0) {
   7008   int64x1_t __ret;
   7009   __ret = __builtin_shufflevector(__p0, __p0, 0);
   7010   return __ret;
   7011 }
   7012 #else
   7013 __ai int64x1_t vget_low_s64(int64x2_t __p0) {
   7014   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   7015   int64x1_t __ret;
   7016   __ret = __builtin_shufflevector(__rev0, __rev0, 0);
   7017   return __ret;
   7018 }
   7019 #endif
   7020 
   7021 #ifdef __LITTLE_ENDIAN__
   7022 __ai int16x4_t vget_low_s16(int16x8_t __p0) {
   7023   int16x4_t __ret;
   7024   __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
   7025   return __ret;
   7026 }
   7027 #else
   7028 __ai int16x4_t vget_low_s16(int16x8_t __p0) {
   7029   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   7030   int16x4_t __ret;
   7031   __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
   7032   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   7033   return __ret;
   7034 }
   7035 #endif
   7036 
   7037 #ifdef __LITTLE_ENDIAN__
   7038 __ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   7039   uint8x16_t __ret;
   7040   __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   7041   return __ret;
   7042 }
   7043 #else
   7044 __ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   7045   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7046   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7047   uint8x16_t __ret;
   7048   __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   7049   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7050   return __ret;
   7051 }
   7052 #endif
   7053 
   7054 #ifdef __LITTLE_ENDIAN__
   7055 __ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   7056   uint32x4_t __ret;
   7057   __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   7058   return __ret;
   7059 }
   7060 #else
   7061 __ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   7062   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   7063   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   7064   uint32x4_t __ret;
   7065   __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   7066   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   7067   return __ret;
   7068 }
   7069 #endif
   7070 
   7071 #ifdef __LITTLE_ENDIAN__
   7072 __ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   7073   uint16x8_t __ret;
   7074   __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   7075   return __ret;
   7076 }
   7077 #else
   7078 __ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   7079   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   7080   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   7081   uint16x8_t __ret;
   7082   __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   7083   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   7084   return __ret;
   7085 }
   7086 #endif
   7087 
   7088 #ifdef __LITTLE_ENDIAN__
   7089 __ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   7090   int8x16_t __ret;
   7091   __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   7092   return __ret;
   7093 }
   7094 #else
   7095 __ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   7096   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7097   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7098   int8x16_t __ret;
   7099   __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   7100   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7101   return __ret;
   7102 }
   7103 #endif
   7104 
   7105 #ifdef __LITTLE_ENDIAN__
   7106 __ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   7107   int32x4_t __ret;
   7108   __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   7109   return __ret;
   7110 }
   7111 #else
   7112 __ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   7113   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   7114   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   7115   int32x4_t __ret;
   7116   __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   7117   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   7118   return __ret;
   7119 }
   7120 #endif
   7121 
   7122 #ifdef __LITTLE_ENDIAN__
   7123 __ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   7124   int16x8_t __ret;
   7125   __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   7126   return __ret;
   7127 }
   7128 #else
   7129 __ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   7130   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   7131   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   7132   int16x8_t __ret;
   7133   __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   7134   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   7135   return __ret;
   7136 }
   7137 #endif
   7138 
   7139 #ifdef __LITTLE_ENDIAN__
   7140 __ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   7141   uint8x8_t __ret;
   7142   __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   7143   return __ret;
   7144 }
   7145 #else
   7146 __ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   7147   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   7148   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   7149   uint8x8_t __ret;
   7150   __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   7151   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   7152   return __ret;
   7153 }
   7154 #endif
   7155 
   7156 #ifdef __LITTLE_ENDIAN__
   7157 __ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   7158   uint32x2_t __ret;
   7159   __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   7160   return __ret;
   7161 }
   7162 #else
   7163 __ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   7164   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   7165   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   7166   uint32x2_t __ret;
   7167   __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   7168   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   7169   return __ret;
   7170 }
   7171 #endif
   7172 
   7173 #ifdef __LITTLE_ENDIAN__
   7174 __ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   7175   uint16x4_t __ret;
   7176   __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   7177   return __ret;
   7178 }
   7179 #else
   7180 __ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   7181   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   7182   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   7183   uint16x4_t __ret;
   7184   __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   7185   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   7186   return __ret;
   7187 }
   7188 #endif
   7189 
   7190 #ifdef __LITTLE_ENDIAN__
   7191 __ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
   7192   int8x8_t __ret;
   7193   __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   7194   return __ret;
   7195 }
   7196 #else
   7197 __ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
   7198   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   7199   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   7200   int8x8_t __ret;
   7201   __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   7202   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   7203   return __ret;
   7204 }
   7205 #endif
   7206 
   7207 #ifdef __LITTLE_ENDIAN__
   7208 __ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
   7209   int32x2_t __ret;
   7210   __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   7211   return __ret;
   7212 }
   7213 #else
   7214 __ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
   7215   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   7216   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   7217   int32x2_t __ret;
   7218   __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   7219   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   7220   return __ret;
   7221 }
   7222 #endif
   7223 
   7224 #ifdef __LITTLE_ENDIAN__
   7225 __ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
   7226   int16x4_t __ret;
   7227   __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   7228   return __ret;
   7229 }
   7230 #else
   7231 __ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
   7232   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   7233   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   7234   int16x4_t __ret;
   7235   __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   7236   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   7237   return __ret;
   7238 }
   7239 #endif
   7240 
   7241 #ifdef __LITTLE_ENDIAN__
   7242 __ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   7243   uint8x16_t __ret;
   7244   __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   7245   return __ret;
   7246 }
   7247 #else
   7248 __ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   7249   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7250   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7251   uint8x16_t __ret;
   7252   __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   7253   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7254   return __ret;
   7255 }
   7256 #endif
   7257 
   7258 #ifdef __LITTLE_ENDIAN__
   7259 __ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   7260   uint32x4_t __ret;
   7261   __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   7262   return __ret;
   7263 }
   7264 #else
   7265 __ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   7266   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   7267   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   7268   uint32x4_t __ret;
   7269   __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   7270   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   7271   return __ret;
   7272 }
   7273 #endif
   7274 
   7275 #ifdef __LITTLE_ENDIAN__
   7276 __ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   7277   uint16x8_t __ret;
   7278   __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   7279   return __ret;
   7280 }
   7281 #else
   7282 __ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   7283   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   7284   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   7285   uint16x8_t __ret;
   7286   __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   7287   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   7288   return __ret;
   7289 }
   7290 #endif
   7291 
   7292 #ifdef __LITTLE_ENDIAN__
   7293 __ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
   7294   int8x16_t __ret;
   7295   __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   7296   return __ret;
   7297 }
   7298 #else
   7299 __ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
   7300   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7301   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7302   int8x16_t __ret;
   7303   __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   7304   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   7305   return __ret;
   7306 }
   7307 #endif
   7308 
   7309 #ifdef __LITTLE_ENDIAN__
   7310 __ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
   7311   int32x4_t __ret;
   7312   __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   7313   return __ret;
   7314 }
   7315 #else
   7316 __ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
   7317   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   7318   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   7319   int32x4_t __ret;
   7320   __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   7321   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   7322   return __ret;
   7323 }
   7324 #endif
   7325 
   7326 #ifdef __LITTLE_ENDIAN__
   7327 __ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
   7328   int16x8_t __ret;
   7329   __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   7330   return __ret;
   7331 }
   7332 #else
   7333 __ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
   7334   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   7335   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   7336   int16x8_t __ret;
   7337   __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   7338   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   7339   return __ret;
   7340 }
   7341 #endif
   7342 
   7343 #ifdef __LITTLE_ENDIAN__
   7344 __ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
   7345   uint8x8_t __ret;
   7346   __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   7347   return __ret;
   7348 }
   7349 #else
   7350 __ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
   7351   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   7352   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   7353   uint8x8_t __ret;
   7354   __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   7355   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   7356   return __ret;
   7357 }
   7358 #endif
   7359 
   7360 #ifdef __LITTLE_ENDIAN__
   7361 __ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
   7362   uint32x2_t __ret;
   7363   __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   7364   return __ret;
   7365 }
   7366 #else
   7367 __ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
   7368   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   7369   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   7370   uint32x2_t __ret;
   7371   __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   7372   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   7373   return __ret;
   7374 }
   7375 #endif
   7376 
   7377 #ifdef __LITTLE_ENDIAN__
   7378 __ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
   7379   uint16x4_t __ret;
   7380   __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   7381   return __ret;
   7382 }
   7383 #else
   7384 __ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
   7385   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   7386   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   7387   uint16x4_t __ret;
   7388   __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   7389   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   7390   return __ret;
   7391 }
   7392 #endif
   7393 
   7394 #ifdef __LITTLE_ENDIAN__
   7395 __ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
   7396   int8x8_t __ret;
   7397   __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   7398   return __ret;
   7399 }
   7400 #else
   7401 __ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
   7402   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   7403   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   7404   int8x8_t __ret;
   7405   __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   7406   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   7407   return __ret;
   7408 }
   7409 #endif
   7410 
   7411 #ifdef __LITTLE_ENDIAN__
   7412 __ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
   7413   int32x2_t __ret;
   7414   __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   7415   return __ret;
   7416 }
   7417 #else
   7418 __ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
   7419   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   7420   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   7421   int32x2_t __ret;
   7422   __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   7423   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   7424   return __ret;
   7425 }
   7426 #endif
   7427 
   7428 #ifdef __LITTLE_ENDIAN__
   7429 __ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
   7430   int16x4_t __ret;
   7431   __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   7432   return __ret;
   7433 }
   7434 #else
   7435 __ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
   7436   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   7437   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   7438   int16x4_t __ret;
   7439   __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   7440   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   7441   return __ret;
   7442 }
   7443 #endif
   7444 
   7445 #ifdef __LITTLE_ENDIAN__
   7446 #define vld1_p8(__p0) __extension__ ({ \
   7447   poly8x8_t __ret; \
   7448   __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
   7449   __ret; \
   7450 })
   7451 #else
   7452 #define vld1_p8(__p0) __extension__ ({ \
   7453   poly8x8_t __ret; \
   7454   __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
   7455   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7456   __ret; \
   7457 })
   7458 #endif
   7459 
   7460 #ifdef __LITTLE_ENDIAN__
   7461 #define vld1_p16(__p0) __extension__ ({ \
   7462   poly16x4_t __ret; \
   7463   __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
   7464   __ret; \
   7465 })
   7466 #else
   7467 #define vld1_p16(__p0) __extension__ ({ \
   7468   poly16x4_t __ret; \
   7469   __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
   7470   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7471   __ret; \
   7472 })
   7473 #endif
   7474 
   7475 #ifdef __LITTLE_ENDIAN__
   7476 #define vld1q_p8(__p0) __extension__ ({ \
   7477   poly8x16_t __ret; \
   7478   __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
   7479   __ret; \
   7480 })
   7481 #else
   7482 #define vld1q_p8(__p0) __extension__ ({ \
   7483   poly8x16_t __ret; \
   7484   __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
   7485   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   7486   __ret; \
   7487 })
   7488 #endif
   7489 
   7490 #ifdef __LITTLE_ENDIAN__
   7491 #define vld1q_p16(__p0) __extension__ ({ \
   7492   poly16x8_t __ret; \
   7493   __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
   7494   __ret; \
   7495 })
   7496 #else
   7497 #define vld1q_p16(__p0) __extension__ ({ \
   7498   poly16x8_t __ret; \
   7499   __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
   7500   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7501   __ret; \
   7502 })
   7503 #endif
   7504 
   7505 #ifdef __LITTLE_ENDIAN__
   7506 #define vld1q_u8(__p0) __extension__ ({ \
   7507   uint8x16_t __ret; \
   7508   __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
   7509   __ret; \
   7510 })
   7511 #else
   7512 #define vld1q_u8(__p0) __extension__ ({ \
   7513   uint8x16_t __ret; \
   7514   __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
   7515   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   7516   __ret; \
   7517 })
   7518 #endif
   7519 
   7520 #ifdef __LITTLE_ENDIAN__
   7521 #define vld1q_u32(__p0) __extension__ ({ \
   7522   uint32x4_t __ret; \
   7523   __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
   7524   __ret; \
   7525 })
   7526 #else
   7527 #define vld1q_u32(__p0) __extension__ ({ \
   7528   uint32x4_t __ret; \
   7529   __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
   7530   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7531   __ret; \
   7532 })
   7533 #endif
   7534 
   7535 #ifdef __LITTLE_ENDIAN__
   7536 #define vld1q_u64(__p0) __extension__ ({ \
   7537   uint64x2_t __ret; \
   7538   __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
   7539   __ret; \
   7540 })
   7541 #else
   7542 #define vld1q_u64(__p0) __extension__ ({ \
   7543   uint64x2_t __ret; \
   7544   __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
   7545   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   7546   __ret; \
   7547 })
   7548 #endif
   7549 
   7550 #ifdef __LITTLE_ENDIAN__
   7551 #define vld1q_u16(__p0) __extension__ ({ \
   7552   uint16x8_t __ret; \
   7553   __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
   7554   __ret; \
   7555 })
   7556 #else
   7557 #define vld1q_u16(__p0) __extension__ ({ \
   7558   uint16x8_t __ret; \
   7559   __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
   7560   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7561   __ret; \
   7562 })
   7563 #endif
   7564 
   7565 #ifdef __LITTLE_ENDIAN__
   7566 #define vld1q_s8(__p0) __extension__ ({ \
   7567   int8x16_t __ret; \
   7568   __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
   7569   __ret; \
   7570 })
   7571 #else
   7572 #define vld1q_s8(__p0) __extension__ ({ \
   7573   int8x16_t __ret; \
   7574   __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
   7575   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   7576   __ret; \
   7577 })
   7578 #endif
   7579 
   7580 #ifdef __LITTLE_ENDIAN__
   7581 #define vld1q_f32(__p0) __extension__ ({ \
   7582   float32x4_t __ret; \
   7583   __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
   7584   __ret; \
   7585 })
   7586 #else
   7587 #define vld1q_f32(__p0) __extension__ ({ \
   7588   float32x4_t __ret; \
   7589   __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
   7590   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7591   __ret; \
   7592 })
   7593 #endif
   7594 
   7595 #ifdef __LITTLE_ENDIAN__
   7596 #define vld1q_f16(__p0) __extension__ ({ \
   7597   float16x8_t __ret; \
   7598   __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
   7599   __ret; \
   7600 })
   7601 #else
   7602 #define vld1q_f16(__p0) __extension__ ({ \
   7603   float16x8_t __ret; \
   7604   __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
   7605   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7606   __ret; \
   7607 })
   7608 #endif
   7609 
   7610 #ifdef __LITTLE_ENDIAN__
   7611 #define vld1q_s32(__p0) __extension__ ({ \
   7612   int32x4_t __ret; \
   7613   __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
   7614   __ret; \
   7615 })
   7616 #else
   7617 #define vld1q_s32(__p0) __extension__ ({ \
   7618   int32x4_t __ret; \
   7619   __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
   7620   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7621   __ret; \
   7622 })
   7623 #endif
   7624 
   7625 #ifdef __LITTLE_ENDIAN__
   7626 #define vld1q_s64(__p0) __extension__ ({ \
   7627   int64x2_t __ret; \
   7628   __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
   7629   __ret; \
   7630 })
   7631 #else
   7632 #define vld1q_s64(__p0) __extension__ ({ \
   7633   int64x2_t __ret; \
   7634   __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
   7635   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   7636   __ret; \
   7637 })
   7638 #endif
   7639 
   7640 #ifdef __LITTLE_ENDIAN__
   7641 #define vld1q_s16(__p0) __extension__ ({ \
   7642   int16x8_t __ret; \
   7643   __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
   7644   __ret; \
   7645 })
   7646 #else
   7647 #define vld1q_s16(__p0) __extension__ ({ \
   7648   int16x8_t __ret; \
   7649   __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
   7650   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7651   __ret; \
   7652 })
   7653 #endif
   7654 
   7655 #ifdef __LITTLE_ENDIAN__
   7656 #define vld1_u8(__p0) __extension__ ({ \
   7657   uint8x8_t __ret; \
   7658   __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
   7659   __ret; \
   7660 })
   7661 #else
   7662 #define vld1_u8(__p0) __extension__ ({ \
   7663   uint8x8_t __ret; \
   7664   __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
   7665   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7666   __ret; \
   7667 })
   7668 #endif
   7669 
   7670 #ifdef __LITTLE_ENDIAN__
   7671 #define vld1_u32(__p0) __extension__ ({ \
   7672   uint32x2_t __ret; \
   7673   __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
   7674   __ret; \
   7675 })
   7676 #else
   7677 #define vld1_u32(__p0) __extension__ ({ \
   7678   uint32x2_t __ret; \
   7679   __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
   7680   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   7681   __ret; \
   7682 })
   7683 #endif
   7684 
   7685 #ifdef __LITTLE_ENDIAN__
   7686 #define vld1_u64(__p0) __extension__ ({ \
   7687   uint64x1_t __ret; \
   7688   __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
   7689   __ret; \
   7690 })
   7691 #else
   7692 #define vld1_u64(__p0) __extension__ ({ \
   7693   uint64x1_t __ret; \
   7694   __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
   7695   __ret; \
   7696 })
   7697 #endif
   7698 
   7699 #ifdef __LITTLE_ENDIAN__
   7700 #define vld1_u16(__p0) __extension__ ({ \
   7701   uint16x4_t __ret; \
   7702   __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
   7703   __ret; \
   7704 })
   7705 #else
   7706 #define vld1_u16(__p0) __extension__ ({ \
   7707   uint16x4_t __ret; \
   7708   __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
   7709   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7710   __ret; \
   7711 })
   7712 #endif
   7713 
   7714 #ifdef __LITTLE_ENDIAN__
   7715 #define vld1_s8(__p0) __extension__ ({ \
   7716   int8x8_t __ret; \
   7717   __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
   7718   __ret; \
   7719 })
   7720 #else
   7721 #define vld1_s8(__p0) __extension__ ({ \
   7722   int8x8_t __ret; \
   7723   __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
   7724   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7725   __ret; \
   7726 })
   7727 #endif
   7728 
   7729 #ifdef __LITTLE_ENDIAN__
   7730 #define vld1_f32(__p0) __extension__ ({ \
   7731   float32x2_t __ret; \
   7732   __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
   7733   __ret; \
   7734 })
   7735 #else
   7736 #define vld1_f32(__p0) __extension__ ({ \
   7737   float32x2_t __ret; \
   7738   __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
   7739   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   7740   __ret; \
   7741 })
   7742 #endif
   7743 
   7744 #ifdef __LITTLE_ENDIAN__
   7745 #define vld1_f16(__p0) __extension__ ({ \
   7746   float16x4_t __ret; \
   7747   __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
   7748   __ret; \
   7749 })
   7750 #else
   7751 #define vld1_f16(__p0) __extension__ ({ \
   7752   float16x4_t __ret; \
   7753   __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
   7754   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7755   __ret; \
   7756 })
   7757 #endif
   7758 
   7759 #ifdef __LITTLE_ENDIAN__
   7760 #define vld1_s32(__p0) __extension__ ({ \
   7761   int32x2_t __ret; \
   7762   __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
   7763   __ret; \
   7764 })
   7765 #else
   7766 #define vld1_s32(__p0) __extension__ ({ \
   7767   int32x2_t __ret; \
   7768   __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
   7769   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   7770   __ret; \
   7771 })
   7772 #endif
   7773 
   7774 #ifdef __LITTLE_ENDIAN__
   7775 #define vld1_s64(__p0) __extension__ ({ \
   7776   int64x1_t __ret; \
   7777   __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
   7778   __ret; \
   7779 })
   7780 #else
   7781 #define vld1_s64(__p0) __extension__ ({ \
   7782   int64x1_t __ret; \
   7783   __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
   7784   __ret; \
   7785 })
   7786 #endif
   7787 
   7788 #ifdef __LITTLE_ENDIAN__
   7789 #define vld1_s16(__p0) __extension__ ({ \
   7790   int16x4_t __ret; \
   7791   __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
   7792   __ret; \
   7793 })
   7794 #else
   7795 #define vld1_s16(__p0) __extension__ ({ \
   7796   int16x4_t __ret; \
   7797   __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
   7798   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7799   __ret; \
   7800 })
   7801 #endif
   7802 
   7803 #ifdef __LITTLE_ENDIAN__
   7804 #define vld1_dup_p8(__p0) __extension__ ({ \
   7805   poly8x8_t __ret; \
   7806   __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
   7807   __ret; \
   7808 })
   7809 #else
   7810 #define vld1_dup_p8(__p0) __extension__ ({ \
   7811   poly8x8_t __ret; \
   7812   __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
   7813   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7814   __ret; \
   7815 })
   7816 #endif
   7817 
   7818 #ifdef __LITTLE_ENDIAN__
   7819 #define vld1_dup_p16(__p0) __extension__ ({ \
   7820   poly16x4_t __ret; \
   7821   __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
   7822   __ret; \
   7823 })
   7824 #else
   7825 #define vld1_dup_p16(__p0) __extension__ ({ \
   7826   poly16x4_t __ret; \
   7827   __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
   7828   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7829   __ret; \
   7830 })
   7831 #endif
   7832 
   7833 #ifdef __LITTLE_ENDIAN__
   7834 #define vld1q_dup_p8(__p0) __extension__ ({ \
   7835   poly8x16_t __ret; \
   7836   __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
   7837   __ret; \
   7838 })
   7839 #else
   7840 #define vld1q_dup_p8(__p0) __extension__ ({ \
   7841   poly8x16_t __ret; \
   7842   __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
   7843   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   7844   __ret; \
   7845 })
   7846 #endif
   7847 
   7848 #ifdef __LITTLE_ENDIAN__
   7849 #define vld1q_dup_p16(__p0) __extension__ ({ \
   7850   poly16x8_t __ret; \
   7851   __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
   7852   __ret; \
   7853 })
   7854 #else
   7855 #define vld1q_dup_p16(__p0) __extension__ ({ \
   7856   poly16x8_t __ret; \
   7857   __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
   7858   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7859   __ret; \
   7860 })
   7861 #endif
   7862 
   7863 #ifdef __LITTLE_ENDIAN__
   7864 #define vld1q_dup_u8(__p0) __extension__ ({ \
   7865   uint8x16_t __ret; \
   7866   __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
   7867   __ret; \
   7868 })
   7869 #else
   7870 #define vld1q_dup_u8(__p0) __extension__ ({ \
   7871   uint8x16_t __ret; \
   7872   __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
   7873   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   7874   __ret; \
   7875 })
   7876 #endif
   7877 
   7878 #ifdef __LITTLE_ENDIAN__
   7879 #define vld1q_dup_u32(__p0) __extension__ ({ \
   7880   uint32x4_t __ret; \
   7881   __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
   7882   __ret; \
   7883 })
   7884 #else
   7885 #define vld1q_dup_u32(__p0) __extension__ ({ \
   7886   uint32x4_t __ret; \
   7887   __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
   7888   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7889   __ret; \
   7890 })
   7891 #endif
   7892 
   7893 #ifdef __LITTLE_ENDIAN__
   7894 #define vld1q_dup_u64(__p0) __extension__ ({ \
   7895   uint64x2_t __ret; \
   7896   __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
   7897   __ret; \
   7898 })
   7899 #else
   7900 #define vld1q_dup_u64(__p0) __extension__ ({ \
   7901   uint64x2_t __ret; \
   7902   __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
   7903   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   7904   __ret; \
   7905 })
   7906 #endif
   7907 
   7908 #ifdef __LITTLE_ENDIAN__
   7909 #define vld1q_dup_u16(__p0) __extension__ ({ \
   7910   uint16x8_t __ret; \
   7911   __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
   7912   __ret; \
   7913 })
   7914 #else
   7915 #define vld1q_dup_u16(__p0) __extension__ ({ \
   7916   uint16x8_t __ret; \
   7917   __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
   7918   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7919   __ret; \
   7920 })
   7921 #endif
   7922 
   7923 #ifdef __LITTLE_ENDIAN__
   7924 #define vld1q_dup_s8(__p0) __extension__ ({ \
   7925   int8x16_t __ret; \
   7926   __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
   7927   __ret; \
   7928 })
   7929 #else
   7930 #define vld1q_dup_s8(__p0) __extension__ ({ \
   7931   int8x16_t __ret; \
   7932   __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
   7933   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   7934   __ret; \
   7935 })
   7936 #endif
   7937 
   7938 #ifdef __LITTLE_ENDIAN__
   7939 #define vld1q_dup_f32(__p0) __extension__ ({ \
   7940   float32x4_t __ret; \
   7941   __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
   7942   __ret; \
   7943 })
   7944 #else
   7945 #define vld1q_dup_f32(__p0) __extension__ ({ \
   7946   float32x4_t __ret; \
   7947   __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
   7948   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7949   __ret; \
   7950 })
   7951 #endif
   7952 
   7953 #ifdef __LITTLE_ENDIAN__
   7954 #define vld1q_dup_f16(__p0) __extension__ ({ \
   7955   float16x8_t __ret; \
   7956   __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
   7957   __ret; \
   7958 })
   7959 #else
   7960 #define vld1q_dup_f16(__p0) __extension__ ({ \
   7961   float16x8_t __ret; \
   7962   __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
   7963   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   7964   __ret; \
   7965 })
   7966 #endif
   7967 
   7968 #ifdef __LITTLE_ENDIAN__
   7969 #define vld1q_dup_s32(__p0) __extension__ ({ \
   7970   int32x4_t __ret; \
   7971   __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
   7972   __ret; \
   7973 })
   7974 #else
   7975 #define vld1q_dup_s32(__p0) __extension__ ({ \
   7976   int32x4_t __ret; \
   7977   __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
   7978   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   7979   __ret; \
   7980 })
   7981 #endif
   7982 
   7983 #ifdef __LITTLE_ENDIAN__
   7984 #define vld1q_dup_s64(__p0) __extension__ ({ \
   7985   int64x2_t __ret; \
   7986   __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
   7987   __ret; \
   7988 })
   7989 #else
   7990 #define vld1q_dup_s64(__p0) __extension__ ({ \
   7991   int64x2_t __ret; \
   7992   __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
   7993   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   7994   __ret; \
   7995 })
   7996 #endif
   7997 
   7998 #ifdef __LITTLE_ENDIAN__
   7999 #define vld1q_dup_s16(__p0) __extension__ ({ \
   8000   int16x8_t __ret; \
   8001   __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
   8002   __ret; \
   8003 })
   8004 #else
   8005 #define vld1q_dup_s16(__p0) __extension__ ({ \
   8006   int16x8_t __ret; \
   8007   __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
   8008   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8009   __ret; \
   8010 })
   8011 #endif
   8012 
   8013 #ifdef __LITTLE_ENDIAN__
   8014 #define vld1_dup_u8(__p0) __extension__ ({ \
   8015   uint8x8_t __ret; \
   8016   __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
   8017   __ret; \
   8018 })
   8019 #else
   8020 #define vld1_dup_u8(__p0) __extension__ ({ \
   8021   uint8x8_t __ret; \
   8022   __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
   8023   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8024   __ret; \
   8025 })
   8026 #endif
   8027 
   8028 #ifdef __LITTLE_ENDIAN__
   8029 #define vld1_dup_u32(__p0) __extension__ ({ \
   8030   uint32x2_t __ret; \
   8031   __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
   8032   __ret; \
   8033 })
   8034 #else
   8035 #define vld1_dup_u32(__p0) __extension__ ({ \
   8036   uint32x2_t __ret; \
   8037   __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
   8038   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   8039   __ret; \
   8040 })
   8041 #endif
   8042 
   8043 #ifdef __LITTLE_ENDIAN__
   8044 #define vld1_dup_u64(__p0) __extension__ ({ \
   8045   uint64x1_t __ret; \
   8046   __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
   8047   __ret; \
   8048 })
   8049 #else
   8050 #define vld1_dup_u64(__p0) __extension__ ({ \
   8051   uint64x1_t __ret; \
   8052   __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
   8053   __ret; \
   8054 })
   8055 #endif
   8056 
   8057 #ifdef __LITTLE_ENDIAN__
   8058 #define vld1_dup_u16(__p0) __extension__ ({ \
   8059   uint16x4_t __ret; \
   8060   __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
   8061   __ret; \
   8062 })
   8063 #else
   8064 #define vld1_dup_u16(__p0) __extension__ ({ \
   8065   uint16x4_t __ret; \
   8066   __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
   8067   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8068   __ret; \
   8069 })
   8070 #endif
   8071 
   8072 #ifdef __LITTLE_ENDIAN__
   8073 #define vld1_dup_s8(__p0) __extension__ ({ \
   8074   int8x8_t __ret; \
   8075   __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
   8076   __ret; \
   8077 })
   8078 #else
   8079 #define vld1_dup_s8(__p0) __extension__ ({ \
   8080   int8x8_t __ret; \
   8081   __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
   8082   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8083   __ret; \
   8084 })
   8085 #endif
   8086 
   8087 #ifdef __LITTLE_ENDIAN__
   8088 #define vld1_dup_f32(__p0) __extension__ ({ \
   8089   float32x2_t __ret; \
   8090   __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
   8091   __ret; \
   8092 })
   8093 #else
   8094 #define vld1_dup_f32(__p0) __extension__ ({ \
   8095   float32x2_t __ret; \
   8096   __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
   8097   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   8098   __ret; \
   8099 })
   8100 #endif
   8101 
   8102 #ifdef __LITTLE_ENDIAN__
   8103 #define vld1_dup_f16(__p0) __extension__ ({ \
   8104   float16x4_t __ret; \
   8105   __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
   8106   __ret; \
   8107 })
   8108 #else
   8109 #define vld1_dup_f16(__p0) __extension__ ({ \
   8110   float16x4_t __ret; \
   8111   __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
   8112   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8113   __ret; \
   8114 })
   8115 #endif
   8116 
   8117 #ifdef __LITTLE_ENDIAN__
   8118 #define vld1_dup_s32(__p0) __extension__ ({ \
   8119   int32x2_t __ret; \
   8120   __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
   8121   __ret; \
   8122 })
   8123 #else
   8124 #define vld1_dup_s32(__p0) __extension__ ({ \
   8125   int32x2_t __ret; \
   8126   __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
   8127   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   8128   __ret; \
   8129 })
   8130 #endif
   8131 
   8132 #ifdef __LITTLE_ENDIAN__
   8133 #define vld1_dup_s64(__p0) __extension__ ({ \
   8134   int64x1_t __ret; \
   8135   __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
   8136   __ret; \
   8137 })
   8138 #else
   8139 #define vld1_dup_s64(__p0) __extension__ ({ \
   8140   int64x1_t __ret; \
   8141   __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
   8142   __ret; \
   8143 })
   8144 #endif
   8145 
   8146 #ifdef __LITTLE_ENDIAN__
   8147 #define vld1_dup_s16(__p0) __extension__ ({ \
   8148   int16x4_t __ret; \
   8149   __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
   8150   __ret; \
   8151 })
   8152 #else
   8153 #define vld1_dup_s16(__p0) __extension__ ({ \
   8154   int16x4_t __ret; \
   8155   __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
   8156   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8157   __ret; \
   8158 })
   8159 #endif
   8160 
   8161 #ifdef __LITTLE_ENDIAN__
   8162 #define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   8163   poly8x8_t __s1 = __p1; \
   8164   poly8x8_t __ret; \
   8165   __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
   8166   __ret; \
   8167 })
   8168 #else
   8169 #define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   8170   poly8x8_t __s1 = __p1; \
   8171   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   8172   poly8x8_t __ret; \
   8173   __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
   8174   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8175   __ret; \
   8176 })
   8177 #endif
   8178 
   8179 #ifdef __LITTLE_ENDIAN__
   8180 #define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   8181   poly16x4_t __s1 = __p1; \
   8182   poly16x4_t __ret; \
   8183   __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
   8184   __ret; \
   8185 })
   8186 #else
   8187 #define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   8188   poly16x4_t __s1 = __p1; \
   8189   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   8190   poly16x4_t __ret; \
   8191   __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
   8192   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8193   __ret; \
   8194 })
   8195 #endif
   8196 
   8197 #ifdef __LITTLE_ENDIAN__
   8198 #define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   8199   poly8x16_t __s1 = __p1; \
   8200   poly8x16_t __ret; \
   8201   __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
   8202   __ret; \
   8203 })
   8204 #else
   8205 #define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   8206   poly8x16_t __s1 = __p1; \
   8207   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8208   poly8x16_t __ret; \
   8209   __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
   8210   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8211   __ret; \
   8212 })
   8213 #endif
   8214 
   8215 #ifdef __LITTLE_ENDIAN__
   8216 #define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   8217   poly16x8_t __s1 = __p1; \
   8218   poly16x8_t __ret; \
   8219   __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
   8220   __ret; \
   8221 })
   8222 #else
   8223 #define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   8224   poly16x8_t __s1 = __p1; \
   8225   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   8226   poly16x8_t __ret; \
   8227   __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
   8228   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8229   __ret; \
   8230 })
   8231 #endif
   8232 
   8233 #ifdef __LITTLE_ENDIAN__
   8234 #define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   8235   uint8x16_t __s1 = __p1; \
   8236   uint8x16_t __ret; \
   8237   __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
   8238   __ret; \
   8239 })
   8240 #else
   8241 #define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   8242   uint8x16_t __s1 = __p1; \
   8243   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8244   uint8x16_t __ret; \
   8245   __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
   8246   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8247   __ret; \
   8248 })
   8249 #endif
   8250 
   8251 #ifdef __LITTLE_ENDIAN__
   8252 #define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   8253   uint32x4_t __s1 = __p1; \
   8254   uint32x4_t __ret; \
   8255   __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
   8256   __ret; \
   8257 })
   8258 #else
   8259 #define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   8260   uint32x4_t __s1 = __p1; \
   8261   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   8262   uint32x4_t __ret; \
   8263   __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
   8264   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8265   __ret; \
   8266 })
   8267 #endif
   8268 
   8269 #ifdef __LITTLE_ENDIAN__
   8270 #define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   8271   uint64x2_t __s1 = __p1; \
   8272   uint64x2_t __ret; \
   8273   __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
   8274   __ret; \
   8275 })
   8276 #else
   8277 #define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   8278   uint64x2_t __s1 = __p1; \
   8279   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   8280   uint64x2_t __ret; \
   8281   __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
   8282   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   8283   __ret; \
   8284 })
   8285 #endif
   8286 
   8287 #ifdef __LITTLE_ENDIAN__
   8288 #define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   8289   uint16x8_t __s1 = __p1; \
   8290   uint16x8_t __ret; \
   8291   __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
   8292   __ret; \
   8293 })
   8294 #else
   8295 #define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   8296   uint16x8_t __s1 = __p1; \
   8297   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   8298   uint16x8_t __ret; \
   8299   __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
   8300   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8301   __ret; \
   8302 })
   8303 #endif
   8304 
   8305 #ifdef __LITTLE_ENDIAN__
   8306 #define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   8307   int8x16_t __s1 = __p1; \
   8308   int8x16_t __ret; \
   8309   __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
   8310   __ret; \
   8311 })
   8312 #else
   8313 #define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   8314   int8x16_t __s1 = __p1; \
   8315   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8316   int8x16_t __ret; \
   8317   __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
   8318   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8319   __ret; \
   8320 })
   8321 #endif
   8322 
   8323 #ifdef __LITTLE_ENDIAN__
   8324 #define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   8325   float32x4_t __s1 = __p1; \
   8326   float32x4_t __ret; \
   8327   __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
   8328   __ret; \
   8329 })
   8330 #else
   8331 #define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   8332   float32x4_t __s1 = __p1; \
   8333   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   8334   float32x4_t __ret; \
   8335   __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
   8336   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8337   __ret; \
   8338 })
   8339 #endif
   8340 
   8341 #ifdef __LITTLE_ENDIAN__
   8342 #define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   8343   float16x8_t __s1 = __p1; \
   8344   float16x8_t __ret; \
   8345   __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
   8346   __ret; \
   8347 })
   8348 #else
   8349 #define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   8350   float16x8_t __s1 = __p1; \
   8351   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   8352   float16x8_t __ret; \
   8353   __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
   8354   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8355   __ret; \
   8356 })
   8357 #endif
   8358 
   8359 #ifdef __LITTLE_ENDIAN__
   8360 #define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   8361   int32x4_t __s1 = __p1; \
   8362   int32x4_t __ret; \
   8363   __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
   8364   __ret; \
   8365 })
   8366 #else
   8367 #define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   8368   int32x4_t __s1 = __p1; \
   8369   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   8370   int32x4_t __ret; \
   8371   __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
   8372   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8373   __ret; \
   8374 })
   8375 #endif
   8376 
   8377 #ifdef __LITTLE_ENDIAN__
   8378 #define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   8379   int64x2_t __s1 = __p1; \
   8380   int64x2_t __ret; \
   8381   __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
   8382   __ret; \
   8383 })
   8384 #else
   8385 #define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   8386   int64x2_t __s1 = __p1; \
   8387   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   8388   int64x2_t __ret; \
   8389   __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
   8390   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   8391   __ret; \
   8392 })
   8393 #endif
   8394 
   8395 #ifdef __LITTLE_ENDIAN__
   8396 #define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   8397   int16x8_t __s1 = __p1; \
   8398   int16x8_t __ret; \
   8399   __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
   8400   __ret; \
   8401 })
   8402 #else
   8403 #define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   8404   int16x8_t __s1 = __p1; \
   8405   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   8406   int16x8_t __ret; \
   8407   __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
   8408   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8409   __ret; \
   8410 })
   8411 #endif
   8412 
   8413 #ifdef __LITTLE_ENDIAN__
   8414 #define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   8415   uint8x8_t __s1 = __p1; \
   8416   uint8x8_t __ret; \
   8417   __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
   8418   __ret; \
   8419 })
   8420 #else
   8421 #define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   8422   uint8x8_t __s1 = __p1; \
   8423   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   8424   uint8x8_t __ret; \
   8425   __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
   8426   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8427   __ret; \
   8428 })
   8429 #endif
   8430 
   8431 #ifdef __LITTLE_ENDIAN__
   8432 #define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   8433   uint32x2_t __s1 = __p1; \
   8434   uint32x2_t __ret; \
   8435   __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
   8436   __ret; \
   8437 })
   8438 #else
   8439 #define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   8440   uint32x2_t __s1 = __p1; \
   8441   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   8442   uint32x2_t __ret; \
   8443   __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
   8444   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   8445   __ret; \
   8446 })
   8447 #endif
   8448 
   8449 #ifdef __LITTLE_ENDIAN__
   8450 #define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   8451   uint64x1_t __s1 = __p1; \
   8452   uint64x1_t __ret; \
   8453   __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
   8454   __ret; \
   8455 })
   8456 #else
   8457 #define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   8458   uint64x1_t __s1 = __p1; \
   8459   uint64x1_t __ret; \
   8460   __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
   8461   __ret; \
   8462 })
   8463 #endif
   8464 
   8465 #ifdef __LITTLE_ENDIAN__
   8466 #define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   8467   uint16x4_t __s1 = __p1; \
   8468   uint16x4_t __ret; \
   8469   __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
   8470   __ret; \
   8471 })
   8472 #else
   8473 #define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   8474   uint16x4_t __s1 = __p1; \
   8475   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   8476   uint16x4_t __ret; \
   8477   __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
   8478   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8479   __ret; \
   8480 })
   8481 #endif
   8482 
   8483 #ifdef __LITTLE_ENDIAN__
   8484 #define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   8485   int8x8_t __s1 = __p1; \
   8486   int8x8_t __ret; \
   8487   __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
   8488   __ret; \
   8489 })
   8490 #else
   8491 #define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   8492   int8x8_t __s1 = __p1; \
   8493   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   8494   int8x8_t __ret; \
   8495   __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
   8496   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   8497   __ret; \
   8498 })
   8499 #endif
   8500 
   8501 #ifdef __LITTLE_ENDIAN__
   8502 #define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   8503   float32x2_t __s1 = __p1; \
   8504   float32x2_t __ret; \
   8505   __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
   8506   __ret; \
   8507 })
   8508 #else
   8509 #define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   8510   float32x2_t __s1 = __p1; \
   8511   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   8512   float32x2_t __ret; \
   8513   __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
   8514   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   8515   __ret; \
   8516 })
   8517 #endif
   8518 
   8519 #ifdef __LITTLE_ENDIAN__
   8520 #define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   8521   float16x4_t __s1 = __p1; \
   8522   float16x4_t __ret; \
   8523   __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
   8524   __ret; \
   8525 })
   8526 #else
   8527 #define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   8528   float16x4_t __s1 = __p1; \
   8529   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   8530   float16x4_t __ret; \
   8531   __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
   8532   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8533   __ret; \
   8534 })
   8535 #endif
   8536 
   8537 #ifdef __LITTLE_ENDIAN__
   8538 #define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   8539   int32x2_t __s1 = __p1; \
   8540   int32x2_t __ret; \
   8541   __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
   8542   __ret; \
   8543 })
   8544 #else
   8545 #define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   8546   int32x2_t __s1 = __p1; \
   8547   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   8548   int32x2_t __ret; \
   8549   __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
   8550   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   8551   __ret; \
   8552 })
   8553 #endif
   8554 
   8555 #ifdef __LITTLE_ENDIAN__
   8556 #define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   8557   int64x1_t __s1 = __p1; \
   8558   int64x1_t __ret; \
   8559   __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
   8560   __ret; \
   8561 })
   8562 #else
   8563 #define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   8564   int64x1_t __s1 = __p1; \
   8565   int64x1_t __ret; \
   8566   __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
   8567   __ret; \
   8568 })
   8569 #endif
   8570 
   8571 #ifdef __LITTLE_ENDIAN__
   8572 #define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   8573   int16x4_t __s1 = __p1; \
   8574   int16x4_t __ret; \
   8575   __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
   8576   __ret; \
   8577 })
   8578 #else
   8579 #define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   8580   int16x4_t __s1 = __p1; \
   8581   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   8582   int16x4_t __ret; \
   8583   __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
   8584   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   8585   __ret; \
   8586 })
   8587 #endif
   8588 
   8589 #ifdef __LITTLE_ENDIAN__
   8590 #define vld2_p8(__p0) __extension__ ({ \
   8591   poly8x8x2_t __ret; \
   8592   __builtin_neon_vld2_v(&__ret, __p0, 4); \
   8593   __ret; \
   8594 })
   8595 #else
   8596 #define vld2_p8(__p0) __extension__ ({ \
   8597   poly8x8x2_t __ret; \
   8598   __builtin_neon_vld2_v(&__ret, __p0, 4); \
   8599  \
   8600   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   8601   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   8602   __ret; \
   8603 })
   8604 #endif
   8605 
   8606 #ifdef __LITTLE_ENDIAN__
   8607 #define vld2_p16(__p0) __extension__ ({ \
   8608   poly16x4x2_t __ret; \
   8609   __builtin_neon_vld2_v(&__ret, __p0, 5); \
   8610   __ret; \
   8611 })
   8612 #else
   8613 #define vld2_p16(__p0) __extension__ ({ \
   8614   poly16x4x2_t __ret; \
   8615   __builtin_neon_vld2_v(&__ret, __p0, 5); \
   8616  \
   8617   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   8618   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   8619   __ret; \
   8620 })
   8621 #endif
   8622 
   8623 #ifdef __LITTLE_ENDIAN__
   8624 #define vld2q_p8(__p0) __extension__ ({ \
   8625   poly8x16x2_t __ret; \
   8626   __builtin_neon_vld2q_v(&__ret, __p0, 36); \
   8627   __ret; \
   8628 })
   8629 #else
   8630 #define vld2q_p8(__p0) __extension__ ({ \
   8631   poly8x16x2_t __ret; \
   8632   __builtin_neon_vld2q_v(&__ret, __p0, 36); \
   8633  \
   8634   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8635   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8636   __ret; \
   8637 })
   8638 #endif
   8639 
   8640 #ifdef __LITTLE_ENDIAN__
   8641 #define vld2q_p16(__p0) __extension__ ({ \
   8642   poly16x8x2_t __ret; \
   8643   __builtin_neon_vld2q_v(&__ret, __p0, 37); \
   8644   __ret; \
   8645 })
   8646 #else
   8647 #define vld2q_p16(__p0) __extension__ ({ \
   8648   poly16x8x2_t __ret; \
   8649   __builtin_neon_vld2q_v(&__ret, __p0, 37); \
   8650  \
   8651   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   8652   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   8653   __ret; \
   8654 })
   8655 #endif
   8656 
   8657 #ifdef __LITTLE_ENDIAN__
   8658 #define vld2q_u8(__p0) __extension__ ({ \
   8659   uint8x16x2_t __ret; \
   8660   __builtin_neon_vld2q_v(&__ret, __p0, 48); \
   8661   __ret; \
   8662 })
   8663 #else
   8664 #define vld2q_u8(__p0) __extension__ ({ \
   8665   uint8x16x2_t __ret; \
   8666   __builtin_neon_vld2q_v(&__ret, __p0, 48); \
   8667  \
   8668   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8669   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8670   __ret; \
   8671 })
   8672 #endif
   8673 
   8674 #ifdef __LITTLE_ENDIAN__
   8675 #define vld2q_u32(__p0) __extension__ ({ \
   8676   uint32x4x2_t __ret; \
   8677   __builtin_neon_vld2q_v(&__ret, __p0, 50); \
   8678   __ret; \
   8679 })
   8680 #else
   8681 #define vld2q_u32(__p0) __extension__ ({ \
   8682   uint32x4x2_t __ret; \
   8683   __builtin_neon_vld2q_v(&__ret, __p0, 50); \
   8684  \
   8685   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   8686   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   8687   __ret; \
   8688 })
   8689 #endif
   8690 
   8691 #ifdef __LITTLE_ENDIAN__
   8692 #define vld2q_u16(__p0) __extension__ ({ \
   8693   uint16x8x2_t __ret; \
   8694   __builtin_neon_vld2q_v(&__ret, __p0, 49); \
   8695   __ret; \
   8696 })
   8697 #else
   8698 #define vld2q_u16(__p0) __extension__ ({ \
   8699   uint16x8x2_t __ret; \
   8700   __builtin_neon_vld2q_v(&__ret, __p0, 49); \
   8701  \
   8702   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   8703   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   8704   __ret; \
   8705 })
   8706 #endif
   8707 
   8708 #ifdef __LITTLE_ENDIAN__
   8709 #define vld2q_s8(__p0) __extension__ ({ \
   8710   int8x16x2_t __ret; \
   8711   __builtin_neon_vld2q_v(&__ret, __p0, 32); \
   8712   __ret; \
   8713 })
   8714 #else
   8715 #define vld2q_s8(__p0) __extension__ ({ \
   8716   int8x16x2_t __ret; \
   8717   __builtin_neon_vld2q_v(&__ret, __p0, 32); \
   8718  \
   8719   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8720   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   8721   __ret; \
   8722 })
   8723 #endif
   8724 
   8725 #ifdef __LITTLE_ENDIAN__
   8726 #define vld2q_f32(__p0) __extension__ ({ \
   8727   float32x4x2_t __ret; \
   8728   __builtin_neon_vld2q_v(&__ret, __p0, 41); \
   8729   __ret; \
   8730 })
   8731 #else
   8732 #define vld2q_f32(__p0) __extension__ ({ \
   8733   float32x4x2_t __ret; \
   8734   __builtin_neon_vld2q_v(&__ret, __p0, 41); \
   8735  \
   8736   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   8737   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   8738   __ret; \
   8739 })
   8740 #endif
   8741 
   8742 #ifdef __LITTLE_ENDIAN__
   8743 #define vld2q_f16(__p0) __extension__ ({ \
   8744   float16x8x2_t __ret; \
   8745   __builtin_neon_vld2q_v(&__ret, __p0, 40); \
   8746   __ret; \
   8747 })
   8748 #else
   8749 #define vld2q_f16(__p0) __extension__ ({ \
   8750   float16x8x2_t __ret; \
   8751   __builtin_neon_vld2q_v(&__ret, __p0, 40); \
   8752  \
   8753   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   8754   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   8755   __ret; \
   8756 })
   8757 #endif
   8758 
   8759 #ifdef __LITTLE_ENDIAN__
   8760 #define vld2q_s32(__p0) __extension__ ({ \
   8761   int32x4x2_t __ret; \
   8762   __builtin_neon_vld2q_v(&__ret, __p0, 34); \
   8763   __ret; \
   8764 })
   8765 #else
   8766 #define vld2q_s32(__p0) __extension__ ({ \
   8767   int32x4x2_t __ret; \
   8768   __builtin_neon_vld2q_v(&__ret, __p0, 34); \
   8769  \
   8770   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   8771   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   8772   __ret; \
   8773 })
   8774 #endif
   8775 
   8776 #ifdef __LITTLE_ENDIAN__
   8777 #define vld2q_s16(__p0) __extension__ ({ \
   8778   int16x8x2_t __ret; \
   8779   __builtin_neon_vld2q_v(&__ret, __p0, 33); \
   8780   __ret; \
   8781 })
   8782 #else
   8783 #define vld2q_s16(__p0) __extension__ ({ \
   8784   int16x8x2_t __ret; \
   8785   __builtin_neon_vld2q_v(&__ret, __p0, 33); \
   8786  \
   8787   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   8788   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   8789   __ret; \
   8790 })
   8791 #endif
   8792 
   8793 #ifdef __LITTLE_ENDIAN__
   8794 #define vld2_u8(__p0) __extension__ ({ \
   8795   uint8x8x2_t __ret; \
   8796   __builtin_neon_vld2_v(&__ret, __p0, 16); \
   8797   __ret; \
   8798 })
   8799 #else
   8800 #define vld2_u8(__p0) __extension__ ({ \
   8801   uint8x8x2_t __ret; \
   8802   __builtin_neon_vld2_v(&__ret, __p0, 16); \
   8803  \
   8804   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   8805   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   8806   __ret; \
   8807 })
   8808 #endif
   8809 
   8810 #ifdef __LITTLE_ENDIAN__
   8811 #define vld2_u32(__p0) __extension__ ({ \
   8812   uint32x2x2_t __ret; \
   8813   __builtin_neon_vld2_v(&__ret, __p0, 18); \
   8814   __ret; \
   8815 })
   8816 #else
   8817 #define vld2_u32(__p0) __extension__ ({ \
   8818   uint32x2x2_t __ret; \
   8819   __builtin_neon_vld2_v(&__ret, __p0, 18); \
   8820  \
   8821   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   8822   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   8823   __ret; \
   8824 })
   8825 #endif
   8826 
   8827 #ifdef __LITTLE_ENDIAN__
   8828 #define vld2_u64(__p0) __extension__ ({ \
   8829   uint64x1x2_t __ret; \
   8830   __builtin_neon_vld2_v(&__ret, __p0, 19); \
   8831   __ret; \
   8832 })
   8833 #else
   8834 #define vld2_u64(__p0) __extension__ ({ \
   8835   uint64x1x2_t __ret; \
   8836   __builtin_neon_vld2_v(&__ret, __p0, 19); \
   8837   __ret; \
   8838 })
   8839 #endif
   8840 
   8841 #ifdef __LITTLE_ENDIAN__
   8842 #define vld2_u16(__p0) __extension__ ({ \
   8843   uint16x4x2_t __ret; \
   8844   __builtin_neon_vld2_v(&__ret, __p0, 17); \
   8845   __ret; \
   8846 })
   8847 #else
   8848 #define vld2_u16(__p0) __extension__ ({ \
   8849   uint16x4x2_t __ret; \
   8850   __builtin_neon_vld2_v(&__ret, __p0, 17); \
   8851  \
   8852   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   8853   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   8854   __ret; \
   8855 })
   8856 #endif
   8857 
   8858 #ifdef __LITTLE_ENDIAN__
   8859 #define vld2_s8(__p0) __extension__ ({ \
   8860   int8x8x2_t __ret; \
   8861   __builtin_neon_vld2_v(&__ret, __p0, 0); \
   8862   __ret; \
   8863 })
   8864 #else
   8865 #define vld2_s8(__p0) __extension__ ({ \
   8866   int8x8x2_t __ret; \
   8867   __builtin_neon_vld2_v(&__ret, __p0, 0); \
   8868  \
   8869   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   8870   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   8871   __ret; \
   8872 })
   8873 #endif
   8874 
   8875 #ifdef __LITTLE_ENDIAN__
   8876 #define vld2_f32(__p0) __extension__ ({ \
   8877   float32x2x2_t __ret; \
   8878   __builtin_neon_vld2_v(&__ret, __p0, 9); \
   8879   __ret; \
   8880 })
   8881 #else
   8882 #define vld2_f32(__p0) __extension__ ({ \
   8883   float32x2x2_t __ret; \
   8884   __builtin_neon_vld2_v(&__ret, __p0, 9); \
   8885  \
   8886   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   8887   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   8888   __ret; \
   8889 })
   8890 #endif
   8891 
   8892 #ifdef __LITTLE_ENDIAN__
   8893 #define vld2_f16(__p0) __extension__ ({ \
   8894   float16x4x2_t __ret; \
   8895   __builtin_neon_vld2_v(&__ret, __p0, 8); \
   8896   __ret; \
   8897 })
   8898 #else
   8899 #define vld2_f16(__p0) __extension__ ({ \
   8900   float16x4x2_t __ret; \
   8901   __builtin_neon_vld2_v(&__ret, __p0, 8); \
   8902  \
   8903   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   8904   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   8905   __ret; \
   8906 })
   8907 #endif
   8908 
   8909 #ifdef __LITTLE_ENDIAN__
   8910 #define vld2_s32(__p0) __extension__ ({ \
   8911   int32x2x2_t __ret; \
   8912   __builtin_neon_vld2_v(&__ret, __p0, 2); \
   8913   __ret; \
   8914 })
   8915 #else
   8916 #define vld2_s32(__p0) __extension__ ({ \
   8917   int32x2x2_t __ret; \
   8918   __builtin_neon_vld2_v(&__ret, __p0, 2); \
   8919  \
   8920   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   8921   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   8922   __ret; \
   8923 })
   8924 #endif
   8925 
   8926 #ifdef __LITTLE_ENDIAN__
   8927 #define vld2_s64(__p0) __extension__ ({ \
   8928   int64x1x2_t __ret; \
   8929   __builtin_neon_vld2_v(&__ret, __p0, 3); \
   8930   __ret; \
   8931 })
   8932 #else
   8933 #define vld2_s64(__p0) __extension__ ({ \
   8934   int64x1x2_t __ret; \
   8935   __builtin_neon_vld2_v(&__ret, __p0, 3); \
   8936   __ret; \
   8937 })
   8938 #endif
   8939 
   8940 #ifdef __LITTLE_ENDIAN__
   8941 #define vld2_s16(__p0) __extension__ ({ \
   8942   int16x4x2_t __ret; \
   8943   __builtin_neon_vld2_v(&__ret, __p0, 1); \
   8944   __ret; \
   8945 })
   8946 #else
   8947 #define vld2_s16(__p0) __extension__ ({ \
   8948   int16x4x2_t __ret; \
   8949   __builtin_neon_vld2_v(&__ret, __p0, 1); \
   8950  \
   8951   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   8952   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   8953   __ret; \
   8954 })
   8955 #endif
   8956 
   8957 #ifdef __LITTLE_ENDIAN__
   8958 #define vld2_dup_p8(__p0) __extension__ ({ \
   8959   poly8x8x2_t __ret; \
   8960   __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
   8961   __ret; \
   8962 })
   8963 #else
   8964 #define vld2_dup_p8(__p0) __extension__ ({ \
   8965   poly8x8x2_t __ret; \
   8966   __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
   8967  \
   8968   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   8969   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   8970   __ret; \
   8971 })
   8972 #endif
   8973 
   8974 #ifdef __LITTLE_ENDIAN__
   8975 #define vld2_dup_p16(__p0) __extension__ ({ \
   8976   poly16x4x2_t __ret; \
   8977   __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
   8978   __ret; \
   8979 })
   8980 #else
   8981 #define vld2_dup_p16(__p0) __extension__ ({ \
   8982   poly16x4x2_t __ret; \
   8983   __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
   8984  \
   8985   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   8986   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   8987   __ret; \
   8988 })
   8989 #endif
   8990 
   8991 #ifdef __LITTLE_ENDIAN__
   8992 #define vld2_dup_u8(__p0) __extension__ ({ \
   8993   uint8x8x2_t __ret; \
   8994   __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
   8995   __ret; \
   8996 })
   8997 #else
   8998 #define vld2_dup_u8(__p0) __extension__ ({ \
   8999   uint8x8x2_t __ret; \
   9000   __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
   9001  \
   9002   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9003   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9004   __ret; \
   9005 })
   9006 #endif
   9007 
   9008 #ifdef __LITTLE_ENDIAN__
   9009 #define vld2_dup_u32(__p0) __extension__ ({ \
   9010   uint32x2x2_t __ret; \
   9011   __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
   9012   __ret; \
   9013 })
   9014 #else
   9015 #define vld2_dup_u32(__p0) __extension__ ({ \
   9016   uint32x2x2_t __ret; \
   9017   __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
   9018  \
   9019   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9020   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9021   __ret; \
   9022 })
   9023 #endif
   9024 
   9025 #ifdef __LITTLE_ENDIAN__
   9026 #define vld2_dup_u64(__p0) __extension__ ({ \
   9027   uint64x1x2_t __ret; \
   9028   __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
   9029   __ret; \
   9030 })
   9031 #else
   9032 #define vld2_dup_u64(__p0) __extension__ ({ \
   9033   uint64x1x2_t __ret; \
   9034   __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
   9035   __ret; \
   9036 })
   9037 #endif
   9038 
   9039 #ifdef __LITTLE_ENDIAN__
   9040 #define vld2_dup_u16(__p0) __extension__ ({ \
   9041   uint16x4x2_t __ret; \
   9042   __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
   9043   __ret; \
   9044 })
   9045 #else
   9046 #define vld2_dup_u16(__p0) __extension__ ({ \
   9047   uint16x4x2_t __ret; \
   9048   __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
   9049  \
   9050   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9051   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9052   __ret; \
   9053 })
   9054 #endif
   9055 
   9056 #ifdef __LITTLE_ENDIAN__
   9057 #define vld2_dup_s8(__p0) __extension__ ({ \
   9058   int8x8x2_t __ret; \
   9059   __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
   9060   __ret; \
   9061 })
   9062 #else
   9063 #define vld2_dup_s8(__p0) __extension__ ({ \
   9064   int8x8x2_t __ret; \
   9065   __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
   9066  \
   9067   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9068   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9069   __ret; \
   9070 })
   9071 #endif
   9072 
   9073 #ifdef __LITTLE_ENDIAN__
   9074 #define vld2_dup_f32(__p0) __extension__ ({ \
   9075   float32x2x2_t __ret; \
   9076   __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
   9077   __ret; \
   9078 })
   9079 #else
   9080 #define vld2_dup_f32(__p0) __extension__ ({ \
   9081   float32x2x2_t __ret; \
   9082   __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
   9083  \
   9084   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9085   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9086   __ret; \
   9087 })
   9088 #endif
   9089 
   9090 #ifdef __LITTLE_ENDIAN__
   9091 #define vld2_dup_f16(__p0) __extension__ ({ \
   9092   float16x4x2_t __ret; \
   9093   __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
   9094   __ret; \
   9095 })
   9096 #else
   9097 #define vld2_dup_f16(__p0) __extension__ ({ \
   9098   float16x4x2_t __ret; \
   9099   __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
   9100  \
   9101   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9102   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9103   __ret; \
   9104 })
   9105 #endif
   9106 
   9107 #ifdef __LITTLE_ENDIAN__
   9108 #define vld2_dup_s32(__p0) __extension__ ({ \
   9109   int32x2x2_t __ret; \
   9110   __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
   9111   __ret; \
   9112 })
   9113 #else
   9114 #define vld2_dup_s32(__p0) __extension__ ({ \
   9115   int32x2x2_t __ret; \
   9116   __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
   9117  \
   9118   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9119   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9120   __ret; \
   9121 })
   9122 #endif
   9123 
   9124 #ifdef __LITTLE_ENDIAN__
   9125 #define vld2_dup_s64(__p0) __extension__ ({ \
   9126   int64x1x2_t __ret; \
   9127   __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
   9128   __ret; \
   9129 })
   9130 #else
   9131 #define vld2_dup_s64(__p0) __extension__ ({ \
   9132   int64x1x2_t __ret; \
   9133   __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
   9134   __ret; \
   9135 })
   9136 #endif
   9137 
   9138 #ifdef __LITTLE_ENDIAN__
   9139 #define vld2_dup_s16(__p0) __extension__ ({ \
   9140   int16x4x2_t __ret; \
   9141   __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
   9142   __ret; \
   9143 })
   9144 #else
   9145 #define vld2_dup_s16(__p0) __extension__ ({ \
   9146   int16x4x2_t __ret; \
   9147   __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
   9148  \
   9149   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9150   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9151   __ret; \
   9152 })
   9153 #endif
   9154 
   9155 #ifdef __LITTLE_ENDIAN__
   9156 #define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   9157   poly8x8x2_t __s1 = __p1; \
   9158   poly8x8x2_t __ret; \
   9159   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
   9160   __ret; \
   9161 })
   9162 #else
   9163 #define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   9164   poly8x8x2_t __s1 = __p1; \
   9165   poly8x8x2_t __rev1; \
   9166   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9167   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9168   poly8x8x2_t __ret; \
   9169   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
   9170  \
   9171   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9172   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9173   __ret; \
   9174 })
   9175 #endif
   9176 
   9177 #ifdef __LITTLE_ENDIAN__
   9178 #define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   9179   poly16x4x2_t __s1 = __p1; \
   9180   poly16x4x2_t __ret; \
   9181   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
   9182   __ret; \
   9183 })
   9184 #else
   9185 #define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   9186   poly16x4x2_t __s1 = __p1; \
   9187   poly16x4x2_t __rev1; \
   9188   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   9189   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   9190   poly16x4x2_t __ret; \
   9191   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
   9192  \
   9193   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9194   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9195   __ret; \
   9196 })
   9197 #endif
   9198 
   9199 #ifdef __LITTLE_ENDIAN__
   9200 #define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   9201   poly16x8x2_t __s1 = __p1; \
   9202   poly16x8x2_t __ret; \
   9203   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
   9204   __ret; \
   9205 })
   9206 #else
   9207 #define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   9208   poly16x8x2_t __s1 = __p1; \
   9209   poly16x8x2_t __rev1; \
   9210   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9211   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9212   poly16x8x2_t __ret; \
   9213   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
   9214  \
   9215   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9216   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9217   __ret; \
   9218 })
   9219 #endif
   9220 
   9221 #ifdef __LITTLE_ENDIAN__
   9222 #define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   9223   uint32x4x2_t __s1 = __p1; \
   9224   uint32x4x2_t __ret; \
   9225   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
   9226   __ret; \
   9227 })
   9228 #else
   9229 #define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   9230   uint32x4x2_t __s1 = __p1; \
   9231   uint32x4x2_t __rev1; \
   9232   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   9233   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   9234   uint32x4x2_t __ret; \
   9235   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
   9236  \
   9237   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9238   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9239   __ret; \
   9240 })
   9241 #endif
   9242 
   9243 #ifdef __LITTLE_ENDIAN__
   9244 #define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   9245   uint16x8x2_t __s1 = __p1; \
   9246   uint16x8x2_t __ret; \
   9247   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
   9248   __ret; \
   9249 })
   9250 #else
   9251 #define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   9252   uint16x8x2_t __s1 = __p1; \
   9253   uint16x8x2_t __rev1; \
   9254   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9255   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9256   uint16x8x2_t __ret; \
   9257   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
   9258  \
   9259   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9260   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9261   __ret; \
   9262 })
   9263 #endif
   9264 
   9265 #ifdef __LITTLE_ENDIAN__
   9266 #define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   9267   float32x4x2_t __s1 = __p1; \
   9268   float32x4x2_t __ret; \
   9269   __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 41); \
   9270   __ret; \
   9271 })
   9272 #else
   9273 #define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   9274   float32x4x2_t __s1 = __p1; \
   9275   float32x4x2_t __rev1; \
   9276   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   9277   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   9278   float32x4x2_t __ret; \
   9279   __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
   9280  \
   9281   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9282   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9283   __ret; \
   9284 })
   9285 #endif
   9286 
   9287 #ifdef __LITTLE_ENDIAN__
   9288 #define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   9289   float16x8x2_t __s1 = __p1; \
   9290   float16x8x2_t __ret; \
   9291   __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 40); \
   9292   __ret; \
   9293 })
   9294 #else
   9295 #define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   9296   float16x8x2_t __s1 = __p1; \
   9297   float16x8x2_t __rev1; \
   9298   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9299   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9300   float16x8x2_t __ret; \
   9301   __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
   9302  \
   9303   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9304   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9305   __ret; \
   9306 })
   9307 #endif
   9308 
   9309 #ifdef __LITTLE_ENDIAN__
   9310 #define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   9311   int32x4x2_t __s1 = __p1; \
   9312   int32x4x2_t __ret; \
   9313   __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 34); \
   9314   __ret; \
   9315 })
   9316 #else
   9317 #define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   9318   int32x4x2_t __s1 = __p1; \
   9319   int32x4x2_t __rev1; \
   9320   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   9321   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   9322   int32x4x2_t __ret; \
   9323   __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
   9324  \
   9325   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9326   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9327   __ret; \
   9328 })
   9329 #endif
   9330 
   9331 #ifdef __LITTLE_ENDIAN__
   9332 #define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   9333   int16x8x2_t __s1 = __p1; \
   9334   int16x8x2_t __ret; \
   9335   __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 33); \
   9336   __ret; \
   9337 })
   9338 #else
   9339 #define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   9340   int16x8x2_t __s1 = __p1; \
   9341   int16x8x2_t __rev1; \
   9342   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9343   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9344   int16x8x2_t __ret; \
   9345   __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
   9346  \
   9347   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9348   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9349   __ret; \
   9350 })
   9351 #endif
   9352 
   9353 #ifdef __LITTLE_ENDIAN__
   9354 #define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   9355   uint8x8x2_t __s1 = __p1; \
   9356   uint8x8x2_t __ret; \
   9357   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
   9358   __ret; \
   9359 })
   9360 #else
   9361 #define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   9362   uint8x8x2_t __s1 = __p1; \
   9363   uint8x8x2_t __rev1; \
   9364   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9365   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9366   uint8x8x2_t __ret; \
   9367   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
   9368  \
   9369   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9370   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9371   __ret; \
   9372 })
   9373 #endif
   9374 
   9375 #ifdef __LITTLE_ENDIAN__
   9376 #define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   9377   uint32x2x2_t __s1 = __p1; \
   9378   uint32x2x2_t __ret; \
   9379   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
   9380   __ret; \
   9381 })
   9382 #else
   9383 #define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   9384   uint32x2x2_t __s1 = __p1; \
   9385   uint32x2x2_t __rev1; \
   9386   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   9387   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   9388   uint32x2x2_t __ret; \
   9389   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
   9390  \
   9391   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9392   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9393   __ret; \
   9394 })
   9395 #endif
   9396 
   9397 #ifdef __LITTLE_ENDIAN__
   9398 #define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   9399   uint16x4x2_t __s1 = __p1; \
   9400   uint16x4x2_t __ret; \
   9401   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
   9402   __ret; \
   9403 })
   9404 #else
   9405 #define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   9406   uint16x4x2_t __s1 = __p1; \
   9407   uint16x4x2_t __rev1; \
   9408   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   9409   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   9410   uint16x4x2_t __ret; \
   9411   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
   9412  \
   9413   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9414   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9415   __ret; \
   9416 })
   9417 #endif
   9418 
   9419 #ifdef __LITTLE_ENDIAN__
   9420 #define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   9421   int8x8x2_t __s1 = __p1; \
   9422   int8x8x2_t __ret; \
   9423   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
   9424   __ret; \
   9425 })
   9426 #else
   9427 #define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   9428   int8x8x2_t __s1 = __p1; \
   9429   int8x8x2_t __rev1; \
   9430   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9431   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9432   int8x8x2_t __ret; \
   9433   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
   9434  \
   9435   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9436   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9437   __ret; \
   9438 })
   9439 #endif
   9440 
   9441 #ifdef __LITTLE_ENDIAN__
   9442 #define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   9443   float32x2x2_t __s1 = __p1; \
   9444   float32x2x2_t __ret; \
   9445   __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 9); \
   9446   __ret; \
   9447 })
   9448 #else
   9449 #define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   9450   float32x2x2_t __s1 = __p1; \
   9451   float32x2x2_t __rev1; \
   9452   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   9453   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   9454   float32x2x2_t __ret; \
   9455   __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
   9456  \
   9457   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9458   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9459   __ret; \
   9460 })
   9461 #endif
   9462 
   9463 #ifdef __LITTLE_ENDIAN__
   9464 #define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   9465   float16x4x2_t __s1 = __p1; \
   9466   float16x4x2_t __ret; \
   9467   __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 8); \
   9468   __ret; \
   9469 })
   9470 #else
   9471 #define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   9472   float16x4x2_t __s1 = __p1; \
   9473   float16x4x2_t __rev1; \
   9474   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   9475   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   9476   float16x4x2_t __ret; \
   9477   __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
   9478  \
   9479   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9480   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9481   __ret; \
   9482 })
   9483 #endif
   9484 
   9485 #ifdef __LITTLE_ENDIAN__
   9486 #define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   9487   int32x2x2_t __s1 = __p1; \
   9488   int32x2x2_t __ret; \
   9489   __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 2); \
   9490   __ret; \
   9491 })
   9492 #else
   9493 #define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   9494   int32x2x2_t __s1 = __p1; \
   9495   int32x2x2_t __rev1; \
   9496   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   9497   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   9498   int32x2x2_t __ret; \
   9499   __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
   9500  \
   9501   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9502   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9503   __ret; \
   9504 })
   9505 #endif
   9506 
   9507 #ifdef __LITTLE_ENDIAN__
   9508 #define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   9509   int16x4x2_t __s1 = __p1; \
   9510   int16x4x2_t __ret; \
   9511   __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 1); \
   9512   __ret; \
   9513 })
   9514 #else
   9515 #define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   9516   int16x4x2_t __s1 = __p1; \
   9517   int16x4x2_t __rev1; \
   9518   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   9519   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   9520   int16x4x2_t __ret; \
   9521   __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
   9522  \
   9523   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9524   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9525   __ret; \
   9526 })
   9527 #endif
   9528 
   9529 #ifdef __LITTLE_ENDIAN__
   9530 #define vld3_p8(__p0) __extension__ ({ \
   9531   poly8x8x3_t __ret; \
   9532   __builtin_neon_vld3_v(&__ret, __p0, 4); \
   9533   __ret; \
   9534 })
   9535 #else
   9536 #define vld3_p8(__p0) __extension__ ({ \
   9537   poly8x8x3_t __ret; \
   9538   __builtin_neon_vld3_v(&__ret, __p0, 4); \
   9539  \
   9540   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9541   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9542   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   9543   __ret; \
   9544 })
   9545 #endif
   9546 
   9547 #ifdef __LITTLE_ENDIAN__
   9548 #define vld3_p16(__p0) __extension__ ({ \
   9549   poly16x4x3_t __ret; \
   9550   __builtin_neon_vld3_v(&__ret, __p0, 5); \
   9551   __ret; \
   9552 })
   9553 #else
   9554 #define vld3_p16(__p0) __extension__ ({ \
   9555   poly16x4x3_t __ret; \
   9556   __builtin_neon_vld3_v(&__ret, __p0, 5); \
   9557  \
   9558   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9559   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9560   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   9561   __ret; \
   9562 })
   9563 #endif
   9564 
   9565 #ifdef __LITTLE_ENDIAN__
   9566 #define vld3q_p8(__p0) __extension__ ({ \
   9567   poly8x16x3_t __ret; \
   9568   __builtin_neon_vld3q_v(&__ret, __p0, 36); \
   9569   __ret; \
   9570 })
   9571 #else
   9572 #define vld3q_p8(__p0) __extension__ ({ \
   9573   poly8x16x3_t __ret; \
   9574   __builtin_neon_vld3q_v(&__ret, __p0, 36); \
   9575  \
   9576   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   9577   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   9578   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   9579   __ret; \
   9580 })
   9581 #endif
   9582 
   9583 #ifdef __LITTLE_ENDIAN__
   9584 #define vld3q_p16(__p0) __extension__ ({ \
   9585   poly16x8x3_t __ret; \
   9586   __builtin_neon_vld3q_v(&__ret, __p0, 37); \
   9587   __ret; \
   9588 })
   9589 #else
   9590 #define vld3q_p16(__p0) __extension__ ({ \
   9591   poly16x8x3_t __ret; \
   9592   __builtin_neon_vld3q_v(&__ret, __p0, 37); \
   9593  \
   9594   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9595   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9596   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   9597   __ret; \
   9598 })
   9599 #endif
   9600 
   9601 #ifdef __LITTLE_ENDIAN__
   9602 #define vld3q_u8(__p0) __extension__ ({ \
   9603   uint8x16x3_t __ret; \
   9604   __builtin_neon_vld3q_v(&__ret, __p0, 48); \
   9605   __ret; \
   9606 })
   9607 #else
   9608 #define vld3q_u8(__p0) __extension__ ({ \
   9609   uint8x16x3_t __ret; \
   9610   __builtin_neon_vld3q_v(&__ret, __p0, 48); \
   9611  \
   9612   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   9613   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   9614   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   9615   __ret; \
   9616 })
   9617 #endif
   9618 
   9619 #ifdef __LITTLE_ENDIAN__
   9620 #define vld3q_u32(__p0) __extension__ ({ \
   9621   uint32x4x3_t __ret; \
   9622   __builtin_neon_vld3q_v(&__ret, __p0, 50); \
   9623   __ret; \
   9624 })
   9625 #else
   9626 #define vld3q_u32(__p0) __extension__ ({ \
   9627   uint32x4x3_t __ret; \
   9628   __builtin_neon_vld3q_v(&__ret, __p0, 50); \
   9629  \
   9630   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9631   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9632   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   9633   __ret; \
   9634 })
   9635 #endif
   9636 
   9637 #ifdef __LITTLE_ENDIAN__
   9638 #define vld3q_u16(__p0) __extension__ ({ \
   9639   uint16x8x3_t __ret; \
   9640   __builtin_neon_vld3q_v(&__ret, __p0, 49); \
   9641   __ret; \
   9642 })
   9643 #else
   9644 #define vld3q_u16(__p0) __extension__ ({ \
   9645   uint16x8x3_t __ret; \
   9646   __builtin_neon_vld3q_v(&__ret, __p0, 49); \
   9647  \
   9648   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9649   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9650   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   9651   __ret; \
   9652 })
   9653 #endif
   9654 
   9655 #ifdef __LITTLE_ENDIAN__
   9656 #define vld3q_s8(__p0) __extension__ ({ \
   9657   int8x16x3_t __ret; \
   9658   __builtin_neon_vld3q_v(&__ret, __p0, 32); \
   9659   __ret; \
   9660 })
   9661 #else
   9662 #define vld3q_s8(__p0) __extension__ ({ \
   9663   int8x16x3_t __ret; \
   9664   __builtin_neon_vld3q_v(&__ret, __p0, 32); \
   9665  \
   9666   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   9667   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   9668   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   9669   __ret; \
   9670 })
   9671 #endif
   9672 
   9673 #ifdef __LITTLE_ENDIAN__
   9674 #define vld3q_f32(__p0) __extension__ ({ \
   9675   float32x4x3_t __ret; \
   9676   __builtin_neon_vld3q_v(&__ret, __p0, 41); \
   9677   __ret; \
   9678 })
   9679 #else
   9680 #define vld3q_f32(__p0) __extension__ ({ \
   9681   float32x4x3_t __ret; \
   9682   __builtin_neon_vld3q_v(&__ret, __p0, 41); \
   9683  \
   9684   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9685   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9686   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   9687   __ret; \
   9688 })
   9689 #endif
   9690 
   9691 #ifdef __LITTLE_ENDIAN__
   9692 #define vld3q_f16(__p0) __extension__ ({ \
   9693   float16x8x3_t __ret; \
   9694   __builtin_neon_vld3q_v(&__ret, __p0, 40); \
   9695   __ret; \
   9696 })
   9697 #else
   9698 #define vld3q_f16(__p0) __extension__ ({ \
   9699   float16x8x3_t __ret; \
   9700   __builtin_neon_vld3q_v(&__ret, __p0, 40); \
   9701  \
   9702   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9703   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9704   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   9705   __ret; \
   9706 })
   9707 #endif
   9708 
   9709 #ifdef __LITTLE_ENDIAN__
   9710 #define vld3q_s32(__p0) __extension__ ({ \
   9711   int32x4x3_t __ret; \
   9712   __builtin_neon_vld3q_v(&__ret, __p0, 34); \
   9713   __ret; \
   9714 })
   9715 #else
   9716 #define vld3q_s32(__p0) __extension__ ({ \
   9717   int32x4x3_t __ret; \
   9718   __builtin_neon_vld3q_v(&__ret, __p0, 34); \
   9719  \
   9720   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9721   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9722   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   9723   __ret; \
   9724 })
   9725 #endif
   9726 
   9727 #ifdef __LITTLE_ENDIAN__
   9728 #define vld3q_s16(__p0) __extension__ ({ \
   9729   int16x8x3_t __ret; \
   9730   __builtin_neon_vld3q_v(&__ret, __p0, 33); \
   9731   __ret; \
   9732 })
   9733 #else
   9734 #define vld3q_s16(__p0) __extension__ ({ \
   9735   int16x8x3_t __ret; \
   9736   __builtin_neon_vld3q_v(&__ret, __p0, 33); \
   9737  \
   9738   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9739   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9740   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   9741   __ret; \
   9742 })
   9743 #endif
   9744 
   9745 #ifdef __LITTLE_ENDIAN__
   9746 #define vld3_u8(__p0) __extension__ ({ \
   9747   uint8x8x3_t __ret; \
   9748   __builtin_neon_vld3_v(&__ret, __p0, 16); \
   9749   __ret; \
   9750 })
   9751 #else
   9752 #define vld3_u8(__p0) __extension__ ({ \
   9753   uint8x8x3_t __ret; \
   9754   __builtin_neon_vld3_v(&__ret, __p0, 16); \
   9755  \
   9756   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9757   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9758   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   9759   __ret; \
   9760 })
   9761 #endif
   9762 
   9763 #ifdef __LITTLE_ENDIAN__
   9764 #define vld3_u32(__p0) __extension__ ({ \
   9765   uint32x2x3_t __ret; \
   9766   __builtin_neon_vld3_v(&__ret, __p0, 18); \
   9767   __ret; \
   9768 })
   9769 #else
   9770 #define vld3_u32(__p0) __extension__ ({ \
   9771   uint32x2x3_t __ret; \
   9772   __builtin_neon_vld3_v(&__ret, __p0, 18); \
   9773  \
   9774   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9775   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9776   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   9777   __ret; \
   9778 })
   9779 #endif
   9780 
   9781 #ifdef __LITTLE_ENDIAN__
   9782 #define vld3_u64(__p0) __extension__ ({ \
   9783   uint64x1x3_t __ret; \
   9784   __builtin_neon_vld3_v(&__ret, __p0, 19); \
   9785   __ret; \
   9786 })
   9787 #else
   9788 #define vld3_u64(__p0) __extension__ ({ \
   9789   uint64x1x3_t __ret; \
   9790   __builtin_neon_vld3_v(&__ret, __p0, 19); \
   9791   __ret; \
   9792 })
   9793 #endif
   9794 
   9795 #ifdef __LITTLE_ENDIAN__
   9796 #define vld3_u16(__p0) __extension__ ({ \
   9797   uint16x4x3_t __ret; \
   9798   __builtin_neon_vld3_v(&__ret, __p0, 17); \
   9799   __ret; \
   9800 })
   9801 #else
   9802 #define vld3_u16(__p0) __extension__ ({ \
   9803   uint16x4x3_t __ret; \
   9804   __builtin_neon_vld3_v(&__ret, __p0, 17); \
   9805  \
   9806   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9807   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9808   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   9809   __ret; \
   9810 })
   9811 #endif
   9812 
   9813 #ifdef __LITTLE_ENDIAN__
   9814 #define vld3_s8(__p0) __extension__ ({ \
   9815   int8x8x3_t __ret; \
   9816   __builtin_neon_vld3_v(&__ret, __p0, 0); \
   9817   __ret; \
   9818 })
   9819 #else
   9820 #define vld3_s8(__p0) __extension__ ({ \
   9821   int8x8x3_t __ret; \
   9822   __builtin_neon_vld3_v(&__ret, __p0, 0); \
   9823  \
   9824   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9825   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9826   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   9827   __ret; \
   9828 })
   9829 #endif
   9830 
   9831 #ifdef __LITTLE_ENDIAN__
   9832 #define vld3_f32(__p0) __extension__ ({ \
   9833   float32x2x3_t __ret; \
   9834   __builtin_neon_vld3_v(&__ret, __p0, 9); \
   9835   __ret; \
   9836 })
   9837 #else
   9838 #define vld3_f32(__p0) __extension__ ({ \
   9839   float32x2x3_t __ret; \
   9840   __builtin_neon_vld3_v(&__ret, __p0, 9); \
   9841  \
   9842   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9843   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9844   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   9845   __ret; \
   9846 })
   9847 #endif
   9848 
   9849 #ifdef __LITTLE_ENDIAN__
   9850 #define vld3_f16(__p0) __extension__ ({ \
   9851   float16x4x3_t __ret; \
   9852   __builtin_neon_vld3_v(&__ret, __p0, 8); \
   9853   __ret; \
   9854 })
   9855 #else
   9856 #define vld3_f16(__p0) __extension__ ({ \
   9857   float16x4x3_t __ret; \
   9858   __builtin_neon_vld3_v(&__ret, __p0, 8); \
   9859  \
   9860   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9861   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9862   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   9863   __ret; \
   9864 })
   9865 #endif
   9866 
   9867 #ifdef __LITTLE_ENDIAN__
   9868 #define vld3_s32(__p0) __extension__ ({ \
   9869   int32x2x3_t __ret; \
   9870   __builtin_neon_vld3_v(&__ret, __p0, 2); \
   9871   __ret; \
   9872 })
   9873 #else
   9874 #define vld3_s32(__p0) __extension__ ({ \
   9875   int32x2x3_t __ret; \
   9876   __builtin_neon_vld3_v(&__ret, __p0, 2); \
   9877  \
   9878   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9879   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9880   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   9881   __ret; \
   9882 })
   9883 #endif
   9884 
   9885 #ifdef __LITTLE_ENDIAN__
   9886 #define vld3_s64(__p0) __extension__ ({ \
   9887   int64x1x3_t __ret; \
   9888   __builtin_neon_vld3_v(&__ret, __p0, 3); \
   9889   __ret; \
   9890 })
   9891 #else
   9892 #define vld3_s64(__p0) __extension__ ({ \
   9893   int64x1x3_t __ret; \
   9894   __builtin_neon_vld3_v(&__ret, __p0, 3); \
   9895   __ret; \
   9896 })
   9897 #endif
   9898 
   9899 #ifdef __LITTLE_ENDIAN__
   9900 #define vld3_s16(__p0) __extension__ ({ \
   9901   int16x4x3_t __ret; \
   9902   __builtin_neon_vld3_v(&__ret, __p0, 1); \
   9903   __ret; \
   9904 })
   9905 #else
   9906 #define vld3_s16(__p0) __extension__ ({ \
   9907   int16x4x3_t __ret; \
   9908   __builtin_neon_vld3_v(&__ret, __p0, 1); \
   9909  \
   9910   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9911   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9912   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   9913   __ret; \
   9914 })
   9915 #endif
   9916 
   9917 #ifdef __LITTLE_ENDIAN__
   9918 #define vld3_dup_p8(__p0) __extension__ ({ \
   9919   poly8x8x3_t __ret; \
   9920   __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
   9921   __ret; \
   9922 })
   9923 #else
   9924 #define vld3_dup_p8(__p0) __extension__ ({ \
   9925   poly8x8x3_t __ret; \
   9926   __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
   9927  \
   9928   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9929   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9930   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   9931   __ret; \
   9932 })
   9933 #endif
   9934 
   9935 #ifdef __LITTLE_ENDIAN__
   9936 #define vld3_dup_p16(__p0) __extension__ ({ \
   9937   poly16x4x3_t __ret; \
   9938   __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
   9939   __ret; \
   9940 })
   9941 #else
   9942 #define vld3_dup_p16(__p0) __extension__ ({ \
   9943   poly16x4x3_t __ret; \
   9944   __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
   9945  \
   9946   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   9947   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   9948   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   9949   __ret; \
   9950 })
   9951 #endif
   9952 
   9953 #ifdef __LITTLE_ENDIAN__
   9954 #define vld3_dup_u8(__p0) __extension__ ({ \
   9955   uint8x8x3_t __ret; \
   9956   __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
   9957   __ret; \
   9958 })
   9959 #else
   9960 #define vld3_dup_u8(__p0) __extension__ ({ \
   9961   uint8x8x3_t __ret; \
   9962   __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
   9963  \
   9964   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   9965   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   9966   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   9967   __ret; \
   9968 })
   9969 #endif
   9970 
   9971 #ifdef __LITTLE_ENDIAN__
   9972 #define vld3_dup_u32(__p0) __extension__ ({ \
   9973   uint32x2x3_t __ret; \
   9974   __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
   9975   __ret; \
   9976 })
   9977 #else
   9978 #define vld3_dup_u32(__p0) __extension__ ({ \
   9979   uint32x2x3_t __ret; \
   9980   __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
   9981  \
   9982   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   9983   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   9984   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   9985   __ret; \
   9986 })
   9987 #endif
   9988 
   9989 #ifdef __LITTLE_ENDIAN__
   9990 #define vld3_dup_u64(__p0) __extension__ ({ \
   9991   uint64x1x3_t __ret; \
   9992   __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
   9993   __ret; \
   9994 })
   9995 #else
   9996 #define vld3_dup_u64(__p0) __extension__ ({ \
   9997   uint64x1x3_t __ret; \
   9998   __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
   9999   __ret; \
   10000 })
   10001 #endif
   10002 
   10003 #ifdef __LITTLE_ENDIAN__
   10004 #define vld3_dup_u16(__p0) __extension__ ({ \
   10005   uint16x4x3_t __ret; \
   10006   __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
   10007   __ret; \
   10008 })
   10009 #else
   10010 #define vld3_dup_u16(__p0) __extension__ ({ \
   10011   uint16x4x3_t __ret; \
   10012   __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
   10013  \
   10014   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10015   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10016   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10017   __ret; \
   10018 })
   10019 #endif
   10020 
   10021 #ifdef __LITTLE_ENDIAN__
   10022 #define vld3_dup_s8(__p0) __extension__ ({ \
   10023   int8x8x3_t __ret; \
   10024   __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
   10025   __ret; \
   10026 })
   10027 #else
   10028 #define vld3_dup_s8(__p0) __extension__ ({ \
   10029   int8x8x3_t __ret; \
   10030   __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
   10031  \
   10032   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10033   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10034   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10035   __ret; \
   10036 })
   10037 #endif
   10038 
   10039 #ifdef __LITTLE_ENDIAN__
   10040 #define vld3_dup_f32(__p0) __extension__ ({ \
   10041   float32x2x3_t __ret; \
   10042   __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
   10043   __ret; \
   10044 })
   10045 #else
   10046 #define vld3_dup_f32(__p0) __extension__ ({ \
   10047   float32x2x3_t __ret; \
   10048   __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
   10049  \
   10050   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   10051   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   10052   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   10053   __ret; \
   10054 })
   10055 #endif
   10056 
   10057 #ifdef __LITTLE_ENDIAN__
   10058 #define vld3_dup_f16(__p0) __extension__ ({ \
   10059   float16x4x3_t __ret; \
   10060   __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
   10061   __ret; \
   10062 })
   10063 #else
   10064 #define vld3_dup_f16(__p0) __extension__ ({ \
   10065   float16x4x3_t __ret; \
   10066   __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
   10067  \
   10068   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10069   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10070   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10071   __ret; \
   10072 })
   10073 #endif
   10074 
   10075 #ifdef __LITTLE_ENDIAN__
   10076 #define vld3_dup_s32(__p0) __extension__ ({ \
   10077   int32x2x3_t __ret; \
   10078   __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
   10079   __ret; \
   10080 })
   10081 #else
   10082 #define vld3_dup_s32(__p0) __extension__ ({ \
   10083   int32x2x3_t __ret; \
   10084   __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
   10085  \
   10086   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   10087   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   10088   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   10089   __ret; \
   10090 })
   10091 #endif
   10092 
   10093 #ifdef __LITTLE_ENDIAN__
   10094 #define vld3_dup_s64(__p0) __extension__ ({ \
   10095   int64x1x3_t __ret; \
   10096   __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
   10097   __ret; \
   10098 })
   10099 #else
   10100 #define vld3_dup_s64(__p0) __extension__ ({ \
   10101   int64x1x3_t __ret; \
   10102   __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
   10103   __ret; \
   10104 })
   10105 #endif
   10106 
   10107 #ifdef __LITTLE_ENDIAN__
   10108 #define vld3_dup_s16(__p0) __extension__ ({ \
   10109   int16x4x3_t __ret; \
   10110   __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
   10111   __ret; \
   10112 })
   10113 #else
   10114 #define vld3_dup_s16(__p0) __extension__ ({ \
   10115   int16x4x3_t __ret; \
   10116   __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
   10117  \
   10118   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10119   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10120   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10121   __ret; \
   10122 })
   10123 #endif
   10124 
   10125 #ifdef __LITTLE_ENDIAN__
   10126 #define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   10127   poly8x8x3_t __s1 = __p1; \
   10128   poly8x8x3_t __ret; \
   10129   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
   10130   __ret; \
   10131 })
   10132 #else
   10133 #define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   10134   poly8x8x3_t __s1 = __p1; \
   10135   poly8x8x3_t __rev1; \
   10136   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10137   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10138   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10139   poly8x8x3_t __ret; \
   10140   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
   10141  \
   10142   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10143   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10144   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10145   __ret; \
   10146 })
   10147 #endif
   10148 
   10149 #ifdef __LITTLE_ENDIAN__
   10150 #define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   10151   poly16x4x3_t __s1 = __p1; \
   10152   poly16x4x3_t __ret; \
   10153   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
   10154   __ret; \
   10155 })
   10156 #else
   10157 #define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   10158   poly16x4x3_t __s1 = __p1; \
   10159   poly16x4x3_t __rev1; \
   10160   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   10161   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   10162   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   10163   poly16x4x3_t __ret; \
   10164   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
   10165  \
   10166   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10167   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10168   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10169   __ret; \
   10170 })
   10171 #endif
   10172 
   10173 #ifdef __LITTLE_ENDIAN__
   10174 #define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   10175   poly16x8x3_t __s1 = __p1; \
   10176   poly16x8x3_t __ret; \
   10177   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
   10178   __ret; \
   10179 })
   10180 #else
   10181 #define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   10182   poly16x8x3_t __s1 = __p1; \
   10183   poly16x8x3_t __rev1; \
   10184   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10185   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10186   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10187   poly16x8x3_t __ret; \
   10188   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
   10189  \
   10190   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10191   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10192   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10193   __ret; \
   10194 })
   10195 #endif
   10196 
   10197 #ifdef __LITTLE_ENDIAN__
   10198 #define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   10199   uint32x4x3_t __s1 = __p1; \
   10200   uint32x4x3_t __ret; \
   10201   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
   10202   __ret; \
   10203 })
   10204 #else
   10205 #define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   10206   uint32x4x3_t __s1 = __p1; \
   10207   uint32x4x3_t __rev1; \
   10208   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   10209   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   10210   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   10211   uint32x4x3_t __ret; \
   10212   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
   10213  \
   10214   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10215   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10216   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10217   __ret; \
   10218 })
   10219 #endif
   10220 
   10221 #ifdef __LITTLE_ENDIAN__
   10222 #define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   10223   uint16x8x3_t __s1 = __p1; \
   10224   uint16x8x3_t __ret; \
   10225   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
   10226   __ret; \
   10227 })
   10228 #else
   10229 #define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   10230   uint16x8x3_t __s1 = __p1; \
   10231   uint16x8x3_t __rev1; \
   10232   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10233   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10234   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10235   uint16x8x3_t __ret; \
   10236   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
   10237  \
   10238   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10239   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10240   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10241   __ret; \
   10242 })
   10243 #endif
   10244 
   10245 #ifdef __LITTLE_ENDIAN__
   10246 #define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   10247   float32x4x3_t __s1 = __p1; \
   10248   float32x4x3_t __ret; \
   10249   __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
   10250   __ret; \
   10251 })
   10252 #else
   10253 #define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   10254   float32x4x3_t __s1 = __p1; \
   10255   float32x4x3_t __rev1; \
   10256   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   10257   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   10258   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   10259   float32x4x3_t __ret; \
   10260   __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
   10261  \
   10262   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10263   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10264   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10265   __ret; \
   10266 })
   10267 #endif
   10268 
   10269 #ifdef __LITTLE_ENDIAN__
   10270 #define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   10271   float16x8x3_t __s1 = __p1; \
   10272   float16x8x3_t __ret; \
   10273   __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
   10274   __ret; \
   10275 })
   10276 #else
   10277 #define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   10278   float16x8x3_t __s1 = __p1; \
   10279   float16x8x3_t __rev1; \
   10280   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10281   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10282   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10283   float16x8x3_t __ret; \
   10284   __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
   10285  \
   10286   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10287   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10288   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10289   __ret; \
   10290 })
   10291 #endif
   10292 
   10293 #ifdef __LITTLE_ENDIAN__
   10294 #define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   10295   int32x4x3_t __s1 = __p1; \
   10296   int32x4x3_t __ret; \
   10297   __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
   10298   __ret; \
   10299 })
   10300 #else
   10301 #define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   10302   int32x4x3_t __s1 = __p1; \
   10303   int32x4x3_t __rev1; \
   10304   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   10305   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   10306   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   10307   int32x4x3_t __ret; \
   10308   __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
   10309  \
   10310   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10311   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10312   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10313   __ret; \
   10314 })
   10315 #endif
   10316 
   10317 #ifdef __LITTLE_ENDIAN__
   10318 #define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   10319   int16x8x3_t __s1 = __p1; \
   10320   int16x8x3_t __ret; \
   10321   __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
   10322   __ret; \
   10323 })
   10324 #else
   10325 #define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   10326   int16x8x3_t __s1 = __p1; \
   10327   int16x8x3_t __rev1; \
   10328   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10329   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10330   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10331   int16x8x3_t __ret; \
   10332   __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
   10333  \
   10334   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10335   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10336   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10337   __ret; \
   10338 })
   10339 #endif
   10340 
   10341 #ifdef __LITTLE_ENDIAN__
   10342 #define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   10343   uint8x8x3_t __s1 = __p1; \
   10344   uint8x8x3_t __ret; \
   10345   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
   10346   __ret; \
   10347 })
   10348 #else
   10349 #define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   10350   uint8x8x3_t __s1 = __p1; \
   10351   uint8x8x3_t __rev1; \
   10352   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10353   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10354   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10355   uint8x8x3_t __ret; \
   10356   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
   10357  \
   10358   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10359   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10360   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10361   __ret; \
   10362 })
   10363 #endif
   10364 
   10365 #ifdef __LITTLE_ENDIAN__
   10366 #define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   10367   uint32x2x3_t __s1 = __p1; \
   10368   uint32x2x3_t __ret; \
   10369   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
   10370   __ret; \
   10371 })
   10372 #else
   10373 #define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   10374   uint32x2x3_t __s1 = __p1; \
   10375   uint32x2x3_t __rev1; \
   10376   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   10377   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   10378   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   10379   uint32x2x3_t __ret; \
   10380   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
   10381  \
   10382   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   10383   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   10384   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   10385   __ret; \
   10386 })
   10387 #endif
   10388 
   10389 #ifdef __LITTLE_ENDIAN__
   10390 #define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   10391   uint16x4x3_t __s1 = __p1; \
   10392   uint16x4x3_t __ret; \
   10393   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
   10394   __ret; \
   10395 })
   10396 #else
   10397 #define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   10398   uint16x4x3_t __s1 = __p1; \
   10399   uint16x4x3_t __rev1; \
   10400   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   10401   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   10402   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   10403   uint16x4x3_t __ret; \
   10404   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
   10405  \
   10406   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10407   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10408   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10409   __ret; \
   10410 })
   10411 #endif
   10412 
   10413 #ifdef __LITTLE_ENDIAN__
   10414 #define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   10415   int8x8x3_t __s1 = __p1; \
   10416   int8x8x3_t __ret; \
   10417   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
   10418   __ret; \
   10419 })
   10420 #else
   10421 #define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   10422   int8x8x3_t __s1 = __p1; \
   10423   int8x8x3_t __rev1; \
   10424   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10425   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10426   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10427   int8x8x3_t __ret; \
   10428   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
   10429  \
   10430   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10431   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10432   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10433   __ret; \
   10434 })
   10435 #endif
   10436 
   10437 #ifdef __LITTLE_ENDIAN__
   10438 #define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   10439   float32x2x3_t __s1 = __p1; \
   10440   float32x2x3_t __ret; \
   10441   __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
   10442   __ret; \
   10443 })
   10444 #else
   10445 #define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   10446   float32x2x3_t __s1 = __p1; \
   10447   float32x2x3_t __rev1; \
   10448   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   10449   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   10450   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   10451   float32x2x3_t __ret; \
   10452   __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
   10453  \
   10454   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   10455   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   10456   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   10457   __ret; \
   10458 })
   10459 #endif
   10460 
   10461 #ifdef __LITTLE_ENDIAN__
   10462 #define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   10463   float16x4x3_t __s1 = __p1; \
   10464   float16x4x3_t __ret; \
   10465   __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
   10466   __ret; \
   10467 })
   10468 #else
   10469 #define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   10470   float16x4x3_t __s1 = __p1; \
   10471   float16x4x3_t __rev1; \
   10472   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   10473   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   10474   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   10475   float16x4x3_t __ret; \
   10476   __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
   10477  \
   10478   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10479   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10480   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10481   __ret; \
   10482 })
   10483 #endif
   10484 
   10485 #ifdef __LITTLE_ENDIAN__
   10486 #define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   10487   int32x2x3_t __s1 = __p1; \
   10488   int32x2x3_t __ret; \
   10489   __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
   10490   __ret; \
   10491 })
   10492 #else
   10493 #define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   10494   int32x2x3_t __s1 = __p1; \
   10495   int32x2x3_t __rev1; \
   10496   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   10497   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   10498   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   10499   int32x2x3_t __ret; \
   10500   __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
   10501  \
   10502   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   10503   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   10504   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   10505   __ret; \
   10506 })
   10507 #endif
   10508 
   10509 #ifdef __LITTLE_ENDIAN__
   10510 #define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   10511   int16x4x3_t __s1 = __p1; \
   10512   int16x4x3_t __ret; \
   10513   __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
   10514   __ret; \
   10515 })
   10516 #else
   10517 #define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   10518   int16x4x3_t __s1 = __p1; \
   10519   int16x4x3_t __rev1; \
   10520   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   10521   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   10522   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   10523   int16x4x3_t __ret; \
   10524   __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
   10525  \
   10526   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10527   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10528   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10529   __ret; \
   10530 })
   10531 #endif
   10532 
   10533 #ifdef __LITTLE_ENDIAN__
   10534 #define vld4_p8(__p0) __extension__ ({ \
   10535   poly8x8x4_t __ret; \
   10536   __builtin_neon_vld4_v(&__ret, __p0, 4); \
   10537   __ret; \
   10538 })
   10539 #else
   10540 #define vld4_p8(__p0) __extension__ ({ \
   10541   poly8x8x4_t __ret; \
   10542   __builtin_neon_vld4_v(&__ret, __p0, 4); \
   10543  \
   10544   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10545   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10546   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10547   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   10548   __ret; \
   10549 })
   10550 #endif
   10551 
   10552 #ifdef __LITTLE_ENDIAN__
   10553 #define vld4_p16(__p0) __extension__ ({ \
   10554   poly16x4x4_t __ret; \
   10555   __builtin_neon_vld4_v(&__ret, __p0, 5); \
   10556   __ret; \
   10557 })
   10558 #else
   10559 #define vld4_p16(__p0) __extension__ ({ \
   10560   poly16x4x4_t __ret; \
   10561   __builtin_neon_vld4_v(&__ret, __p0, 5); \
   10562  \
   10563   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10564   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10565   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10566   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   10567   __ret; \
   10568 })
   10569 #endif
   10570 
   10571 #ifdef __LITTLE_ENDIAN__
   10572 #define vld4q_p8(__p0) __extension__ ({ \
   10573   poly8x16x4_t __ret; \
   10574   __builtin_neon_vld4q_v(&__ret, __p0, 36); \
   10575   __ret; \
   10576 })
   10577 #else
   10578 #define vld4q_p8(__p0) __extension__ ({ \
   10579   poly8x16x4_t __ret; \
   10580   __builtin_neon_vld4q_v(&__ret, __p0, 36); \
   10581  \
   10582   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10583   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10584   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10585   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10586   __ret; \
   10587 })
   10588 #endif
   10589 
   10590 #ifdef __LITTLE_ENDIAN__
   10591 #define vld4q_p16(__p0) __extension__ ({ \
   10592   poly16x8x4_t __ret; \
   10593   __builtin_neon_vld4q_v(&__ret, __p0, 37); \
   10594   __ret; \
   10595 })
   10596 #else
   10597 #define vld4q_p16(__p0) __extension__ ({ \
   10598   poly16x8x4_t __ret; \
   10599   __builtin_neon_vld4q_v(&__ret, __p0, 37); \
   10600  \
   10601   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10602   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10603   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10604   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   10605   __ret; \
   10606 })
   10607 #endif
   10608 
   10609 #ifdef __LITTLE_ENDIAN__
   10610 #define vld4q_u8(__p0) __extension__ ({ \
   10611   uint8x16x4_t __ret; \
   10612   __builtin_neon_vld4q_v(&__ret, __p0, 48); \
   10613   __ret; \
   10614 })
   10615 #else
   10616 #define vld4q_u8(__p0) __extension__ ({ \
   10617   uint8x16x4_t __ret; \
   10618   __builtin_neon_vld4q_v(&__ret, __p0, 48); \
   10619  \
   10620   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10621   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10622   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10623   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10624   __ret; \
   10625 })
   10626 #endif
   10627 
   10628 #ifdef __LITTLE_ENDIAN__
   10629 #define vld4q_u32(__p0) __extension__ ({ \
   10630   uint32x4x4_t __ret; \
   10631   __builtin_neon_vld4q_v(&__ret, __p0, 50); \
   10632   __ret; \
   10633 })
   10634 #else
   10635 #define vld4q_u32(__p0) __extension__ ({ \
   10636   uint32x4x4_t __ret; \
   10637   __builtin_neon_vld4q_v(&__ret, __p0, 50); \
   10638  \
   10639   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10640   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10641   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10642   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   10643   __ret; \
   10644 })
   10645 #endif
   10646 
   10647 #ifdef __LITTLE_ENDIAN__
   10648 #define vld4q_u16(__p0) __extension__ ({ \
   10649   uint16x8x4_t __ret; \
   10650   __builtin_neon_vld4q_v(&__ret, __p0, 49); \
   10651   __ret; \
   10652 })
   10653 #else
   10654 #define vld4q_u16(__p0) __extension__ ({ \
   10655   uint16x8x4_t __ret; \
   10656   __builtin_neon_vld4q_v(&__ret, __p0, 49); \
   10657  \
   10658   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10659   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10660   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10661   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   10662   __ret; \
   10663 })
   10664 #endif
   10665 
   10666 #ifdef __LITTLE_ENDIAN__
   10667 #define vld4q_s8(__p0) __extension__ ({ \
   10668   int8x16x4_t __ret; \
   10669   __builtin_neon_vld4q_v(&__ret, __p0, 32); \
   10670   __ret; \
   10671 })
   10672 #else
   10673 #define vld4q_s8(__p0) __extension__ ({ \
   10674   int8x16x4_t __ret; \
   10675   __builtin_neon_vld4q_v(&__ret, __p0, 32); \
   10676  \
   10677   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10678   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10679   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10680   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   10681   __ret; \
   10682 })
   10683 #endif
   10684 
   10685 #ifdef __LITTLE_ENDIAN__
   10686 #define vld4q_f32(__p0) __extension__ ({ \
   10687   float32x4x4_t __ret; \
   10688   __builtin_neon_vld4q_v(&__ret, __p0, 41); \
   10689   __ret; \
   10690 })
   10691 #else
   10692 #define vld4q_f32(__p0) __extension__ ({ \
   10693   float32x4x4_t __ret; \
   10694   __builtin_neon_vld4q_v(&__ret, __p0, 41); \
   10695  \
   10696   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10697   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10698   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10699   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   10700   __ret; \
   10701 })
   10702 #endif
   10703 
   10704 #ifdef __LITTLE_ENDIAN__
   10705 #define vld4q_f16(__p0) __extension__ ({ \
   10706   float16x8x4_t __ret; \
   10707   __builtin_neon_vld4q_v(&__ret, __p0, 40); \
   10708   __ret; \
   10709 })
   10710 #else
   10711 #define vld4q_f16(__p0) __extension__ ({ \
   10712   float16x8x4_t __ret; \
   10713   __builtin_neon_vld4q_v(&__ret, __p0, 40); \
   10714  \
   10715   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10716   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10717   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10718   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   10719   __ret; \
   10720 })
   10721 #endif
   10722 
   10723 #ifdef __LITTLE_ENDIAN__
   10724 #define vld4q_s32(__p0) __extension__ ({ \
   10725   int32x4x4_t __ret; \
   10726   __builtin_neon_vld4q_v(&__ret, __p0, 34); \
   10727   __ret; \
   10728 })
   10729 #else
   10730 #define vld4q_s32(__p0) __extension__ ({ \
   10731   int32x4x4_t __ret; \
   10732   __builtin_neon_vld4q_v(&__ret, __p0, 34); \
   10733  \
   10734   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10735   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10736   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10737   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   10738   __ret; \
   10739 })
   10740 #endif
   10741 
   10742 #ifdef __LITTLE_ENDIAN__
   10743 #define vld4q_s16(__p0) __extension__ ({ \
   10744   int16x8x4_t __ret; \
   10745   __builtin_neon_vld4q_v(&__ret, __p0, 33); \
   10746   __ret; \
   10747 })
   10748 #else
   10749 #define vld4q_s16(__p0) __extension__ ({ \
   10750   int16x8x4_t __ret; \
   10751   __builtin_neon_vld4q_v(&__ret, __p0, 33); \
   10752  \
   10753   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10754   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10755   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10756   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   10757   __ret; \
   10758 })
   10759 #endif
   10760 
   10761 #ifdef __LITTLE_ENDIAN__
   10762 #define vld4_u8(__p0) __extension__ ({ \
   10763   uint8x8x4_t __ret; \
   10764   __builtin_neon_vld4_v(&__ret, __p0, 16); \
   10765   __ret; \
   10766 })
   10767 #else
   10768 #define vld4_u8(__p0) __extension__ ({ \
   10769   uint8x8x4_t __ret; \
   10770   __builtin_neon_vld4_v(&__ret, __p0, 16); \
   10771  \
   10772   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10773   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10774   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10775   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   10776   __ret; \
   10777 })
   10778 #endif
   10779 
   10780 #ifdef __LITTLE_ENDIAN__
   10781 #define vld4_u32(__p0) __extension__ ({ \
   10782   uint32x2x4_t __ret; \
   10783   __builtin_neon_vld4_v(&__ret, __p0, 18); \
   10784   __ret; \
   10785 })
   10786 #else
   10787 #define vld4_u32(__p0) __extension__ ({ \
   10788   uint32x2x4_t __ret; \
   10789   __builtin_neon_vld4_v(&__ret, __p0, 18); \
   10790  \
   10791   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   10792   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   10793   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   10794   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   10795   __ret; \
   10796 })
   10797 #endif
   10798 
   10799 #ifdef __LITTLE_ENDIAN__
   10800 #define vld4_u64(__p0) __extension__ ({ \
   10801   uint64x1x4_t __ret; \
   10802   __builtin_neon_vld4_v(&__ret, __p0, 19); \
   10803   __ret; \
   10804 })
   10805 #else
   10806 #define vld4_u64(__p0) __extension__ ({ \
   10807   uint64x1x4_t __ret; \
   10808   __builtin_neon_vld4_v(&__ret, __p0, 19); \
   10809   __ret; \
   10810 })
   10811 #endif
   10812 
   10813 #ifdef __LITTLE_ENDIAN__
   10814 #define vld4_u16(__p0) __extension__ ({ \
   10815   uint16x4x4_t __ret; \
   10816   __builtin_neon_vld4_v(&__ret, __p0, 17); \
   10817   __ret; \
   10818 })
   10819 #else
   10820 #define vld4_u16(__p0) __extension__ ({ \
   10821   uint16x4x4_t __ret; \
   10822   __builtin_neon_vld4_v(&__ret, __p0, 17); \
   10823  \
   10824   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10825   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10826   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10827   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   10828   __ret; \
   10829 })
   10830 #endif
   10831 
   10832 #ifdef __LITTLE_ENDIAN__
   10833 #define vld4_s8(__p0) __extension__ ({ \
   10834   int8x8x4_t __ret; \
   10835   __builtin_neon_vld4_v(&__ret, __p0, 0); \
   10836   __ret; \
   10837 })
   10838 #else
   10839 #define vld4_s8(__p0) __extension__ ({ \
   10840   int8x8x4_t __ret; \
   10841   __builtin_neon_vld4_v(&__ret, __p0, 0); \
   10842  \
   10843   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10844   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10845   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10846   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   10847   __ret; \
   10848 })
   10849 #endif
   10850 
   10851 #ifdef __LITTLE_ENDIAN__
   10852 #define vld4_f32(__p0) __extension__ ({ \
   10853   float32x2x4_t __ret; \
   10854   __builtin_neon_vld4_v(&__ret, __p0, 9); \
   10855   __ret; \
   10856 })
   10857 #else
   10858 #define vld4_f32(__p0) __extension__ ({ \
   10859   float32x2x4_t __ret; \
   10860   __builtin_neon_vld4_v(&__ret, __p0, 9); \
   10861  \
   10862   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   10863   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   10864   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   10865   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   10866   __ret; \
   10867 })
   10868 #endif
   10869 
   10870 #ifdef __LITTLE_ENDIAN__
   10871 #define vld4_f16(__p0) __extension__ ({ \
   10872   float16x4x4_t __ret; \
   10873   __builtin_neon_vld4_v(&__ret, __p0, 8); \
   10874   __ret; \
   10875 })
   10876 #else
   10877 #define vld4_f16(__p0) __extension__ ({ \
   10878   float16x4x4_t __ret; \
   10879   __builtin_neon_vld4_v(&__ret, __p0, 8); \
   10880  \
   10881   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10882   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10883   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10884   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   10885   __ret; \
   10886 })
   10887 #endif
   10888 
   10889 #ifdef __LITTLE_ENDIAN__
   10890 #define vld4_s32(__p0) __extension__ ({ \
   10891   int32x2x4_t __ret; \
   10892   __builtin_neon_vld4_v(&__ret, __p0, 2); \
   10893   __ret; \
   10894 })
   10895 #else
   10896 #define vld4_s32(__p0) __extension__ ({ \
   10897   int32x2x4_t __ret; \
   10898   __builtin_neon_vld4_v(&__ret, __p0, 2); \
   10899  \
   10900   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   10901   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   10902   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   10903   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   10904   __ret; \
   10905 })
   10906 #endif
   10907 
   10908 #ifdef __LITTLE_ENDIAN__
   10909 #define vld4_s64(__p0) __extension__ ({ \
   10910   int64x1x4_t __ret; \
   10911   __builtin_neon_vld4_v(&__ret, __p0, 3); \
   10912   __ret; \
   10913 })
   10914 #else
   10915 #define vld4_s64(__p0) __extension__ ({ \
   10916   int64x1x4_t __ret; \
   10917   __builtin_neon_vld4_v(&__ret, __p0, 3); \
   10918   __ret; \
   10919 })
   10920 #endif
   10921 
   10922 #ifdef __LITTLE_ENDIAN__
   10923 #define vld4_s16(__p0) __extension__ ({ \
   10924   int16x4x4_t __ret; \
   10925   __builtin_neon_vld4_v(&__ret, __p0, 1); \
   10926   __ret; \
   10927 })
   10928 #else
   10929 #define vld4_s16(__p0) __extension__ ({ \
   10930   int16x4x4_t __ret; \
   10931   __builtin_neon_vld4_v(&__ret, __p0, 1); \
   10932  \
   10933   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10934   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10935   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10936   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   10937   __ret; \
   10938 })
   10939 #endif
   10940 
   10941 #ifdef __LITTLE_ENDIAN__
   10942 #define vld4_dup_p8(__p0) __extension__ ({ \
   10943   poly8x8x4_t __ret; \
   10944   __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
   10945   __ret; \
   10946 })
   10947 #else
   10948 #define vld4_dup_p8(__p0) __extension__ ({ \
   10949   poly8x8x4_t __ret; \
   10950   __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
   10951  \
   10952   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10953   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10954   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10955   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   10956   __ret; \
   10957 })
   10958 #endif
   10959 
   10960 #ifdef __LITTLE_ENDIAN__
   10961 #define vld4_dup_p16(__p0) __extension__ ({ \
   10962   poly16x4x4_t __ret; \
   10963   __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
   10964   __ret; \
   10965 })
   10966 #else
   10967 #define vld4_dup_p16(__p0) __extension__ ({ \
   10968   poly16x4x4_t __ret; \
   10969   __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
   10970  \
   10971   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   10972   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   10973   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   10974   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   10975   __ret; \
   10976 })
   10977 #endif
   10978 
   10979 #ifdef __LITTLE_ENDIAN__
   10980 #define vld4_dup_u8(__p0) __extension__ ({ \
   10981   uint8x8x4_t __ret; \
   10982   __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
   10983   __ret; \
   10984 })
   10985 #else
   10986 #define vld4_dup_u8(__p0) __extension__ ({ \
   10987   uint8x8x4_t __ret; \
   10988   __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
   10989  \
   10990   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   10991   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   10992   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   10993   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   10994   __ret; \
   10995 })
   10996 #endif
   10997 
   10998 #ifdef __LITTLE_ENDIAN__
   10999 #define vld4_dup_u32(__p0) __extension__ ({ \
   11000   uint32x2x4_t __ret; \
   11001   __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
   11002   __ret; \
   11003 })
   11004 #else
   11005 #define vld4_dup_u32(__p0) __extension__ ({ \
   11006   uint32x2x4_t __ret; \
   11007   __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
   11008  \
   11009   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   11010   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   11011   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   11012   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   11013   __ret; \
   11014 })
   11015 #endif
   11016 
   11017 #ifdef __LITTLE_ENDIAN__
   11018 #define vld4_dup_u64(__p0) __extension__ ({ \
   11019   uint64x1x4_t __ret; \
   11020   __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
   11021   __ret; \
   11022 })
   11023 #else
   11024 #define vld4_dup_u64(__p0) __extension__ ({ \
   11025   uint64x1x4_t __ret; \
   11026   __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
   11027   __ret; \
   11028 })
   11029 #endif
   11030 
   11031 #ifdef __LITTLE_ENDIAN__
   11032 #define vld4_dup_u16(__p0) __extension__ ({ \
   11033   uint16x4x4_t __ret; \
   11034   __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
   11035   __ret; \
   11036 })
   11037 #else
   11038 #define vld4_dup_u16(__p0) __extension__ ({ \
   11039   uint16x4x4_t __ret; \
   11040   __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
   11041  \
   11042   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11043   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11044   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11045   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11046   __ret; \
   11047 })
   11048 #endif
   11049 
   11050 #ifdef __LITTLE_ENDIAN__
   11051 #define vld4_dup_s8(__p0) __extension__ ({ \
   11052   int8x8x4_t __ret; \
   11053   __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
   11054   __ret; \
   11055 })
   11056 #else
   11057 #define vld4_dup_s8(__p0) __extension__ ({ \
   11058   int8x8x4_t __ret; \
   11059   __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
   11060  \
   11061   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11062   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11063   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11064   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11065   __ret; \
   11066 })
   11067 #endif
   11068 
   11069 #ifdef __LITTLE_ENDIAN__
   11070 #define vld4_dup_f32(__p0) __extension__ ({ \
   11071   float32x2x4_t __ret; \
   11072   __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
   11073   __ret; \
   11074 })
   11075 #else
   11076 #define vld4_dup_f32(__p0) __extension__ ({ \
   11077   float32x2x4_t __ret; \
   11078   __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
   11079  \
   11080   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   11081   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   11082   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   11083   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   11084   __ret; \
   11085 })
   11086 #endif
   11087 
   11088 #ifdef __LITTLE_ENDIAN__
   11089 #define vld4_dup_f16(__p0) __extension__ ({ \
   11090   float16x4x4_t __ret; \
   11091   __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
   11092   __ret; \
   11093 })
   11094 #else
   11095 #define vld4_dup_f16(__p0) __extension__ ({ \
   11096   float16x4x4_t __ret; \
   11097   __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
   11098  \
   11099   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11100   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11101   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11102   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11103   __ret; \
   11104 })
   11105 #endif
   11106 
   11107 #ifdef __LITTLE_ENDIAN__
   11108 #define vld4_dup_s32(__p0) __extension__ ({ \
   11109   int32x2x4_t __ret; \
   11110   __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
   11111   __ret; \
   11112 })
   11113 #else
   11114 #define vld4_dup_s32(__p0) __extension__ ({ \
   11115   int32x2x4_t __ret; \
   11116   __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
   11117  \
   11118   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   11119   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   11120   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   11121   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   11122   __ret; \
   11123 })
   11124 #endif
   11125 
   11126 #ifdef __LITTLE_ENDIAN__
   11127 #define vld4_dup_s64(__p0) __extension__ ({ \
   11128   int64x1x4_t __ret; \
   11129   __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
   11130   __ret; \
   11131 })
   11132 #else
   11133 #define vld4_dup_s64(__p0) __extension__ ({ \
   11134   int64x1x4_t __ret; \
   11135   __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
   11136   __ret; \
   11137 })
   11138 #endif
   11139 
   11140 #ifdef __LITTLE_ENDIAN__
   11141 #define vld4_dup_s16(__p0) __extension__ ({ \
   11142   int16x4x4_t __ret; \
   11143   __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
   11144   __ret; \
   11145 })
   11146 #else
   11147 #define vld4_dup_s16(__p0) __extension__ ({ \
   11148   int16x4x4_t __ret; \
   11149   __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
   11150  \
   11151   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11152   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11153   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11154   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11155   __ret; \
   11156 })
   11157 #endif
   11158 
   11159 #ifdef __LITTLE_ENDIAN__
   11160 #define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   11161   poly8x8x4_t __s1 = __p1; \
   11162   poly8x8x4_t __ret; \
   11163   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
   11164   __ret; \
   11165 })
   11166 #else
   11167 #define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   11168   poly8x8x4_t __s1 = __p1; \
   11169   poly8x8x4_t __rev1; \
   11170   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11171   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11172   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11173   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11174   poly8x8x4_t __ret; \
   11175   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
   11176  \
   11177   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11178   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11179   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11180   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11181   __ret; \
   11182 })
   11183 #endif
   11184 
   11185 #ifdef __LITTLE_ENDIAN__
   11186 #define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   11187   poly16x4x4_t __s1 = __p1; \
   11188   poly16x4x4_t __ret; \
   11189   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
   11190   __ret; \
   11191 })
   11192 #else
   11193 #define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   11194   poly16x4x4_t __s1 = __p1; \
   11195   poly16x4x4_t __rev1; \
   11196   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   11197   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   11198   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   11199   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   11200   poly16x4x4_t __ret; \
   11201   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
   11202  \
   11203   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11204   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11205   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11206   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11207   __ret; \
   11208 })
   11209 #endif
   11210 
   11211 #ifdef __LITTLE_ENDIAN__
   11212 #define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   11213   poly16x8x4_t __s1 = __p1; \
   11214   poly16x8x4_t __ret; \
   11215   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
   11216   __ret; \
   11217 })
   11218 #else
   11219 #define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   11220   poly16x8x4_t __s1 = __p1; \
   11221   poly16x8x4_t __rev1; \
   11222   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11223   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11224   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11225   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11226   poly16x8x4_t __ret; \
   11227   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
   11228  \
   11229   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11230   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11231   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11232   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11233   __ret; \
   11234 })
   11235 #endif
   11236 
   11237 #ifdef __LITTLE_ENDIAN__
   11238 #define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   11239   uint32x4x4_t __s1 = __p1; \
   11240   uint32x4x4_t __ret; \
   11241   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
   11242   __ret; \
   11243 })
   11244 #else
   11245 #define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   11246   uint32x4x4_t __s1 = __p1; \
   11247   uint32x4x4_t __rev1; \
   11248   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   11249   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   11250   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   11251   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   11252   uint32x4x4_t __ret; \
   11253   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
   11254  \
   11255   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11256   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11257   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11258   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11259   __ret; \
   11260 })
   11261 #endif
   11262 
   11263 #ifdef __LITTLE_ENDIAN__
   11264 #define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   11265   uint16x8x4_t __s1 = __p1; \
   11266   uint16x8x4_t __ret; \
   11267   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
   11268   __ret; \
   11269 })
   11270 #else
   11271 #define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   11272   uint16x8x4_t __s1 = __p1; \
   11273   uint16x8x4_t __rev1; \
   11274   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11275   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11276   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11277   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11278   uint16x8x4_t __ret; \
   11279   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
   11280  \
   11281   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11282   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11283   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11284   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11285   __ret; \
   11286 })
   11287 #endif
   11288 
   11289 #ifdef __LITTLE_ENDIAN__
   11290 #define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   11291   float32x4x4_t __s1 = __p1; \
   11292   float32x4x4_t __ret; \
   11293   __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
   11294   __ret; \
   11295 })
   11296 #else
   11297 #define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   11298   float32x4x4_t __s1 = __p1; \
   11299   float32x4x4_t __rev1; \
   11300   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   11301   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   11302   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   11303   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   11304   float32x4x4_t __ret; \
   11305   __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
   11306  \
   11307   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11308   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11309   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11310   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11311   __ret; \
   11312 })
   11313 #endif
   11314 
   11315 #ifdef __LITTLE_ENDIAN__
   11316 #define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   11317   float16x8x4_t __s1 = __p1; \
   11318   float16x8x4_t __ret; \
   11319   __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
   11320   __ret; \
   11321 })
   11322 #else
   11323 #define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   11324   float16x8x4_t __s1 = __p1; \
   11325   float16x8x4_t __rev1; \
   11326   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11327   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11328   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11329   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11330   float16x8x4_t __ret; \
   11331   __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
   11332  \
   11333   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11334   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11335   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11336   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11337   __ret; \
   11338 })
   11339 #endif
   11340 
   11341 #ifdef __LITTLE_ENDIAN__
   11342 #define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   11343   int32x4x4_t __s1 = __p1; \
   11344   int32x4x4_t __ret; \
   11345   __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
   11346   __ret; \
   11347 })
   11348 #else
   11349 #define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   11350   int32x4x4_t __s1 = __p1; \
   11351   int32x4x4_t __rev1; \
   11352   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   11353   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   11354   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   11355   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   11356   int32x4x4_t __ret; \
   11357   __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
   11358  \
   11359   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11360   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11361   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11362   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11363   __ret; \
   11364 })
   11365 #endif
   11366 
   11367 #ifdef __LITTLE_ENDIAN__
   11368 #define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   11369   int16x8x4_t __s1 = __p1; \
   11370   int16x8x4_t __ret; \
   11371   __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
   11372   __ret; \
   11373 })
   11374 #else
   11375 #define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   11376   int16x8x4_t __s1 = __p1; \
   11377   int16x8x4_t __rev1; \
   11378   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11379   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11380   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11381   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11382   int16x8x4_t __ret; \
   11383   __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
   11384  \
   11385   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11386   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11387   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11388   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11389   __ret; \
   11390 })
   11391 #endif
   11392 
   11393 #ifdef __LITTLE_ENDIAN__
   11394 #define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   11395   uint8x8x4_t __s1 = __p1; \
   11396   uint8x8x4_t __ret; \
   11397   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
   11398   __ret; \
   11399 })
   11400 #else
   11401 #define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   11402   uint8x8x4_t __s1 = __p1; \
   11403   uint8x8x4_t __rev1; \
   11404   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11405   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11406   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11407   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11408   uint8x8x4_t __ret; \
   11409   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
   11410  \
   11411   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11412   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11413   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11414   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11415   __ret; \
   11416 })
   11417 #endif
   11418 
   11419 #ifdef __LITTLE_ENDIAN__
   11420 #define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   11421   uint32x2x4_t __s1 = __p1; \
   11422   uint32x2x4_t __ret; \
   11423   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
   11424   __ret; \
   11425 })
   11426 #else
   11427 #define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   11428   uint32x2x4_t __s1 = __p1; \
   11429   uint32x2x4_t __rev1; \
   11430   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   11431   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   11432   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   11433   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   11434   uint32x2x4_t __ret; \
   11435   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
   11436  \
   11437   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   11438   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   11439   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   11440   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   11441   __ret; \
   11442 })
   11443 #endif
   11444 
   11445 #ifdef __LITTLE_ENDIAN__
   11446 #define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   11447   uint16x4x4_t __s1 = __p1; \
   11448   uint16x4x4_t __ret; \
   11449   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
   11450   __ret; \
   11451 })
   11452 #else
   11453 #define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   11454   uint16x4x4_t __s1 = __p1; \
   11455   uint16x4x4_t __rev1; \
   11456   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   11457   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   11458   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   11459   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   11460   uint16x4x4_t __ret; \
   11461   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
   11462  \
   11463   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11464   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11465   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11466   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11467   __ret; \
   11468 })
   11469 #endif
   11470 
   11471 #ifdef __LITTLE_ENDIAN__
   11472 #define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   11473   int8x8x4_t __s1 = __p1; \
   11474   int8x8x4_t __ret; \
   11475   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
   11476   __ret; \
   11477 })
   11478 #else
   11479 #define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   11480   int8x8x4_t __s1 = __p1; \
   11481   int8x8x4_t __rev1; \
   11482   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11483   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11484   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11485   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11486   int8x8x4_t __ret; \
   11487   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
   11488  \
   11489   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   11490   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   11491   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   11492   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   11493   __ret; \
   11494 })
   11495 #endif
   11496 
   11497 #ifdef __LITTLE_ENDIAN__
   11498 #define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   11499   float32x2x4_t __s1 = __p1; \
   11500   float32x2x4_t __ret; \
   11501   __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
   11502   __ret; \
   11503 })
   11504 #else
   11505 #define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   11506   float32x2x4_t __s1 = __p1; \
   11507   float32x2x4_t __rev1; \
   11508   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   11509   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   11510   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   11511   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   11512   float32x2x4_t __ret; \
   11513   __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
   11514  \
   11515   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   11516   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   11517   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   11518   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   11519   __ret; \
   11520 })
   11521 #endif
   11522 
   11523 #ifdef __LITTLE_ENDIAN__
   11524 #define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   11525   float16x4x4_t __s1 = __p1; \
   11526   float16x4x4_t __ret; \
   11527   __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
   11528   __ret; \
   11529 })
   11530 #else
   11531 #define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   11532   float16x4x4_t __s1 = __p1; \
   11533   float16x4x4_t __rev1; \
   11534   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   11535   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   11536   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   11537   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   11538   float16x4x4_t __ret; \
   11539   __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
   11540  \
   11541   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11542   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11543   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11544   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11545   __ret; \
   11546 })
   11547 #endif
   11548 
   11549 #ifdef __LITTLE_ENDIAN__
   11550 #define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   11551   int32x2x4_t __s1 = __p1; \
   11552   int32x2x4_t __ret; \
   11553   __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
   11554   __ret; \
   11555 })
   11556 #else
   11557 #define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   11558   int32x2x4_t __s1 = __p1; \
   11559   int32x2x4_t __rev1; \
   11560   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   11561   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   11562   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   11563   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   11564   int32x2x4_t __ret; \
   11565   __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
   11566  \
   11567   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   11568   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   11569   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   11570   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   11571   __ret; \
   11572 })
   11573 #endif
   11574 
   11575 #ifdef __LITTLE_ENDIAN__
   11576 #define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   11577   int16x4x4_t __s1 = __p1; \
   11578   int16x4x4_t __ret; \
   11579   __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
   11580   __ret; \
   11581 })
   11582 #else
   11583 #define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   11584   int16x4x4_t __s1 = __p1; \
   11585   int16x4x4_t __rev1; \
   11586   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   11587   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   11588   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   11589   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   11590   int16x4x4_t __ret; \
   11591   __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
   11592  \
   11593   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   11594   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   11595   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   11596   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   11597   __ret; \
   11598 })
   11599 #endif
   11600 
   11601 #ifdef __LITTLE_ENDIAN__
   11602 __ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   11603   uint8x16_t __ret;
   11604   __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   11605   return __ret;
   11606 }
   11607 #else
   11608 __ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   11609   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11610   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11611   uint8x16_t __ret;
   11612   __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   11613   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11614   return __ret;
   11615 }
   11616 #endif
   11617 
   11618 #ifdef __LITTLE_ENDIAN__
   11619 __ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   11620   uint32x4_t __ret;
   11621   __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   11622   return __ret;
   11623 }
   11624 #else
   11625 __ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   11626   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   11627   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   11628   uint32x4_t __ret;
   11629   __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   11630   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   11631   return __ret;
   11632 }
   11633 #endif
   11634 
   11635 #ifdef __LITTLE_ENDIAN__
   11636 __ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   11637   uint16x8_t __ret;
   11638   __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   11639   return __ret;
   11640 }
   11641 #else
   11642 __ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   11643   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   11644   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   11645   uint16x8_t __ret;
   11646   __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   11647   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   11648   return __ret;
   11649 }
   11650 #endif
   11651 
   11652 #ifdef __LITTLE_ENDIAN__
   11653 __ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
   11654   int8x16_t __ret;
   11655   __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   11656   return __ret;
   11657 }
   11658 #else
   11659 __ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
   11660   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11661   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11662   int8x16_t __ret;
   11663   __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   11664   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11665   return __ret;
   11666 }
   11667 #endif
   11668 
   11669 #ifdef __LITTLE_ENDIAN__
   11670 __ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
   11671   float32x4_t __ret;
   11672   __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   11673   return __ret;
   11674 }
   11675 #else
   11676 __ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
   11677   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   11678   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   11679   float32x4_t __ret;
   11680   __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   11681   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   11682   return __ret;
   11683 }
   11684 #endif
   11685 
   11686 #ifdef __LITTLE_ENDIAN__
   11687 __ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
   11688   int32x4_t __ret;
   11689   __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   11690   return __ret;
   11691 }
   11692 #else
   11693 __ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
   11694   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   11695   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   11696   int32x4_t __ret;
   11697   __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   11698   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   11699   return __ret;
   11700 }
   11701 #endif
   11702 
   11703 #ifdef __LITTLE_ENDIAN__
   11704 __ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
   11705   int16x8_t __ret;
   11706   __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   11707   return __ret;
   11708 }
   11709 #else
   11710 __ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
   11711   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   11712   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   11713   int16x8_t __ret;
   11714   __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   11715   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   11716   return __ret;
   11717 }
   11718 #endif
   11719 
   11720 #ifdef __LITTLE_ENDIAN__
   11721 __ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
   11722   uint8x8_t __ret;
   11723   __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   11724   return __ret;
   11725 }
   11726 #else
   11727 __ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
   11728   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   11729   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   11730   uint8x8_t __ret;
   11731   __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   11732   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   11733   return __ret;
   11734 }
   11735 #endif
   11736 
   11737 #ifdef __LITTLE_ENDIAN__
   11738 __ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
   11739   uint32x2_t __ret;
   11740   __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   11741   return __ret;
   11742 }
   11743 #else
   11744 __ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
   11745   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   11746   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   11747   uint32x2_t __ret;
   11748   __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   11749   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   11750   return __ret;
   11751 }
   11752 #endif
   11753 
   11754 #ifdef __LITTLE_ENDIAN__
   11755 __ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
   11756   uint16x4_t __ret;
   11757   __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   11758   return __ret;
   11759 }
   11760 #else
   11761 __ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
   11762   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   11763   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   11764   uint16x4_t __ret;
   11765   __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   11766   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   11767   return __ret;
   11768 }
   11769 #endif
   11770 
   11771 #ifdef __LITTLE_ENDIAN__
   11772 __ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
   11773   int8x8_t __ret;
   11774   __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   11775   return __ret;
   11776 }
   11777 #else
   11778 __ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
   11779   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   11780   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   11781   int8x8_t __ret;
   11782   __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   11783   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   11784   return __ret;
   11785 }
   11786 #endif
   11787 
   11788 #ifdef __LITTLE_ENDIAN__
   11789 __ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
   11790   float32x2_t __ret;
   11791   __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   11792   return __ret;
   11793 }
   11794 #else
   11795 __ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
   11796   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   11797   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   11798   float32x2_t __ret;
   11799   __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   11800   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   11801   return __ret;
   11802 }
   11803 #endif
   11804 
   11805 #ifdef __LITTLE_ENDIAN__
   11806 __ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
   11807   int32x2_t __ret;
   11808   __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   11809   return __ret;
   11810 }
   11811 #else
   11812 __ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
   11813   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   11814   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   11815   int32x2_t __ret;
   11816   __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   11817   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   11818   return __ret;
   11819 }
   11820 #endif
   11821 
   11822 #ifdef __LITTLE_ENDIAN__
   11823 __ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
   11824   int16x4_t __ret;
   11825   __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   11826   return __ret;
   11827 }
   11828 #else
   11829 __ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
   11830   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   11831   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   11832   int16x4_t __ret;
   11833   __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   11834   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   11835   return __ret;
   11836 }
   11837 #endif
   11838 
   11839 #ifdef __LITTLE_ENDIAN__
   11840 __ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   11841   uint8x16_t __ret;
   11842   __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   11843   return __ret;
   11844 }
   11845 #else
   11846 __ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   11847   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11848   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11849   uint8x16_t __ret;
   11850   __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   11851   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11852   return __ret;
   11853 }
   11854 #endif
   11855 
   11856 #ifdef __LITTLE_ENDIAN__
   11857 __ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   11858   uint32x4_t __ret;
   11859   __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   11860   return __ret;
   11861 }
   11862 #else
   11863 __ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   11864   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   11865   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   11866   uint32x4_t __ret;
   11867   __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   11868   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   11869   return __ret;
   11870 }
   11871 #endif
   11872 
   11873 #ifdef __LITTLE_ENDIAN__
   11874 __ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   11875   uint16x8_t __ret;
   11876   __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   11877   return __ret;
   11878 }
   11879 #else
   11880 __ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   11881   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   11882   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   11883   uint16x8_t __ret;
   11884   __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   11885   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   11886   return __ret;
   11887 }
   11888 #endif
   11889 
   11890 #ifdef __LITTLE_ENDIAN__
   11891 __ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
   11892   int8x16_t __ret;
   11893   __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   11894   return __ret;
   11895 }
   11896 #else
   11897 __ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
   11898   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11899   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11900   int8x16_t __ret;
   11901   __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   11902   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   11903   return __ret;
   11904 }
   11905 #endif
   11906 
   11907 #ifdef __LITTLE_ENDIAN__
   11908 __ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
   11909   float32x4_t __ret;
   11910   __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   11911   return __ret;
   11912 }
   11913 #else
   11914 __ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
   11915   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   11916   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   11917   float32x4_t __ret;
   11918   __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   11919   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   11920   return __ret;
   11921 }
   11922 #endif
   11923 
   11924 #ifdef __LITTLE_ENDIAN__
   11925 __ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
   11926   int32x4_t __ret;
   11927   __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   11928   return __ret;
   11929 }
   11930 #else
   11931 __ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
   11932   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   11933   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   11934   int32x4_t __ret;
   11935   __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   11936   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   11937   return __ret;
   11938 }
   11939 #endif
   11940 
   11941 #ifdef __LITTLE_ENDIAN__
   11942 __ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
   11943   int16x8_t __ret;
   11944   __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   11945   return __ret;
   11946 }
   11947 #else
   11948 __ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
   11949   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   11950   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   11951   int16x8_t __ret;
   11952   __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   11953   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   11954   return __ret;
   11955 }
   11956 #endif
   11957 
   11958 #ifdef __LITTLE_ENDIAN__
   11959 __ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
   11960   uint8x8_t __ret;
   11961   __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   11962   return __ret;
   11963 }
   11964 #else
   11965 __ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
   11966   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   11967   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   11968   uint8x8_t __ret;
   11969   __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   11970   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   11971   return __ret;
   11972 }
   11973 #endif
   11974 
   11975 #ifdef __LITTLE_ENDIAN__
   11976 __ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
   11977   uint32x2_t __ret;
   11978   __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   11979   return __ret;
   11980 }
   11981 #else
   11982 __ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
   11983   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   11984   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   11985   uint32x2_t __ret;
   11986   __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   11987   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   11988   return __ret;
   11989 }
   11990 #endif
   11991 
   11992 #ifdef __LITTLE_ENDIAN__
   11993 __ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
   11994   uint16x4_t __ret;
   11995   __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   11996   return __ret;
   11997 }
   11998 #else
   11999 __ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
   12000   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12001   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12002   uint16x4_t __ret;
   12003   __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   12004   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12005   return __ret;
   12006 }
   12007 #endif
   12008 
   12009 #ifdef __LITTLE_ENDIAN__
   12010 __ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
   12011   int8x8_t __ret;
   12012   __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   12013   return __ret;
   12014 }
   12015 #else
   12016 __ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
   12017   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12018   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12019   int8x8_t __ret;
   12020   __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   12021   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12022   return __ret;
   12023 }
   12024 #endif
   12025 
   12026 #ifdef __LITTLE_ENDIAN__
   12027 __ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
   12028   float32x2_t __ret;
   12029   __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   12030   return __ret;
   12031 }
   12032 #else
   12033 __ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
   12034   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12035   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12036   float32x2_t __ret;
   12037   __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   12038   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12039   return __ret;
   12040 }
   12041 #endif
   12042 
   12043 #ifdef __LITTLE_ENDIAN__
   12044 __ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
   12045   int32x2_t __ret;
   12046   __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   12047   return __ret;
   12048 }
   12049 #else
   12050 __ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
   12051   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12052   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12053   int32x2_t __ret;
   12054   __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   12055   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12056   return __ret;
   12057 }
   12058 #endif
   12059 
   12060 #ifdef __LITTLE_ENDIAN__
   12061 __ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
   12062   int16x4_t __ret;
   12063   __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   12064   return __ret;
   12065 }
   12066 #else
   12067 __ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
   12068   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12069   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12070   int16x4_t __ret;
   12071   __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   12072   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12073   return __ret;
   12074 }
   12075 #endif
   12076 
   12077 #ifdef __LITTLE_ENDIAN__
   12078 __ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   12079   uint8x16_t __ret;
   12080   __ret = __p0 + __p1 * __p2;
   12081   return __ret;
   12082 }
   12083 #else
   12084 __ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   12085   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12086   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12087   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12088   uint8x16_t __ret;
   12089   __ret = __rev0 + __rev1 * __rev2;
   12090   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12091   return __ret;
   12092 }
   12093 #endif
   12094 
   12095 #ifdef __LITTLE_ENDIAN__
   12096 __ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   12097   uint32x4_t __ret;
   12098   __ret = __p0 + __p1 * __p2;
   12099   return __ret;
   12100 }
   12101 #else
   12102 __ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   12103   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12104   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12105   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12106   uint32x4_t __ret;
   12107   __ret = __rev0 + __rev1 * __rev2;
   12108   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12109   return __ret;
   12110 }
   12111 #endif
   12112 
   12113 #ifdef __LITTLE_ENDIAN__
   12114 __ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   12115   uint16x8_t __ret;
   12116   __ret = __p0 + __p1 * __p2;
   12117   return __ret;
   12118 }
   12119 #else
   12120 __ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   12121   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12122   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12123   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   12124   uint16x8_t __ret;
   12125   __ret = __rev0 + __rev1 * __rev2;
   12126   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12127   return __ret;
   12128 }
   12129 #endif
   12130 
   12131 #ifdef __LITTLE_ENDIAN__
   12132 __ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   12133   int8x16_t __ret;
   12134   __ret = __p0 + __p1 * __p2;
   12135   return __ret;
   12136 }
   12137 #else
   12138 __ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   12139   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12140   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12141   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12142   int8x16_t __ret;
   12143   __ret = __rev0 + __rev1 * __rev2;
   12144   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12145   return __ret;
   12146 }
   12147 #endif
   12148 
   12149 #ifdef __LITTLE_ENDIAN__
   12150 __ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   12151   float32x4_t __ret;
   12152   __ret = __p0 + __p1 * __p2;
   12153   return __ret;
   12154 }
   12155 #else
   12156 __ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   12157   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12158   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12159   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12160   float32x4_t __ret;
   12161   __ret = __rev0 + __rev1 * __rev2;
   12162   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12163   return __ret;
   12164 }
   12165 #endif
   12166 
   12167 #ifdef __LITTLE_ENDIAN__
   12168 __ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   12169   int32x4_t __ret;
   12170   __ret = __p0 + __p1 * __p2;
   12171   return __ret;
   12172 }
   12173 #else
   12174 __ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   12175   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12176   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12177   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12178   int32x4_t __ret;
   12179   __ret = __rev0 + __rev1 * __rev2;
   12180   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12181   return __ret;
   12182 }
   12183 #endif
   12184 
   12185 #ifdef __LITTLE_ENDIAN__
   12186 __ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   12187   int16x8_t __ret;
   12188   __ret = __p0 + __p1 * __p2;
   12189   return __ret;
   12190 }
   12191 #else
   12192 __ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   12193   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12194   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12195   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   12196   int16x8_t __ret;
   12197   __ret = __rev0 + __rev1 * __rev2;
   12198   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12199   return __ret;
   12200 }
   12201 #endif
   12202 
   12203 #ifdef __LITTLE_ENDIAN__
   12204 __ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   12205   uint8x8_t __ret;
   12206   __ret = __p0 + __p1 * __p2;
   12207   return __ret;
   12208 }
   12209 #else
   12210 __ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   12211   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12212   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12213   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   12214   uint8x8_t __ret;
   12215   __ret = __rev0 + __rev1 * __rev2;
   12216   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12217   return __ret;
   12218 }
   12219 #endif
   12220 
   12221 #ifdef __LITTLE_ENDIAN__
   12222 __ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   12223   uint32x2_t __ret;
   12224   __ret = __p0 + __p1 * __p2;
   12225   return __ret;
   12226 }
   12227 #else
   12228 __ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   12229   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12230   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12231   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   12232   uint32x2_t __ret;
   12233   __ret = __rev0 + __rev1 * __rev2;
   12234   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12235   return __ret;
   12236 }
   12237 #endif
   12238 
   12239 #ifdef __LITTLE_ENDIAN__
   12240 __ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   12241   uint16x4_t __ret;
   12242   __ret = __p0 + __p1 * __p2;
   12243   return __ret;
   12244 }
   12245 #else
   12246 __ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   12247   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12248   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12249   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12250   uint16x4_t __ret;
   12251   __ret = __rev0 + __rev1 * __rev2;
   12252   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12253   return __ret;
   12254 }
   12255 #endif
   12256 
   12257 #ifdef __LITTLE_ENDIAN__
   12258 __ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   12259   int8x8_t __ret;
   12260   __ret = __p0 + __p1 * __p2;
   12261   return __ret;
   12262 }
   12263 #else
   12264 __ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   12265   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12266   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12267   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   12268   int8x8_t __ret;
   12269   __ret = __rev0 + __rev1 * __rev2;
   12270   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12271   return __ret;
   12272 }
   12273 #endif
   12274 
   12275 #ifdef __LITTLE_ENDIAN__
   12276 __ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   12277   float32x2_t __ret;
   12278   __ret = __p0 + __p1 * __p2;
   12279   return __ret;
   12280 }
   12281 #else
   12282 __ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   12283   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12284   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12285   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   12286   float32x2_t __ret;
   12287   __ret = __rev0 + __rev1 * __rev2;
   12288   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12289   return __ret;
   12290 }
   12291 #endif
   12292 
   12293 #ifdef __LITTLE_ENDIAN__
   12294 __ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   12295   int32x2_t __ret;
   12296   __ret = __p0 + __p1 * __p2;
   12297   return __ret;
   12298 }
   12299 #else
   12300 __ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   12301   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12302   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12303   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   12304   int32x2_t __ret;
   12305   __ret = __rev0 + __rev1 * __rev2;
   12306   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12307   return __ret;
   12308 }
   12309 #endif
   12310 
   12311 #ifdef __LITTLE_ENDIAN__
   12312 __ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   12313   int16x4_t __ret;
   12314   __ret = __p0 + __p1 * __p2;
   12315   return __ret;
   12316 }
   12317 #else
   12318 __ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   12319   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12320   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12321   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12322   int16x4_t __ret;
   12323   __ret = __rev0 + __rev1 * __rev2;
   12324   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12325   return __ret;
   12326 }
   12327 #endif
   12328 
   12329 #ifdef __LITTLE_ENDIAN__
   12330 #define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12331   uint32x4_t __s0 = __p0; \
   12332   uint32x4_t __s1 = __p1; \
   12333   uint32x2_t __s2 = __p2; \
   12334   uint32x4_t __ret; \
   12335   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   12336   __ret; \
   12337 })
   12338 #else
   12339 #define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12340   uint32x4_t __s0 = __p0; \
   12341   uint32x4_t __s1 = __p1; \
   12342   uint32x2_t __s2 = __p2; \
   12343   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   12344   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   12345   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   12346   uint32x4_t __ret; \
   12347   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   12348   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   12349   __ret; \
   12350 })
   12351 #endif
   12352 
   12353 #ifdef __LITTLE_ENDIAN__
   12354 #define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   12355   uint16x8_t __s0 = __p0; \
   12356   uint16x8_t __s1 = __p1; \
   12357   uint16x4_t __s2 = __p2; \
   12358   uint16x8_t __ret; \
   12359   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   12360   __ret; \
   12361 })
   12362 #else
   12363 #define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   12364   uint16x8_t __s0 = __p0; \
   12365   uint16x8_t __s1 = __p1; \
   12366   uint16x4_t __s2 = __p2; \
   12367   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   12368   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   12369   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   12370   uint16x8_t __ret; \
   12371   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   12372   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   12373   __ret; \
   12374 })
   12375 #endif
   12376 
   12377 #ifdef __LITTLE_ENDIAN__
   12378 #define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12379   float32x4_t __s0 = __p0; \
   12380   float32x4_t __s1 = __p1; \
   12381   float32x2_t __s2 = __p2; \
   12382   float32x4_t __ret; \
   12383   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   12384   __ret; \
   12385 })
   12386 #else
   12387 #define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12388   float32x4_t __s0 = __p0; \
   12389   float32x4_t __s1 = __p1; \
   12390   float32x2_t __s2 = __p2; \
   12391   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   12392   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   12393   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   12394   float32x4_t __ret; \
   12395   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   12396   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   12397   __ret; \
   12398 })
   12399 #endif
   12400 
   12401 #ifdef __LITTLE_ENDIAN__
   12402 #define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12403   int32x4_t __s0 = __p0; \
   12404   int32x4_t __s1 = __p1; \
   12405   int32x2_t __s2 = __p2; \
   12406   int32x4_t __ret; \
   12407   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   12408   __ret; \
   12409 })
   12410 #else
   12411 #define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12412   int32x4_t __s0 = __p0; \
   12413   int32x4_t __s1 = __p1; \
   12414   int32x2_t __s2 = __p2; \
   12415   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   12416   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   12417   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   12418   int32x4_t __ret; \
   12419   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   12420   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   12421   __ret; \
   12422 })
   12423 #endif
   12424 
   12425 #ifdef __LITTLE_ENDIAN__
   12426 #define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   12427   int16x8_t __s0 = __p0; \
   12428   int16x8_t __s1 = __p1; \
   12429   int16x4_t __s2 = __p2; \
   12430   int16x8_t __ret; \
   12431   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   12432   __ret; \
   12433 })
   12434 #else
   12435 #define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   12436   int16x8_t __s0 = __p0; \
   12437   int16x8_t __s1 = __p1; \
   12438   int16x4_t __s2 = __p2; \
   12439   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   12440   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   12441   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   12442   int16x8_t __ret; \
   12443   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   12444   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   12445   __ret; \
   12446 })
   12447 #endif
   12448 
   12449 #ifdef __LITTLE_ENDIAN__
   12450 #define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12451   uint32x2_t __s0 = __p0; \
   12452   uint32x2_t __s1 = __p1; \
   12453   uint32x2_t __s2 = __p2; \
   12454   uint32x2_t __ret; \
   12455   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   12456   __ret; \
   12457 })
   12458 #else
   12459 #define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12460   uint32x2_t __s0 = __p0; \
   12461   uint32x2_t __s1 = __p1; \
   12462   uint32x2_t __s2 = __p2; \
   12463   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   12464   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   12465   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   12466   uint32x2_t __ret; \
   12467   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   12468   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   12469   __ret; \
   12470 })
   12471 #endif
   12472 
   12473 #ifdef __LITTLE_ENDIAN__
   12474 #define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   12475   uint16x4_t __s0 = __p0; \
   12476   uint16x4_t __s1 = __p1; \
   12477   uint16x4_t __s2 = __p2; \
   12478   uint16x4_t __ret; \
   12479   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   12480   __ret; \
   12481 })
   12482 #else
   12483 #define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   12484   uint16x4_t __s0 = __p0; \
   12485   uint16x4_t __s1 = __p1; \
   12486   uint16x4_t __s2 = __p2; \
   12487   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   12488   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   12489   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   12490   uint16x4_t __ret; \
   12491   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   12492   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   12493   __ret; \
   12494 })
   12495 #endif
   12496 
   12497 #ifdef __LITTLE_ENDIAN__
   12498 #define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12499   float32x2_t __s0 = __p0; \
   12500   float32x2_t __s1 = __p1; \
   12501   float32x2_t __s2 = __p2; \
   12502   float32x2_t __ret; \
   12503   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   12504   __ret; \
   12505 })
   12506 #else
   12507 #define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12508   float32x2_t __s0 = __p0; \
   12509   float32x2_t __s1 = __p1; \
   12510   float32x2_t __s2 = __p2; \
   12511   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   12512   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   12513   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   12514   float32x2_t __ret; \
   12515   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   12516   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   12517   __ret; \
   12518 })
   12519 #endif
   12520 
   12521 #ifdef __LITTLE_ENDIAN__
   12522 #define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12523   int32x2_t __s0 = __p0; \
   12524   int32x2_t __s1 = __p1; \
   12525   int32x2_t __s2 = __p2; \
   12526   int32x2_t __ret; \
   12527   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   12528   __ret; \
   12529 })
   12530 #else
   12531 #define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12532   int32x2_t __s0 = __p0; \
   12533   int32x2_t __s1 = __p1; \
   12534   int32x2_t __s2 = __p2; \
   12535   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   12536   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   12537   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   12538   int32x2_t __ret; \
   12539   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   12540   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   12541   __ret; \
   12542 })
   12543 #endif
   12544 
   12545 #ifdef __LITTLE_ENDIAN__
   12546 #define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   12547   int16x4_t __s0 = __p0; \
   12548   int16x4_t __s1 = __p1; \
   12549   int16x4_t __s2 = __p2; \
   12550   int16x4_t __ret; \
   12551   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   12552   __ret; \
   12553 })
   12554 #else
   12555 #define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   12556   int16x4_t __s0 = __p0; \
   12557   int16x4_t __s1 = __p1; \
   12558   int16x4_t __s2 = __p2; \
   12559   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   12560   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   12561   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   12562   int16x4_t __ret; \
   12563   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   12564   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   12565   __ret; \
   12566 })
   12567 #endif
   12568 
   12569 #ifdef __LITTLE_ENDIAN__
   12570 __ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
   12571   uint32x4_t __ret;
   12572   __ret = __p0 + __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
   12573   return __ret;
   12574 }
   12575 #else
   12576 __ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
   12577   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12578   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12579   uint32x4_t __ret;
   12580   __ret = __rev0 + __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
   12581   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12582   return __ret;
   12583 }
   12584 #endif
   12585 
   12586 #ifdef __LITTLE_ENDIAN__
   12587 __ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
   12588   uint16x8_t __ret;
   12589   __ret = __p0 + __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
   12590   return __ret;
   12591 }
   12592 #else
   12593 __ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
   12594   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12595   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12596   uint16x8_t __ret;
   12597   __ret = __rev0 + __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
   12598   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12599   return __ret;
   12600 }
   12601 #endif
   12602 
   12603 #ifdef __LITTLE_ENDIAN__
   12604 __ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
   12605   float32x4_t __ret;
   12606   __ret = __p0 + __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
   12607   return __ret;
   12608 }
   12609 #else
   12610 __ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
   12611   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12612   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12613   float32x4_t __ret;
   12614   __ret = __rev0 + __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
   12615   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12616   return __ret;
   12617 }
   12618 #endif
   12619 
   12620 #ifdef __LITTLE_ENDIAN__
   12621 __ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
   12622   int32x4_t __ret;
   12623   __ret = __p0 + __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
   12624   return __ret;
   12625 }
   12626 #else
   12627 __ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
   12628   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12629   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12630   int32x4_t __ret;
   12631   __ret = __rev0 + __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
   12632   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12633   return __ret;
   12634 }
   12635 #endif
   12636 
   12637 #ifdef __LITTLE_ENDIAN__
   12638 __ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
   12639   int16x8_t __ret;
   12640   __ret = __p0 + __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
   12641   return __ret;
   12642 }
   12643 #else
   12644 __ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
   12645   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12646   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12647   int16x8_t __ret;
   12648   __ret = __rev0 + __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
   12649   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12650   return __ret;
   12651 }
   12652 #endif
   12653 
   12654 #ifdef __LITTLE_ENDIAN__
   12655 __ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   12656   uint32x2_t __ret;
   12657   __ret = __p0 + __p1 * (uint32x2_t) {__p2, __p2};
   12658   return __ret;
   12659 }
   12660 #else
   12661 __ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   12662   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12663   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12664   uint32x2_t __ret;
   12665   __ret = __rev0 + __rev1 * (uint32x2_t) {__p2, __p2};
   12666   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12667   return __ret;
   12668 }
   12669 #endif
   12670 
   12671 #ifdef __LITTLE_ENDIAN__
   12672 __ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   12673   uint16x4_t __ret;
   12674   __ret = __p0 + __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
   12675   return __ret;
   12676 }
   12677 #else
   12678 __ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   12679   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12680   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12681   uint16x4_t __ret;
   12682   __ret = __rev0 + __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
   12683   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12684   return __ret;
   12685 }
   12686 #endif
   12687 
   12688 #ifdef __LITTLE_ENDIAN__
   12689 __ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
   12690   float32x2_t __ret;
   12691   __ret = __p0 + __p1 * (float32x2_t) {__p2, __p2};
   12692   return __ret;
   12693 }
   12694 #else
   12695 __ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
   12696   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12697   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12698   float32x2_t __ret;
   12699   __ret = __rev0 + __rev1 * (float32x2_t) {__p2, __p2};
   12700   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12701   return __ret;
   12702 }
   12703 #endif
   12704 
   12705 #ifdef __LITTLE_ENDIAN__
   12706 __ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
   12707   int32x2_t __ret;
   12708   __ret = __p0 + __p1 * (int32x2_t) {__p2, __p2};
   12709   return __ret;
   12710 }
   12711 #else
   12712 __ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
   12713   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12714   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12715   int32x2_t __ret;
   12716   __ret = __rev0 + __rev1 * (int32x2_t) {__p2, __p2};
   12717   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12718   return __ret;
   12719 }
   12720 #endif
   12721 
   12722 #ifdef __LITTLE_ENDIAN__
   12723 __ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
   12724   int16x4_t __ret;
   12725   __ret = __p0 + __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
   12726   return __ret;
   12727 }
   12728 #else
   12729 __ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
   12730   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12731   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12732   int16x4_t __ret;
   12733   __ret = __rev0 + __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
   12734   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12735   return __ret;
   12736 }
   12737 #endif
   12738 
   12739 #ifdef __LITTLE_ENDIAN__
   12740 __ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   12741   uint8x16_t __ret;
   12742   __ret = __p0 - __p1 * __p2;
   12743   return __ret;
   12744 }
   12745 #else
   12746 __ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   12747   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12748   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12749   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12750   uint8x16_t __ret;
   12751   __ret = __rev0 - __rev1 * __rev2;
   12752   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12753   return __ret;
   12754 }
   12755 #endif
   12756 
   12757 #ifdef __LITTLE_ENDIAN__
   12758 __ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   12759   uint32x4_t __ret;
   12760   __ret = __p0 - __p1 * __p2;
   12761   return __ret;
   12762 }
   12763 #else
   12764 __ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   12765   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12766   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12767   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12768   uint32x4_t __ret;
   12769   __ret = __rev0 - __rev1 * __rev2;
   12770   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12771   return __ret;
   12772 }
   12773 #endif
   12774 
   12775 #ifdef __LITTLE_ENDIAN__
   12776 __ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   12777   uint16x8_t __ret;
   12778   __ret = __p0 - __p1 * __p2;
   12779   return __ret;
   12780 }
   12781 #else
   12782 __ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   12783   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12784   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12785   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   12786   uint16x8_t __ret;
   12787   __ret = __rev0 - __rev1 * __rev2;
   12788   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12789   return __ret;
   12790 }
   12791 #endif
   12792 
   12793 #ifdef __LITTLE_ENDIAN__
   12794 __ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   12795   int8x16_t __ret;
   12796   __ret = __p0 - __p1 * __p2;
   12797   return __ret;
   12798 }
   12799 #else
   12800 __ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   12801   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12802   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12803   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12804   int8x16_t __ret;
   12805   __ret = __rev0 - __rev1 * __rev2;
   12806   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   12807   return __ret;
   12808 }
   12809 #endif
   12810 
   12811 #ifdef __LITTLE_ENDIAN__
   12812 __ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   12813   float32x4_t __ret;
   12814   __ret = __p0 - __p1 * __p2;
   12815   return __ret;
   12816 }
   12817 #else
   12818 __ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   12819   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12820   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12821   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12822   float32x4_t __ret;
   12823   __ret = __rev0 - __rev1 * __rev2;
   12824   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12825   return __ret;
   12826 }
   12827 #endif
   12828 
   12829 #ifdef __LITTLE_ENDIAN__
   12830 __ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   12831   int32x4_t __ret;
   12832   __ret = __p0 - __p1 * __p2;
   12833   return __ret;
   12834 }
   12835 #else
   12836 __ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   12837   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12838   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12839   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12840   int32x4_t __ret;
   12841   __ret = __rev0 - __rev1 * __rev2;
   12842   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12843   return __ret;
   12844 }
   12845 #endif
   12846 
   12847 #ifdef __LITTLE_ENDIAN__
   12848 __ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   12849   int16x8_t __ret;
   12850   __ret = __p0 - __p1 * __p2;
   12851   return __ret;
   12852 }
   12853 #else
   12854 __ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   12855   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12856   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12857   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   12858   int16x8_t __ret;
   12859   __ret = __rev0 - __rev1 * __rev2;
   12860   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12861   return __ret;
   12862 }
   12863 #endif
   12864 
   12865 #ifdef __LITTLE_ENDIAN__
   12866 __ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   12867   uint8x8_t __ret;
   12868   __ret = __p0 - __p1 * __p2;
   12869   return __ret;
   12870 }
   12871 #else
   12872 __ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   12873   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12874   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12875   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   12876   uint8x8_t __ret;
   12877   __ret = __rev0 - __rev1 * __rev2;
   12878   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12879   return __ret;
   12880 }
   12881 #endif
   12882 
   12883 #ifdef __LITTLE_ENDIAN__
   12884 __ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   12885   uint32x2_t __ret;
   12886   __ret = __p0 - __p1 * __p2;
   12887   return __ret;
   12888 }
   12889 #else
   12890 __ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   12891   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12892   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12893   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   12894   uint32x2_t __ret;
   12895   __ret = __rev0 - __rev1 * __rev2;
   12896   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12897   return __ret;
   12898 }
   12899 #endif
   12900 
   12901 #ifdef __LITTLE_ENDIAN__
   12902 __ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   12903   uint16x4_t __ret;
   12904   __ret = __p0 - __p1 * __p2;
   12905   return __ret;
   12906 }
   12907 #else
   12908 __ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   12909   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12910   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12911   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12912   uint16x4_t __ret;
   12913   __ret = __rev0 - __rev1 * __rev2;
   12914   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12915   return __ret;
   12916 }
   12917 #endif
   12918 
   12919 #ifdef __LITTLE_ENDIAN__
   12920 __ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   12921   int8x8_t __ret;
   12922   __ret = __p0 - __p1 * __p2;
   12923   return __ret;
   12924 }
   12925 #else
   12926 __ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   12927   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   12928   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   12929   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   12930   int8x8_t __ret;
   12931   __ret = __rev0 - __rev1 * __rev2;
   12932   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   12933   return __ret;
   12934 }
   12935 #endif
   12936 
   12937 #ifdef __LITTLE_ENDIAN__
   12938 __ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   12939   float32x2_t __ret;
   12940   __ret = __p0 - __p1 * __p2;
   12941   return __ret;
   12942 }
   12943 #else
   12944 __ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   12945   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12946   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12947   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   12948   float32x2_t __ret;
   12949   __ret = __rev0 - __rev1 * __rev2;
   12950   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12951   return __ret;
   12952 }
   12953 #endif
   12954 
   12955 #ifdef __LITTLE_ENDIAN__
   12956 __ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   12957   int32x2_t __ret;
   12958   __ret = __p0 - __p1 * __p2;
   12959   return __ret;
   12960 }
   12961 #else
   12962 __ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   12963   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   12964   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   12965   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   12966   int32x2_t __ret;
   12967   __ret = __rev0 - __rev1 * __rev2;
   12968   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   12969   return __ret;
   12970 }
   12971 #endif
   12972 
   12973 #ifdef __LITTLE_ENDIAN__
   12974 __ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   12975   int16x4_t __ret;
   12976   __ret = __p0 - __p1 * __p2;
   12977   return __ret;
   12978 }
   12979 #else
   12980 __ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   12981   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   12982   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   12983   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   12984   int16x4_t __ret;
   12985   __ret = __rev0 - __rev1 * __rev2;
   12986   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   12987   return __ret;
   12988 }
   12989 #endif
   12990 
   12991 #ifdef __LITTLE_ENDIAN__
   12992 #define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   12993   uint32x4_t __s0 = __p0; \
   12994   uint32x4_t __s1 = __p1; \
   12995   uint32x2_t __s2 = __p2; \
   12996   uint32x4_t __ret; \
   12997   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   12998   __ret; \
   12999 })
   13000 #else
   13001 #define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13002   uint32x4_t __s0 = __p0; \
   13003   uint32x4_t __s1 = __p1; \
   13004   uint32x2_t __s2 = __p2; \
   13005   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   13006   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   13007   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   13008   uint32x4_t __ret; \
   13009   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   13010   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   13011   __ret; \
   13012 })
   13013 #endif
   13014 
   13015 #ifdef __LITTLE_ENDIAN__
   13016 #define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   13017   uint16x8_t __s0 = __p0; \
   13018   uint16x8_t __s1 = __p1; \
   13019   uint16x4_t __s2 = __p2; \
   13020   uint16x8_t __ret; \
   13021   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   13022   __ret; \
   13023 })
   13024 #else
   13025 #define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   13026   uint16x8_t __s0 = __p0; \
   13027   uint16x8_t __s1 = __p1; \
   13028   uint16x4_t __s2 = __p2; \
   13029   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   13030   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   13031   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   13032   uint16x8_t __ret; \
   13033   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   13034   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   13035   __ret; \
   13036 })
   13037 #endif
   13038 
   13039 #ifdef __LITTLE_ENDIAN__
   13040 #define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13041   float32x4_t __s0 = __p0; \
   13042   float32x4_t __s1 = __p1; \
   13043   float32x2_t __s2 = __p2; \
   13044   float32x4_t __ret; \
   13045   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   13046   __ret; \
   13047 })
   13048 #else
   13049 #define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13050   float32x4_t __s0 = __p0; \
   13051   float32x4_t __s1 = __p1; \
   13052   float32x2_t __s2 = __p2; \
   13053   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   13054   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   13055   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   13056   float32x4_t __ret; \
   13057   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   13058   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   13059   __ret; \
   13060 })
   13061 #endif
   13062 
   13063 #ifdef __LITTLE_ENDIAN__
   13064 #define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13065   int32x4_t __s0 = __p0; \
   13066   int32x4_t __s1 = __p1; \
   13067   int32x2_t __s2 = __p2; \
   13068   int32x4_t __ret; \
   13069   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   13070   __ret; \
   13071 })
   13072 #else
   13073 #define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13074   int32x4_t __s0 = __p0; \
   13075   int32x4_t __s1 = __p1; \
   13076   int32x2_t __s2 = __p2; \
   13077   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   13078   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   13079   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   13080   int32x4_t __ret; \
   13081   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   13082   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   13083   __ret; \
   13084 })
   13085 #endif
   13086 
   13087 #ifdef __LITTLE_ENDIAN__
   13088 #define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   13089   int16x8_t __s0 = __p0; \
   13090   int16x8_t __s1 = __p1; \
   13091   int16x4_t __s2 = __p2; \
   13092   int16x8_t __ret; \
   13093   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   13094   __ret; \
   13095 })
   13096 #else
   13097 #define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   13098   int16x8_t __s0 = __p0; \
   13099   int16x8_t __s1 = __p1; \
   13100   int16x4_t __s2 = __p2; \
   13101   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   13102   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   13103   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   13104   int16x8_t __ret; \
   13105   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   13106   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   13107   __ret; \
   13108 })
   13109 #endif
   13110 
   13111 #ifdef __LITTLE_ENDIAN__
   13112 #define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13113   uint32x2_t __s0 = __p0; \
   13114   uint32x2_t __s1 = __p1; \
   13115   uint32x2_t __s2 = __p2; \
   13116   uint32x2_t __ret; \
   13117   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   13118   __ret; \
   13119 })
   13120 #else
   13121 #define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13122   uint32x2_t __s0 = __p0; \
   13123   uint32x2_t __s1 = __p1; \
   13124   uint32x2_t __s2 = __p2; \
   13125   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   13126   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   13127   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   13128   uint32x2_t __ret; \
   13129   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   13130   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   13131   __ret; \
   13132 })
   13133 #endif
   13134 
   13135 #ifdef __LITTLE_ENDIAN__
   13136 #define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   13137   uint16x4_t __s0 = __p0; \
   13138   uint16x4_t __s1 = __p1; \
   13139   uint16x4_t __s2 = __p2; \
   13140   uint16x4_t __ret; \
   13141   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   13142   __ret; \
   13143 })
   13144 #else
   13145 #define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   13146   uint16x4_t __s0 = __p0; \
   13147   uint16x4_t __s1 = __p1; \
   13148   uint16x4_t __s2 = __p2; \
   13149   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   13150   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   13151   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   13152   uint16x4_t __ret; \
   13153   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   13154   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   13155   __ret; \
   13156 })
   13157 #endif
   13158 
   13159 #ifdef __LITTLE_ENDIAN__
   13160 #define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13161   float32x2_t __s0 = __p0; \
   13162   float32x2_t __s1 = __p1; \
   13163   float32x2_t __s2 = __p2; \
   13164   float32x2_t __ret; \
   13165   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   13166   __ret; \
   13167 })
   13168 #else
   13169 #define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13170   float32x2_t __s0 = __p0; \
   13171   float32x2_t __s1 = __p1; \
   13172   float32x2_t __s2 = __p2; \
   13173   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   13174   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   13175   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   13176   float32x2_t __ret; \
   13177   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   13178   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   13179   __ret; \
   13180 })
   13181 #endif
   13182 
   13183 #ifdef __LITTLE_ENDIAN__
   13184 #define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13185   int32x2_t __s0 = __p0; \
   13186   int32x2_t __s1 = __p1; \
   13187   int32x2_t __s2 = __p2; \
   13188   int32x2_t __ret; \
   13189   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   13190   __ret; \
   13191 })
   13192 #else
   13193 #define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   13194   int32x2_t __s0 = __p0; \
   13195   int32x2_t __s1 = __p1; \
   13196   int32x2_t __s2 = __p2; \
   13197   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   13198   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   13199   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   13200   int32x2_t __ret; \
   13201   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   13202   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   13203   __ret; \
   13204 })
   13205 #endif
   13206 
   13207 #ifdef __LITTLE_ENDIAN__
   13208 #define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   13209   int16x4_t __s0 = __p0; \
   13210   int16x4_t __s1 = __p1; \
   13211   int16x4_t __s2 = __p2; \
   13212   int16x4_t __ret; \
   13213   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   13214   __ret; \
   13215 })
   13216 #else
   13217 #define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   13218   int16x4_t __s0 = __p0; \
   13219   int16x4_t __s1 = __p1; \
   13220   int16x4_t __s2 = __p2; \
   13221   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   13222   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   13223   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   13224   int16x4_t __ret; \
   13225   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   13226   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   13227   __ret; \
   13228 })
   13229 #endif
   13230 
   13231 #ifdef __LITTLE_ENDIAN__
   13232 __ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
   13233   uint32x4_t __ret;
   13234   __ret = __p0 - __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
   13235   return __ret;
   13236 }
   13237 #else
   13238 __ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
   13239   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   13240   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   13241   uint32x4_t __ret;
   13242   __ret = __rev0 - __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
   13243   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13244   return __ret;
   13245 }
   13246 #endif
   13247 
   13248 #ifdef __LITTLE_ENDIAN__
   13249 __ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
   13250   uint16x8_t __ret;
   13251   __ret = __p0 - __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
   13252   return __ret;
   13253 }
   13254 #else
   13255 __ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
   13256   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   13257   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   13258   uint16x8_t __ret;
   13259   __ret = __rev0 - __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
   13260   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13261   return __ret;
   13262 }
   13263 #endif
   13264 
   13265 #ifdef __LITTLE_ENDIAN__
   13266 __ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
   13267   float32x4_t __ret;
   13268   __ret = __p0 - __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
   13269   return __ret;
   13270 }
   13271 #else
   13272 __ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
   13273   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   13274   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   13275   float32x4_t __ret;
   13276   __ret = __rev0 - __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
   13277   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13278   return __ret;
   13279 }
   13280 #endif
   13281 
   13282 #ifdef __LITTLE_ENDIAN__
   13283 __ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
   13284   int32x4_t __ret;
   13285   __ret = __p0 - __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
   13286   return __ret;
   13287 }
   13288 #else
   13289 __ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
   13290   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   13291   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   13292   int32x4_t __ret;
   13293   __ret = __rev0 - __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
   13294   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13295   return __ret;
   13296 }
   13297 #endif
   13298 
   13299 #ifdef __LITTLE_ENDIAN__
   13300 __ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
   13301   int16x8_t __ret;
   13302   __ret = __p0 - __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
   13303   return __ret;
   13304 }
   13305 #else
   13306 __ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
   13307   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   13308   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   13309   int16x8_t __ret;
   13310   __ret = __rev0 - __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
   13311   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13312   return __ret;
   13313 }
   13314 #endif
   13315 
   13316 #ifdef __LITTLE_ENDIAN__
   13317 __ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   13318   uint32x2_t __ret;
   13319   __ret = __p0 - __p1 * (uint32x2_t) {__p2, __p2};
   13320   return __ret;
   13321 }
   13322 #else
   13323 __ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   13324   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   13325   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   13326   uint32x2_t __ret;
   13327   __ret = __rev0 - __rev1 * (uint32x2_t) {__p2, __p2};
   13328   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13329   return __ret;
   13330 }
   13331 #endif
   13332 
   13333 #ifdef __LITTLE_ENDIAN__
   13334 __ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   13335   uint16x4_t __ret;
   13336   __ret = __p0 - __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
   13337   return __ret;
   13338 }
   13339 #else
   13340 __ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   13341   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   13342   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   13343   uint16x4_t __ret;
   13344   __ret = __rev0 - __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
   13345   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13346   return __ret;
   13347 }
   13348 #endif
   13349 
   13350 #ifdef __LITTLE_ENDIAN__
   13351 __ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
   13352   float32x2_t __ret;
   13353   __ret = __p0 - __p1 * (float32x2_t) {__p2, __p2};
   13354   return __ret;
   13355 }
   13356 #else
   13357 __ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
   13358   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   13359   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   13360   float32x2_t __ret;
   13361   __ret = __rev0 - __rev1 * (float32x2_t) {__p2, __p2};
   13362   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13363   return __ret;
   13364 }
   13365 #endif
   13366 
   13367 #ifdef __LITTLE_ENDIAN__
   13368 __ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
   13369   int32x2_t __ret;
   13370   __ret = __p0 - __p1 * (int32x2_t) {__p2, __p2};
   13371   return __ret;
   13372 }
   13373 #else
   13374 __ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
   13375   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   13376   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   13377   int32x2_t __ret;
   13378   __ret = __rev0 - __rev1 * (int32x2_t) {__p2, __p2};
   13379   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13380   return __ret;
   13381 }
   13382 #endif
   13383 
   13384 #ifdef __LITTLE_ENDIAN__
   13385 __ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
   13386   int16x4_t __ret;
   13387   __ret = __p0 - __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
   13388   return __ret;
   13389 }
   13390 #else
   13391 __ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
   13392   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   13393   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   13394   int16x4_t __ret;
   13395   __ret = __rev0 - __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
   13396   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13397   return __ret;
   13398 }
   13399 #endif
   13400 
   13401 #ifdef __LITTLE_ENDIAN__
   13402 __ai poly8x8_t vmov_n_p8(poly8_t __p0) {
   13403   poly8x8_t __ret;
   13404   __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13405   return __ret;
   13406 }
   13407 #else
   13408 __ai poly8x8_t vmov_n_p8(poly8_t __p0) {
   13409   poly8x8_t __ret;
   13410   __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13411   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13412   return __ret;
   13413 }
   13414 #endif
   13415 
   13416 #ifdef __LITTLE_ENDIAN__
   13417 __ai poly16x4_t vmov_n_p16(poly16_t __p0) {
   13418   poly16x4_t __ret;
   13419   __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
   13420   return __ret;
   13421 }
   13422 #else
   13423 __ai poly16x4_t vmov_n_p16(poly16_t __p0) {
   13424   poly16x4_t __ret;
   13425   __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
   13426   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13427   return __ret;
   13428 }
   13429 #endif
   13430 
   13431 #ifdef __LITTLE_ENDIAN__
   13432 __ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
   13433   poly8x16_t __ret;
   13434   __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13435   return __ret;
   13436 }
   13437 #else
   13438 __ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
   13439   poly8x16_t __ret;
   13440   __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13441   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   13442   return __ret;
   13443 }
   13444 #endif
   13445 
   13446 #ifdef __LITTLE_ENDIAN__
   13447 __ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
   13448   poly16x8_t __ret;
   13449   __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13450   return __ret;
   13451 }
   13452 #else
   13453 __ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
   13454   poly16x8_t __ret;
   13455   __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13456   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13457   return __ret;
   13458 }
   13459 #endif
   13460 
   13461 #ifdef __LITTLE_ENDIAN__
   13462 __ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
   13463   uint8x16_t __ret;
   13464   __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13465   return __ret;
   13466 }
   13467 #else
   13468 __ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
   13469   uint8x16_t __ret;
   13470   __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13471   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   13472   return __ret;
   13473 }
   13474 #endif
   13475 
   13476 #ifdef __LITTLE_ENDIAN__
   13477 __ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
   13478   uint32x4_t __ret;
   13479   __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
   13480   return __ret;
   13481 }
   13482 #else
   13483 __ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
   13484   uint32x4_t __ret;
   13485   __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
   13486   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13487   return __ret;
   13488 }
   13489 #endif
   13490 
   13491 #ifdef __LITTLE_ENDIAN__
   13492 __ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
   13493   uint64x2_t __ret;
   13494   __ret = (uint64x2_t) {__p0, __p0};
   13495   return __ret;
   13496 }
   13497 #else
   13498 __ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
   13499   uint64x2_t __ret;
   13500   __ret = (uint64x2_t) {__p0, __p0};
   13501   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13502   return __ret;
   13503 }
   13504 #endif
   13505 
   13506 #ifdef __LITTLE_ENDIAN__
   13507 __ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
   13508   uint16x8_t __ret;
   13509   __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13510   return __ret;
   13511 }
   13512 #else
   13513 __ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
   13514   uint16x8_t __ret;
   13515   __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13516   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13517   return __ret;
   13518 }
   13519 #endif
   13520 
   13521 #ifdef __LITTLE_ENDIAN__
   13522 __ai int8x16_t vmovq_n_s8(int8_t __p0) {
   13523   int8x16_t __ret;
   13524   __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13525   return __ret;
   13526 }
   13527 #else
   13528 __ai int8x16_t vmovq_n_s8(int8_t __p0) {
   13529   int8x16_t __ret;
   13530   __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13531   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   13532   return __ret;
   13533 }
   13534 #endif
   13535 
   13536 #ifdef __LITTLE_ENDIAN__
   13537 __ai float32x4_t vmovq_n_f32(float32_t __p0) {
   13538   float32x4_t __ret;
   13539   __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
   13540   return __ret;
   13541 }
   13542 #else
   13543 __ai float32x4_t vmovq_n_f32(float32_t __p0) {
   13544   float32x4_t __ret;
   13545   __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
   13546   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13547   return __ret;
   13548 }
   13549 #endif
   13550 
   13551 #ifdef __LITTLE_ENDIAN__
   13552 #define vmovq_n_f16(__p0) __extension__ ({ \
   13553   float16_t __s0 = __p0; \
   13554   float16x8_t __ret; \
   13555   __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
   13556   __ret; \
   13557 })
   13558 #else
   13559 #define vmovq_n_f16(__p0) __extension__ ({ \
   13560   float16_t __s0 = __p0; \
   13561   float16x8_t __ret; \
   13562   __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
   13563   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   13564   __ret; \
   13565 })
   13566 #endif
   13567 
   13568 #ifdef __LITTLE_ENDIAN__
   13569 __ai int32x4_t vmovq_n_s32(int32_t __p0) {
   13570   int32x4_t __ret;
   13571   __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
   13572   return __ret;
   13573 }
   13574 #else
   13575 __ai int32x4_t vmovq_n_s32(int32_t __p0) {
   13576   int32x4_t __ret;
   13577   __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
   13578   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13579   return __ret;
   13580 }
   13581 #endif
   13582 
   13583 #ifdef __LITTLE_ENDIAN__
   13584 __ai int64x2_t vmovq_n_s64(int64_t __p0) {
   13585   int64x2_t __ret;
   13586   __ret = (int64x2_t) {__p0, __p0};
   13587   return __ret;
   13588 }
   13589 #else
   13590 __ai int64x2_t vmovq_n_s64(int64_t __p0) {
   13591   int64x2_t __ret;
   13592   __ret = (int64x2_t) {__p0, __p0};
   13593   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13594   return __ret;
   13595 }
   13596 #endif
   13597 
   13598 #ifdef __LITTLE_ENDIAN__
   13599 __ai int16x8_t vmovq_n_s16(int16_t __p0) {
   13600   int16x8_t __ret;
   13601   __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13602   return __ret;
   13603 }
   13604 #else
   13605 __ai int16x8_t vmovq_n_s16(int16_t __p0) {
   13606   int16x8_t __ret;
   13607   __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13608   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13609   return __ret;
   13610 }
   13611 #endif
   13612 
   13613 #ifdef __LITTLE_ENDIAN__
   13614 __ai uint8x8_t vmov_n_u8(uint8_t __p0) {
   13615   uint8x8_t __ret;
   13616   __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13617   return __ret;
   13618 }
   13619 #else
   13620 __ai uint8x8_t vmov_n_u8(uint8_t __p0) {
   13621   uint8x8_t __ret;
   13622   __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13623   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13624   return __ret;
   13625 }
   13626 #endif
   13627 
   13628 #ifdef __LITTLE_ENDIAN__
   13629 __ai uint32x2_t vmov_n_u32(uint32_t __p0) {
   13630   uint32x2_t __ret;
   13631   __ret = (uint32x2_t) {__p0, __p0};
   13632   return __ret;
   13633 }
   13634 #else
   13635 __ai uint32x2_t vmov_n_u32(uint32_t __p0) {
   13636   uint32x2_t __ret;
   13637   __ret = (uint32x2_t) {__p0, __p0};
   13638   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13639   return __ret;
   13640 }
   13641 #endif
   13642 
   13643 #ifdef __LITTLE_ENDIAN__
   13644 __ai uint64x1_t vmov_n_u64(uint64_t __p0) {
   13645   uint64x1_t __ret;
   13646   __ret = (uint64x1_t) {__p0};
   13647   return __ret;
   13648 }
   13649 #else
   13650 __ai uint64x1_t vmov_n_u64(uint64_t __p0) {
   13651   uint64x1_t __ret;
   13652   __ret = (uint64x1_t) {__p0};
   13653   return __ret;
   13654 }
   13655 #endif
   13656 
   13657 #ifdef __LITTLE_ENDIAN__
   13658 __ai uint16x4_t vmov_n_u16(uint16_t __p0) {
   13659   uint16x4_t __ret;
   13660   __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
   13661   return __ret;
   13662 }
   13663 #else
   13664 __ai uint16x4_t vmov_n_u16(uint16_t __p0) {
   13665   uint16x4_t __ret;
   13666   __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
   13667   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13668   return __ret;
   13669 }
   13670 #endif
   13671 
   13672 #ifdef __LITTLE_ENDIAN__
   13673 __ai int8x8_t vmov_n_s8(int8_t __p0) {
   13674   int8x8_t __ret;
   13675   __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13676   return __ret;
   13677 }
   13678 #else
   13679 __ai int8x8_t vmov_n_s8(int8_t __p0) {
   13680   int8x8_t __ret;
   13681   __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
   13682   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13683   return __ret;
   13684 }
   13685 #endif
   13686 
   13687 #ifdef __LITTLE_ENDIAN__
   13688 __ai float32x2_t vmov_n_f32(float32_t __p0) {
   13689   float32x2_t __ret;
   13690   __ret = (float32x2_t) {__p0, __p0};
   13691   return __ret;
   13692 }
   13693 #else
   13694 __ai float32x2_t vmov_n_f32(float32_t __p0) {
   13695   float32x2_t __ret;
   13696   __ret = (float32x2_t) {__p0, __p0};
   13697   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13698   return __ret;
   13699 }
   13700 #endif
   13701 
   13702 #ifdef __LITTLE_ENDIAN__
   13703 #define vmov_n_f16(__p0) __extension__ ({ \
   13704   float16_t __s0 = __p0; \
   13705   float16x4_t __ret; \
   13706   __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
   13707   __ret; \
   13708 })
   13709 #else
   13710 #define vmov_n_f16(__p0) __extension__ ({ \
   13711   float16_t __s0 = __p0; \
   13712   float16x4_t __ret; \
   13713   __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
   13714   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   13715   __ret; \
   13716 })
   13717 #endif
   13718 
   13719 #ifdef __LITTLE_ENDIAN__
   13720 __ai int32x2_t vmov_n_s32(int32_t __p0) {
   13721   int32x2_t __ret;
   13722   __ret = (int32x2_t) {__p0, __p0};
   13723   return __ret;
   13724 }
   13725 #else
   13726 __ai int32x2_t vmov_n_s32(int32_t __p0) {
   13727   int32x2_t __ret;
   13728   __ret = (int32x2_t) {__p0, __p0};
   13729   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13730   return __ret;
   13731 }
   13732 #endif
   13733 
   13734 #ifdef __LITTLE_ENDIAN__
   13735 __ai int64x1_t vmov_n_s64(int64_t __p0) {
   13736   int64x1_t __ret;
   13737   __ret = (int64x1_t) {__p0};
   13738   return __ret;
   13739 }
   13740 #else
   13741 __ai int64x1_t vmov_n_s64(int64_t __p0) {
   13742   int64x1_t __ret;
   13743   __ret = (int64x1_t) {__p0};
   13744   return __ret;
   13745 }
   13746 #endif
   13747 
   13748 #ifdef __LITTLE_ENDIAN__
   13749 __ai int16x4_t vmov_n_s16(int16_t __p0) {
   13750   int16x4_t __ret;
   13751   __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
   13752   return __ret;
   13753 }
   13754 #else
   13755 __ai int16x4_t vmov_n_s16(int16_t __p0) {
   13756   int16x4_t __ret;
   13757   __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
   13758   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13759   return __ret;
   13760 }
   13761 #endif
   13762 
   13763 #ifdef __LITTLE_ENDIAN__
   13764 __ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
   13765   uint16x8_t __ret;
   13766   __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
   13767   return __ret;
   13768 }
   13769 #else
   13770 __ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
   13771   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   13772   uint16x8_t __ret;
   13773   __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 49);
   13774   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13775   return __ret;
   13776 }
   13777 __ai uint16x8_t __noswap_vmovl_u8(uint8x8_t __p0) {
   13778   uint16x8_t __ret;
   13779   __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
   13780   return __ret;
   13781 }
   13782 #endif
   13783 
   13784 #ifdef __LITTLE_ENDIAN__
   13785 __ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
   13786   uint64x2_t __ret;
   13787   __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
   13788   return __ret;
   13789 }
   13790 #else
   13791 __ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
   13792   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   13793   uint64x2_t __ret;
   13794   __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 51);
   13795   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13796   return __ret;
   13797 }
   13798 __ai uint64x2_t __noswap_vmovl_u32(uint32x2_t __p0) {
   13799   uint64x2_t __ret;
   13800   __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
   13801   return __ret;
   13802 }
   13803 #endif
   13804 
   13805 #ifdef __LITTLE_ENDIAN__
   13806 __ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
   13807   uint32x4_t __ret;
   13808   __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
   13809   return __ret;
   13810 }
   13811 #else
   13812 __ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
   13813   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   13814   uint32x4_t __ret;
   13815   __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 50);
   13816   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13817   return __ret;
   13818 }
   13819 __ai uint32x4_t __noswap_vmovl_u16(uint16x4_t __p0) {
   13820   uint32x4_t __ret;
   13821   __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
   13822   return __ret;
   13823 }
   13824 #endif
   13825 
   13826 #ifdef __LITTLE_ENDIAN__
   13827 __ai int16x8_t vmovl_s8(int8x8_t __p0) {
   13828   int16x8_t __ret;
   13829   __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
   13830   return __ret;
   13831 }
   13832 #else
   13833 __ai int16x8_t vmovl_s8(int8x8_t __p0) {
   13834   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   13835   int16x8_t __ret;
   13836   __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 33);
   13837   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13838   return __ret;
   13839 }
   13840 __ai int16x8_t __noswap_vmovl_s8(int8x8_t __p0) {
   13841   int16x8_t __ret;
   13842   __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
   13843   return __ret;
   13844 }
   13845 #endif
   13846 
   13847 #ifdef __LITTLE_ENDIAN__
   13848 __ai int64x2_t vmovl_s32(int32x2_t __p0) {
   13849   int64x2_t __ret;
   13850   __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
   13851   return __ret;
   13852 }
   13853 #else
   13854 __ai int64x2_t vmovl_s32(int32x2_t __p0) {
   13855   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   13856   int64x2_t __ret;
   13857   __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 35);
   13858   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13859   return __ret;
   13860 }
   13861 __ai int64x2_t __noswap_vmovl_s32(int32x2_t __p0) {
   13862   int64x2_t __ret;
   13863   __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
   13864   return __ret;
   13865 }
   13866 #endif
   13867 
   13868 #ifdef __LITTLE_ENDIAN__
   13869 __ai int32x4_t vmovl_s16(int16x4_t __p0) {
   13870   int32x4_t __ret;
   13871   __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
   13872   return __ret;
   13873 }
   13874 #else
   13875 __ai int32x4_t vmovl_s16(int16x4_t __p0) {
   13876   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   13877   int32x4_t __ret;
   13878   __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 34);
   13879   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13880   return __ret;
   13881 }
   13882 __ai int32x4_t __noswap_vmovl_s16(int16x4_t __p0) {
   13883   int32x4_t __ret;
   13884   __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
   13885   return __ret;
   13886 }
   13887 #endif
   13888 
   13889 #ifdef __LITTLE_ENDIAN__
   13890 __ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
   13891   uint16x4_t __ret;
   13892   __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
   13893   return __ret;
   13894 }
   13895 #else
   13896 __ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
   13897   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   13898   uint16x4_t __ret;
   13899   __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 17);
   13900   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13901   return __ret;
   13902 }
   13903 __ai uint16x4_t __noswap_vmovn_u32(uint32x4_t __p0) {
   13904   uint16x4_t __ret;
   13905   __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
   13906   return __ret;
   13907 }
   13908 #endif
   13909 
   13910 #ifdef __LITTLE_ENDIAN__
   13911 __ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
   13912   uint32x2_t __ret;
   13913   __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
   13914   return __ret;
   13915 }
   13916 #else
   13917 __ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
   13918   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   13919   uint32x2_t __ret;
   13920   __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 18);
   13921   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13922   return __ret;
   13923 }
   13924 __ai uint32x2_t __noswap_vmovn_u64(uint64x2_t __p0) {
   13925   uint32x2_t __ret;
   13926   __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
   13927   return __ret;
   13928 }
   13929 #endif
   13930 
   13931 #ifdef __LITTLE_ENDIAN__
   13932 __ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
   13933   uint8x8_t __ret;
   13934   __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
   13935   return __ret;
   13936 }
   13937 #else
   13938 __ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
   13939   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   13940   uint8x8_t __ret;
   13941   __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 16);
   13942   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   13943   return __ret;
   13944 }
   13945 __ai uint8x8_t __noswap_vmovn_u16(uint16x8_t __p0) {
   13946   uint8x8_t __ret;
   13947   __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
   13948   return __ret;
   13949 }
   13950 #endif
   13951 
   13952 #ifdef __LITTLE_ENDIAN__
   13953 __ai int16x4_t vmovn_s32(int32x4_t __p0) {
   13954   int16x4_t __ret;
   13955   __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
   13956   return __ret;
   13957 }
   13958 #else
   13959 __ai int16x4_t vmovn_s32(int32x4_t __p0) {
   13960   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   13961   int16x4_t __ret;
   13962   __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 1);
   13963   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   13964   return __ret;
   13965 }
   13966 __ai int16x4_t __noswap_vmovn_s32(int32x4_t __p0) {
   13967   int16x4_t __ret;
   13968   __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
   13969   return __ret;
   13970 }
   13971 #endif
   13972 
   13973 #ifdef __LITTLE_ENDIAN__
   13974 __ai int32x2_t vmovn_s64(int64x2_t __p0) {
   13975   int32x2_t __ret;
   13976   __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
   13977   return __ret;
   13978 }
   13979 #else
   13980 __ai int32x2_t vmovn_s64(int64x2_t __p0) {
   13981   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   13982   int32x2_t __ret;
   13983   __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 2);
   13984   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   13985   return __ret;
   13986 }
   13987 __ai int32x2_t __noswap_vmovn_s64(int64x2_t __p0) {
   13988   int32x2_t __ret;
   13989   __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
   13990   return __ret;
   13991 }
   13992 #endif
   13993 
   13994 #ifdef __LITTLE_ENDIAN__
   13995 __ai int8x8_t vmovn_s16(int16x8_t __p0) {
   13996   int8x8_t __ret;
   13997   __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
   13998   return __ret;
   13999 }
   14000 #else
   14001 __ai int8x8_t vmovn_s16(int16x8_t __p0) {
   14002   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14003   int8x8_t __ret;
   14004   __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 0);
   14005   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14006   return __ret;
   14007 }
   14008 __ai int8x8_t __noswap_vmovn_s16(int16x8_t __p0) {
   14009   int8x8_t __ret;
   14010   __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
   14011   return __ret;
   14012 }
   14013 #endif
   14014 
   14015 #ifdef __LITTLE_ENDIAN__
   14016 __ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   14017   uint8x16_t __ret;
   14018   __ret = __p0 * __p1;
   14019   return __ret;
   14020 }
   14021 #else
   14022 __ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   14023   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   14024   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   14025   uint8x16_t __ret;
   14026   __ret = __rev0 * __rev1;
   14027   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   14028   return __ret;
   14029 }
   14030 #endif
   14031 
   14032 #ifdef __LITTLE_ENDIAN__
   14033 __ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   14034   uint32x4_t __ret;
   14035   __ret = __p0 * __p1;
   14036   return __ret;
   14037 }
   14038 #else
   14039 __ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   14040   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14041   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   14042   uint32x4_t __ret;
   14043   __ret = __rev0 * __rev1;
   14044   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14045   return __ret;
   14046 }
   14047 #endif
   14048 
   14049 #ifdef __LITTLE_ENDIAN__
   14050 __ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   14051   uint16x8_t __ret;
   14052   __ret = __p0 * __p1;
   14053   return __ret;
   14054 }
   14055 #else
   14056 __ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   14057   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14058   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   14059   uint16x8_t __ret;
   14060   __ret = __rev0 * __rev1;
   14061   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14062   return __ret;
   14063 }
   14064 #endif
   14065 
   14066 #ifdef __LITTLE_ENDIAN__
   14067 __ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
   14068   int8x16_t __ret;
   14069   __ret = __p0 * __p1;
   14070   return __ret;
   14071 }
   14072 #else
   14073 __ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
   14074   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   14075   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   14076   int8x16_t __ret;
   14077   __ret = __rev0 * __rev1;
   14078   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   14079   return __ret;
   14080 }
   14081 #endif
   14082 
   14083 #ifdef __LITTLE_ENDIAN__
   14084 __ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
   14085   float32x4_t __ret;
   14086   __ret = __p0 * __p1;
   14087   return __ret;
   14088 }
   14089 #else
   14090 __ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
   14091   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14092   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   14093   float32x4_t __ret;
   14094   __ret = __rev0 * __rev1;
   14095   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14096   return __ret;
   14097 }
   14098 #endif
   14099 
   14100 #ifdef __LITTLE_ENDIAN__
   14101 __ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
   14102   int32x4_t __ret;
   14103   __ret = __p0 * __p1;
   14104   return __ret;
   14105 }
   14106 #else
   14107 __ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
   14108   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14109   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   14110   int32x4_t __ret;
   14111   __ret = __rev0 * __rev1;
   14112   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14113   return __ret;
   14114 }
   14115 #endif
   14116 
   14117 #ifdef __LITTLE_ENDIAN__
   14118 __ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
   14119   int16x8_t __ret;
   14120   __ret = __p0 * __p1;
   14121   return __ret;
   14122 }
   14123 #else
   14124 __ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
   14125   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14126   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   14127   int16x8_t __ret;
   14128   __ret = __rev0 * __rev1;
   14129   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14130   return __ret;
   14131 }
   14132 #endif
   14133 
   14134 #ifdef __LITTLE_ENDIAN__
   14135 __ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
   14136   uint8x8_t __ret;
   14137   __ret = __p0 * __p1;
   14138   return __ret;
   14139 }
   14140 #else
   14141 __ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
   14142   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14143   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   14144   uint8x8_t __ret;
   14145   __ret = __rev0 * __rev1;
   14146   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14147   return __ret;
   14148 }
   14149 #endif
   14150 
   14151 #ifdef __LITTLE_ENDIAN__
   14152 __ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
   14153   uint32x2_t __ret;
   14154   __ret = __p0 * __p1;
   14155   return __ret;
   14156 }
   14157 #else
   14158 __ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
   14159   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14160   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   14161   uint32x2_t __ret;
   14162   __ret = __rev0 * __rev1;
   14163   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14164   return __ret;
   14165 }
   14166 #endif
   14167 
   14168 #ifdef __LITTLE_ENDIAN__
   14169 __ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
   14170   uint16x4_t __ret;
   14171   __ret = __p0 * __p1;
   14172   return __ret;
   14173 }
   14174 #else
   14175 __ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
   14176   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14177   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   14178   uint16x4_t __ret;
   14179   __ret = __rev0 * __rev1;
   14180   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14181   return __ret;
   14182 }
   14183 #endif
   14184 
   14185 #ifdef __LITTLE_ENDIAN__
   14186 __ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
   14187   int8x8_t __ret;
   14188   __ret = __p0 * __p1;
   14189   return __ret;
   14190 }
   14191 #else
   14192 __ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
   14193   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14194   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   14195   int8x8_t __ret;
   14196   __ret = __rev0 * __rev1;
   14197   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14198   return __ret;
   14199 }
   14200 #endif
   14201 
   14202 #ifdef __LITTLE_ENDIAN__
   14203 __ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
   14204   float32x2_t __ret;
   14205   __ret = __p0 * __p1;
   14206   return __ret;
   14207 }
   14208 #else
   14209 __ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
   14210   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14211   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   14212   float32x2_t __ret;
   14213   __ret = __rev0 * __rev1;
   14214   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14215   return __ret;
   14216 }
   14217 #endif
   14218 
   14219 #ifdef __LITTLE_ENDIAN__
   14220 __ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
   14221   int32x2_t __ret;
   14222   __ret = __p0 * __p1;
   14223   return __ret;
   14224 }
   14225 #else
   14226 __ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
   14227   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14228   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   14229   int32x2_t __ret;
   14230   __ret = __rev0 * __rev1;
   14231   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14232   return __ret;
   14233 }
   14234 #endif
   14235 
   14236 #ifdef __LITTLE_ENDIAN__
   14237 __ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
   14238   int16x4_t __ret;
   14239   __ret = __p0 * __p1;
   14240   return __ret;
   14241 }
   14242 #else
   14243 __ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
   14244   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14245   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   14246   int16x4_t __ret;
   14247   __ret = __rev0 * __rev1;
   14248   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14249   return __ret;
   14250 }
   14251 #endif
   14252 
   14253 #ifdef __LITTLE_ENDIAN__
   14254 __ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
   14255   poly8x8_t __ret;
   14256   __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
   14257   return __ret;
   14258 }
   14259 #else
   14260 __ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
   14261   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14262   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   14263   poly8x8_t __ret;
   14264   __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
   14265   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14266   return __ret;
   14267 }
   14268 #endif
   14269 
   14270 #ifdef __LITTLE_ENDIAN__
   14271 __ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   14272   poly8x16_t __ret;
   14273   __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
   14274   return __ret;
   14275 }
   14276 #else
   14277 __ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   14278   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   14279   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   14280   poly8x16_t __ret;
   14281   __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
   14282   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   14283   return __ret;
   14284 }
   14285 #endif
   14286 
   14287 #ifdef __LITTLE_ENDIAN__
   14288 #define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   14289   uint32x4_t __s0 = __p0; \
   14290   uint32x2_t __s1 = __p1; \
   14291   uint32x4_t __ret; \
   14292   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   14293   __ret; \
   14294 })
   14295 #else
   14296 #define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   14297   uint32x4_t __s0 = __p0; \
   14298   uint32x2_t __s1 = __p1; \
   14299   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   14300   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   14301   uint32x4_t __ret; \
   14302   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   14303   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   14304   __ret; \
   14305 })
   14306 #endif
   14307 
   14308 #ifdef __LITTLE_ENDIAN__
   14309 #define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   14310   uint16x8_t __s0 = __p0; \
   14311   uint16x4_t __s1 = __p1; \
   14312   uint16x8_t __ret; \
   14313   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
   14314   __ret; \
   14315 })
   14316 #else
   14317 #define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   14318   uint16x8_t __s0 = __p0; \
   14319   uint16x4_t __s1 = __p1; \
   14320   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   14321   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   14322   uint16x8_t __ret; \
   14323   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
   14324   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   14325   __ret; \
   14326 })
   14327 #endif
   14328 
   14329 #ifdef __LITTLE_ENDIAN__
   14330 #define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   14331   float32x4_t __s0 = __p0; \
   14332   float32x2_t __s1 = __p1; \
   14333   float32x4_t __ret; \
   14334   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   14335   __ret; \
   14336 })
   14337 #else
   14338 #define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   14339   float32x4_t __s0 = __p0; \
   14340   float32x2_t __s1 = __p1; \
   14341   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   14342   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   14343   float32x4_t __ret; \
   14344   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   14345   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   14346   __ret; \
   14347 })
   14348 #endif
   14349 
   14350 #ifdef __LITTLE_ENDIAN__
   14351 #define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   14352   int32x4_t __s0 = __p0; \
   14353   int32x2_t __s1 = __p1; \
   14354   int32x4_t __ret; \
   14355   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   14356   __ret; \
   14357 })
   14358 #else
   14359 #define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   14360   int32x4_t __s0 = __p0; \
   14361   int32x2_t __s1 = __p1; \
   14362   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   14363   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   14364   int32x4_t __ret; \
   14365   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   14366   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   14367   __ret; \
   14368 })
   14369 #endif
   14370 
   14371 #ifdef __LITTLE_ENDIAN__
   14372 #define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   14373   int16x8_t __s0 = __p0; \
   14374   int16x4_t __s1 = __p1; \
   14375   int16x8_t __ret; \
   14376   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
   14377   __ret; \
   14378 })
   14379 #else
   14380 #define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   14381   int16x8_t __s0 = __p0; \
   14382   int16x4_t __s1 = __p1; \
   14383   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   14384   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   14385   int16x8_t __ret; \
   14386   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
   14387   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   14388   __ret; \
   14389 })
   14390 #endif
   14391 
   14392 #ifdef __LITTLE_ENDIAN__
   14393 #define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   14394   uint32x2_t __s0 = __p0; \
   14395   uint32x2_t __s1 = __p1; \
   14396   uint32x2_t __ret; \
   14397   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
   14398   __ret; \
   14399 })
   14400 #else
   14401 #define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   14402   uint32x2_t __s0 = __p0; \
   14403   uint32x2_t __s1 = __p1; \
   14404   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   14405   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   14406   uint32x2_t __ret; \
   14407   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
   14408   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   14409   __ret; \
   14410 })
   14411 #endif
   14412 
   14413 #ifdef __LITTLE_ENDIAN__
   14414 #define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   14415   uint16x4_t __s0 = __p0; \
   14416   uint16x4_t __s1 = __p1; \
   14417   uint16x4_t __ret; \
   14418   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   14419   __ret; \
   14420 })
   14421 #else
   14422 #define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   14423   uint16x4_t __s0 = __p0; \
   14424   uint16x4_t __s1 = __p1; \
   14425   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   14426   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   14427   uint16x4_t __ret; \
   14428   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   14429   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   14430   __ret; \
   14431 })
   14432 #endif
   14433 
   14434 #ifdef __LITTLE_ENDIAN__
   14435 #define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   14436   float32x2_t __s0 = __p0; \
   14437   float32x2_t __s1 = __p1; \
   14438   float32x2_t __ret; \
   14439   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
   14440   __ret; \
   14441 })
   14442 #else
   14443 #define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   14444   float32x2_t __s0 = __p0; \
   14445   float32x2_t __s1 = __p1; \
   14446   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   14447   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   14448   float32x2_t __ret; \
   14449   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
   14450   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   14451   __ret; \
   14452 })
   14453 #endif
   14454 
   14455 #ifdef __LITTLE_ENDIAN__
   14456 #define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   14457   int32x2_t __s0 = __p0; \
   14458   int32x2_t __s1 = __p1; \
   14459   int32x2_t __ret; \
   14460   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
   14461   __ret; \
   14462 })
   14463 #else
   14464 #define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   14465   int32x2_t __s0 = __p0; \
   14466   int32x2_t __s1 = __p1; \
   14467   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   14468   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   14469   int32x2_t __ret; \
   14470   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
   14471   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   14472   __ret; \
   14473 })
   14474 #endif
   14475 
   14476 #ifdef __LITTLE_ENDIAN__
   14477 #define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   14478   int16x4_t __s0 = __p0; \
   14479   int16x4_t __s1 = __p1; \
   14480   int16x4_t __ret; \
   14481   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   14482   __ret; \
   14483 })
   14484 #else
   14485 #define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   14486   int16x4_t __s0 = __p0; \
   14487   int16x4_t __s1 = __p1; \
   14488   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   14489   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   14490   int16x4_t __ret; \
   14491   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   14492   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   14493   __ret; \
   14494 })
   14495 #endif
   14496 
   14497 #ifdef __LITTLE_ENDIAN__
   14498 __ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
   14499   uint32x4_t __ret;
   14500   __ret = __p0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
   14501   return __ret;
   14502 }
   14503 #else
   14504 __ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
   14505   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14506   uint32x4_t __ret;
   14507   __ret = __rev0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
   14508   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14509   return __ret;
   14510 }
   14511 #endif
   14512 
   14513 #ifdef __LITTLE_ENDIAN__
   14514 __ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
   14515   uint16x8_t __ret;
   14516   __ret = __p0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
   14517   return __ret;
   14518 }
   14519 #else
   14520 __ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
   14521   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14522   uint16x8_t __ret;
   14523   __ret = __rev0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
   14524   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14525   return __ret;
   14526 }
   14527 #endif
   14528 
   14529 #ifdef __LITTLE_ENDIAN__
   14530 __ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
   14531   float32x4_t __ret;
   14532   __ret = __p0 * (float32x4_t) {__p1, __p1, __p1, __p1};
   14533   return __ret;
   14534 }
   14535 #else
   14536 __ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
   14537   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14538   float32x4_t __ret;
   14539   __ret = __rev0 * (float32x4_t) {__p1, __p1, __p1, __p1};
   14540   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14541   return __ret;
   14542 }
   14543 #endif
   14544 
   14545 #ifdef __LITTLE_ENDIAN__
   14546 __ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
   14547   int32x4_t __ret;
   14548   __ret = __p0 * (int32x4_t) {__p1, __p1, __p1, __p1};
   14549   return __ret;
   14550 }
   14551 #else
   14552 __ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
   14553   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14554   int32x4_t __ret;
   14555   __ret = __rev0 * (int32x4_t) {__p1, __p1, __p1, __p1};
   14556   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14557   return __ret;
   14558 }
   14559 #endif
   14560 
   14561 #ifdef __LITTLE_ENDIAN__
   14562 __ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
   14563   int16x8_t __ret;
   14564   __ret = __p0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
   14565   return __ret;
   14566 }
   14567 #else
   14568 __ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
   14569   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14570   int16x8_t __ret;
   14571   __ret = __rev0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
   14572   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14573   return __ret;
   14574 }
   14575 #endif
   14576 
   14577 #ifdef __LITTLE_ENDIAN__
   14578 __ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
   14579   uint32x2_t __ret;
   14580   __ret = __p0 * (uint32x2_t) {__p1, __p1};
   14581   return __ret;
   14582 }
   14583 #else
   14584 __ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
   14585   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14586   uint32x2_t __ret;
   14587   __ret = __rev0 * (uint32x2_t) {__p1, __p1};
   14588   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14589   return __ret;
   14590 }
   14591 #endif
   14592 
   14593 #ifdef __LITTLE_ENDIAN__
   14594 __ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
   14595   uint16x4_t __ret;
   14596   __ret = __p0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
   14597   return __ret;
   14598 }
   14599 #else
   14600 __ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
   14601   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14602   uint16x4_t __ret;
   14603   __ret = __rev0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
   14604   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14605   return __ret;
   14606 }
   14607 #endif
   14608 
   14609 #ifdef __LITTLE_ENDIAN__
   14610 __ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
   14611   float32x2_t __ret;
   14612   __ret = __p0 * (float32x2_t) {__p1, __p1};
   14613   return __ret;
   14614 }
   14615 #else
   14616 __ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
   14617   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14618   float32x2_t __ret;
   14619   __ret = __rev0 * (float32x2_t) {__p1, __p1};
   14620   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14621   return __ret;
   14622 }
   14623 #endif
   14624 
   14625 #ifdef __LITTLE_ENDIAN__
   14626 __ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
   14627   int32x2_t __ret;
   14628   __ret = __p0 * (int32x2_t) {__p1, __p1};
   14629   return __ret;
   14630 }
   14631 #else
   14632 __ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
   14633   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14634   int32x2_t __ret;
   14635   __ret = __rev0 * (int32x2_t) {__p1, __p1};
   14636   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14637   return __ret;
   14638 }
   14639 #endif
   14640 
   14641 #ifdef __LITTLE_ENDIAN__
   14642 __ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
   14643   int16x4_t __ret;
   14644   __ret = __p0 * (int16x4_t) {__p1, __p1, __p1, __p1};
   14645   return __ret;
   14646 }
   14647 #else
   14648 __ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
   14649   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14650   int16x4_t __ret;
   14651   __ret = __rev0 * (int16x4_t) {__p1, __p1, __p1, __p1};
   14652   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14653   return __ret;
   14654 }
   14655 #endif
   14656 
   14657 #ifdef __LITTLE_ENDIAN__
   14658 __ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
   14659   poly16x8_t __ret;
   14660   __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
   14661   return __ret;
   14662 }
   14663 #else
   14664 __ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
   14665   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14666   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   14667   poly16x8_t __ret;
   14668   __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 37);
   14669   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14670   return __ret;
   14671 }
   14672 __ai poly16x8_t __noswap_vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
   14673   poly16x8_t __ret;
   14674   __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
   14675   return __ret;
   14676 }
   14677 #endif
   14678 
   14679 #ifdef __LITTLE_ENDIAN__
   14680 __ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
   14681   uint16x8_t __ret;
   14682   __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
   14683   return __ret;
   14684 }
   14685 #else
   14686 __ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
   14687   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14688   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   14689   uint16x8_t __ret;
   14690   __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 49);
   14691   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14692   return __ret;
   14693 }
   14694 __ai uint16x8_t __noswap_vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
   14695   uint16x8_t __ret;
   14696   __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
   14697   return __ret;
   14698 }
   14699 #endif
   14700 
   14701 #ifdef __LITTLE_ENDIAN__
   14702 __ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
   14703   uint64x2_t __ret;
   14704   __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
   14705   return __ret;
   14706 }
   14707 #else
   14708 __ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
   14709   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14710   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   14711   uint64x2_t __ret;
   14712   __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 51);
   14713   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14714   return __ret;
   14715 }
   14716 __ai uint64x2_t __noswap_vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
   14717   uint64x2_t __ret;
   14718   __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
   14719   return __ret;
   14720 }
   14721 #endif
   14722 
   14723 #ifdef __LITTLE_ENDIAN__
   14724 __ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
   14725   uint32x4_t __ret;
   14726   __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
   14727   return __ret;
   14728 }
   14729 #else
   14730 __ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
   14731   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14732   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   14733   uint32x4_t __ret;
   14734   __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 50);
   14735   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14736   return __ret;
   14737 }
   14738 __ai uint32x4_t __noswap_vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
   14739   uint32x4_t __ret;
   14740   __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
   14741   return __ret;
   14742 }
   14743 #endif
   14744 
   14745 #ifdef __LITTLE_ENDIAN__
   14746 __ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
   14747   int16x8_t __ret;
   14748   __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
   14749   return __ret;
   14750 }
   14751 #else
   14752 __ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
   14753   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14754   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   14755   int16x8_t __ret;
   14756   __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 33);
   14757   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14758   return __ret;
   14759 }
   14760 __ai int16x8_t __noswap_vmull_s8(int8x8_t __p0, int8x8_t __p1) {
   14761   int16x8_t __ret;
   14762   __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
   14763   return __ret;
   14764 }
   14765 #endif
   14766 
   14767 #ifdef __LITTLE_ENDIAN__
   14768 __ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
   14769   int64x2_t __ret;
   14770   __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
   14771   return __ret;
   14772 }
   14773 #else
   14774 __ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
   14775   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14776   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   14777   int64x2_t __ret;
   14778   __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
   14779   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14780   return __ret;
   14781 }
   14782 __ai int64x2_t __noswap_vmull_s32(int32x2_t __p0, int32x2_t __p1) {
   14783   int64x2_t __ret;
   14784   __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
   14785   return __ret;
   14786 }
   14787 #endif
   14788 
   14789 #ifdef __LITTLE_ENDIAN__
   14790 __ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
   14791   int32x4_t __ret;
   14792   __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
   14793   return __ret;
   14794 }
   14795 #else
   14796 __ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
   14797   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14798   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   14799   int32x4_t __ret;
   14800   __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
   14801   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14802   return __ret;
   14803 }
   14804 __ai int32x4_t __noswap_vmull_s16(int16x4_t __p0, int16x4_t __p1) {
   14805   int32x4_t __ret;
   14806   __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
   14807   return __ret;
   14808 }
   14809 #endif
   14810 
   14811 #ifdef __LITTLE_ENDIAN__
   14812 #define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   14813   uint32x2_t __s0 = __p0; \
   14814   uint32x2_t __s1 = __p1; \
   14815   uint64x2_t __ret; \
   14816   __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   14817   __ret; \
   14818 })
   14819 #else
   14820 #define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   14821   uint32x2_t __s0 = __p0; \
   14822   uint32x2_t __s1 = __p1; \
   14823   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   14824   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   14825   uint64x2_t __ret; \
   14826   __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   14827   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   14828   __ret; \
   14829 })
   14830 #endif
   14831 
   14832 #ifdef __LITTLE_ENDIAN__
   14833 #define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   14834   uint16x4_t __s0 = __p0; \
   14835   uint16x4_t __s1 = __p1; \
   14836   uint32x4_t __ret; \
   14837   __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   14838   __ret; \
   14839 })
   14840 #else
   14841 #define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   14842   uint16x4_t __s0 = __p0; \
   14843   uint16x4_t __s1 = __p1; \
   14844   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   14845   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   14846   uint32x4_t __ret; \
   14847   __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   14848   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   14849   __ret; \
   14850 })
   14851 #endif
   14852 
   14853 #ifdef __LITTLE_ENDIAN__
   14854 #define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   14855   int32x2_t __s0 = __p0; \
   14856   int32x2_t __s1 = __p1; \
   14857   int64x2_t __ret; \
   14858   __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   14859   __ret; \
   14860 })
   14861 #else
   14862 #define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   14863   int32x2_t __s0 = __p0; \
   14864   int32x2_t __s1 = __p1; \
   14865   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   14866   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   14867   int64x2_t __ret; \
   14868   __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   14869   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   14870   __ret; \
   14871 })
   14872 #endif
   14873 
   14874 #ifdef __LITTLE_ENDIAN__
   14875 #define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   14876   int16x4_t __s0 = __p0; \
   14877   int16x4_t __s1 = __p1; \
   14878   int32x4_t __ret; \
   14879   __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   14880   __ret; \
   14881 })
   14882 #else
   14883 #define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   14884   int16x4_t __s0 = __p0; \
   14885   int16x4_t __s1 = __p1; \
   14886   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   14887   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   14888   int32x4_t __ret; \
   14889   __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   14890   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   14891   __ret; \
   14892 })
   14893 #endif
   14894 
   14895 #ifdef __LITTLE_ENDIAN__
   14896 __ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
   14897   uint64x2_t __ret;
   14898   __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
   14899   return __ret;
   14900 }
   14901 #else
   14902 __ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
   14903   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14904   uint64x2_t __ret;
   14905   __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
   14906   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14907   return __ret;
   14908 }
   14909 __ai uint64x2_t __noswap_vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
   14910   uint64x2_t __ret;
   14911   __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
   14912   return __ret;
   14913 }
   14914 #endif
   14915 
   14916 #ifdef __LITTLE_ENDIAN__
   14917 __ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
   14918   uint32x4_t __ret;
   14919   __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
   14920   return __ret;
   14921 }
   14922 #else
   14923 __ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
   14924   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14925   uint32x4_t __ret;
   14926   __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
   14927   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14928   return __ret;
   14929 }
   14930 __ai uint32x4_t __noswap_vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
   14931   uint32x4_t __ret;
   14932   __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
   14933   return __ret;
   14934 }
   14935 #endif
   14936 
   14937 #ifdef __LITTLE_ENDIAN__
   14938 __ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
   14939   int64x2_t __ret;
   14940   __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
   14941   return __ret;
   14942 }
   14943 #else
   14944 __ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
   14945   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   14946   int64x2_t __ret;
   14947   __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
   14948   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   14949   return __ret;
   14950 }
   14951 __ai int64x2_t __noswap_vmull_n_s32(int32x2_t __p0, int32_t __p1) {
   14952   int64x2_t __ret;
   14953   __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
   14954   return __ret;
   14955 }
   14956 #endif
   14957 
   14958 #ifdef __LITTLE_ENDIAN__
   14959 __ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
   14960   int32x4_t __ret;
   14961   __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
   14962   return __ret;
   14963 }
   14964 #else
   14965 __ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
   14966   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   14967   int32x4_t __ret;
   14968   __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
   14969   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   14970   return __ret;
   14971 }
   14972 __ai int32x4_t __noswap_vmull_n_s16(int16x4_t __p0, int16_t __p1) {
   14973   int32x4_t __ret;
   14974   __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
   14975   return __ret;
   14976 }
   14977 #endif
   14978 
   14979 #ifdef __LITTLE_ENDIAN__
   14980 __ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
   14981   poly8x8_t __ret;
   14982   __ret = ~__p0;
   14983   return __ret;
   14984 }
   14985 #else
   14986 __ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
   14987   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   14988   poly8x8_t __ret;
   14989   __ret = ~__rev0;
   14990   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   14991   return __ret;
   14992 }
   14993 #endif
   14994 
   14995 #ifdef __LITTLE_ENDIAN__
   14996 __ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
   14997   poly8x16_t __ret;
   14998   __ret = ~__p0;
   14999   return __ret;
   15000 }
   15001 #else
   15002 __ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
   15003   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15004   poly8x16_t __ret;
   15005   __ret = ~__rev0;
   15006   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15007   return __ret;
   15008 }
   15009 #endif
   15010 
   15011 #ifdef __LITTLE_ENDIAN__
   15012 __ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
   15013   uint8x16_t __ret;
   15014   __ret = ~__p0;
   15015   return __ret;
   15016 }
   15017 #else
   15018 __ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
   15019   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15020   uint8x16_t __ret;
   15021   __ret = ~__rev0;
   15022   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15023   return __ret;
   15024 }
   15025 #endif
   15026 
   15027 #ifdef __LITTLE_ENDIAN__
   15028 __ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
   15029   uint32x4_t __ret;
   15030   __ret = ~__p0;
   15031   return __ret;
   15032 }
   15033 #else
   15034 __ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
   15035   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15036   uint32x4_t __ret;
   15037   __ret = ~__rev0;
   15038   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15039   return __ret;
   15040 }
   15041 #endif
   15042 
   15043 #ifdef __LITTLE_ENDIAN__
   15044 __ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
   15045   uint16x8_t __ret;
   15046   __ret = ~__p0;
   15047   return __ret;
   15048 }
   15049 #else
   15050 __ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
   15051   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15052   uint16x8_t __ret;
   15053   __ret = ~__rev0;
   15054   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15055   return __ret;
   15056 }
   15057 #endif
   15058 
   15059 #ifdef __LITTLE_ENDIAN__
   15060 __ai int8x16_t vmvnq_s8(int8x16_t __p0) {
   15061   int8x16_t __ret;
   15062   __ret = ~__p0;
   15063   return __ret;
   15064 }
   15065 #else
   15066 __ai int8x16_t vmvnq_s8(int8x16_t __p0) {
   15067   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15068   int8x16_t __ret;
   15069   __ret = ~__rev0;
   15070   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15071   return __ret;
   15072 }
   15073 #endif
   15074 
   15075 #ifdef __LITTLE_ENDIAN__
   15076 __ai int32x4_t vmvnq_s32(int32x4_t __p0) {
   15077   int32x4_t __ret;
   15078   __ret = ~__p0;
   15079   return __ret;
   15080 }
   15081 #else
   15082 __ai int32x4_t vmvnq_s32(int32x4_t __p0) {
   15083   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15084   int32x4_t __ret;
   15085   __ret = ~__rev0;
   15086   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15087   return __ret;
   15088 }
   15089 #endif
   15090 
   15091 #ifdef __LITTLE_ENDIAN__
   15092 __ai int16x8_t vmvnq_s16(int16x8_t __p0) {
   15093   int16x8_t __ret;
   15094   __ret = ~__p0;
   15095   return __ret;
   15096 }
   15097 #else
   15098 __ai int16x8_t vmvnq_s16(int16x8_t __p0) {
   15099   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15100   int16x8_t __ret;
   15101   __ret = ~__rev0;
   15102   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15103   return __ret;
   15104 }
   15105 #endif
   15106 
   15107 #ifdef __LITTLE_ENDIAN__
   15108 __ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
   15109   uint8x8_t __ret;
   15110   __ret = ~__p0;
   15111   return __ret;
   15112 }
   15113 #else
   15114 __ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
   15115   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15116   uint8x8_t __ret;
   15117   __ret = ~__rev0;
   15118   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15119   return __ret;
   15120 }
   15121 #endif
   15122 
   15123 #ifdef __LITTLE_ENDIAN__
   15124 __ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
   15125   uint32x2_t __ret;
   15126   __ret = ~__p0;
   15127   return __ret;
   15128 }
   15129 #else
   15130 __ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
   15131   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15132   uint32x2_t __ret;
   15133   __ret = ~__rev0;
   15134   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15135   return __ret;
   15136 }
   15137 #endif
   15138 
   15139 #ifdef __LITTLE_ENDIAN__
   15140 __ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
   15141   uint16x4_t __ret;
   15142   __ret = ~__p0;
   15143   return __ret;
   15144 }
   15145 #else
   15146 __ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
   15147   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15148   uint16x4_t __ret;
   15149   __ret = ~__rev0;
   15150   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15151   return __ret;
   15152 }
   15153 #endif
   15154 
   15155 #ifdef __LITTLE_ENDIAN__
   15156 __ai int8x8_t vmvn_s8(int8x8_t __p0) {
   15157   int8x8_t __ret;
   15158   __ret = ~__p0;
   15159   return __ret;
   15160 }
   15161 #else
   15162 __ai int8x8_t vmvn_s8(int8x8_t __p0) {
   15163   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15164   int8x8_t __ret;
   15165   __ret = ~__rev0;
   15166   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15167   return __ret;
   15168 }
   15169 #endif
   15170 
   15171 #ifdef __LITTLE_ENDIAN__
   15172 __ai int32x2_t vmvn_s32(int32x2_t __p0) {
   15173   int32x2_t __ret;
   15174   __ret = ~__p0;
   15175   return __ret;
   15176 }
   15177 #else
   15178 __ai int32x2_t vmvn_s32(int32x2_t __p0) {
   15179   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15180   int32x2_t __ret;
   15181   __ret = ~__rev0;
   15182   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15183   return __ret;
   15184 }
   15185 #endif
   15186 
   15187 #ifdef __LITTLE_ENDIAN__
   15188 __ai int16x4_t vmvn_s16(int16x4_t __p0) {
   15189   int16x4_t __ret;
   15190   __ret = ~__p0;
   15191   return __ret;
   15192 }
   15193 #else
   15194 __ai int16x4_t vmvn_s16(int16x4_t __p0) {
   15195   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15196   int16x4_t __ret;
   15197   __ret = ~__rev0;
   15198   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15199   return __ret;
   15200 }
   15201 #endif
   15202 
   15203 #ifdef __LITTLE_ENDIAN__
   15204 __ai int8x16_t vnegq_s8(int8x16_t __p0) {
   15205   int8x16_t __ret;
   15206   __ret = -__p0;
   15207   return __ret;
   15208 }
   15209 #else
   15210 __ai int8x16_t vnegq_s8(int8x16_t __p0) {
   15211   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15212   int8x16_t __ret;
   15213   __ret = -__rev0;
   15214   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15215   return __ret;
   15216 }
   15217 #endif
   15218 
   15219 #ifdef __LITTLE_ENDIAN__
   15220 __ai float32x4_t vnegq_f32(float32x4_t __p0) {
   15221   float32x4_t __ret;
   15222   __ret = -__p0;
   15223   return __ret;
   15224 }
   15225 #else
   15226 __ai float32x4_t vnegq_f32(float32x4_t __p0) {
   15227   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15228   float32x4_t __ret;
   15229   __ret = -__rev0;
   15230   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15231   return __ret;
   15232 }
   15233 #endif
   15234 
   15235 #ifdef __LITTLE_ENDIAN__
   15236 __ai int32x4_t vnegq_s32(int32x4_t __p0) {
   15237   int32x4_t __ret;
   15238   __ret = -__p0;
   15239   return __ret;
   15240 }
   15241 #else
   15242 __ai int32x4_t vnegq_s32(int32x4_t __p0) {
   15243   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15244   int32x4_t __ret;
   15245   __ret = -__rev0;
   15246   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15247   return __ret;
   15248 }
   15249 #endif
   15250 
   15251 #ifdef __LITTLE_ENDIAN__
   15252 __ai int16x8_t vnegq_s16(int16x8_t __p0) {
   15253   int16x8_t __ret;
   15254   __ret = -__p0;
   15255   return __ret;
   15256 }
   15257 #else
   15258 __ai int16x8_t vnegq_s16(int16x8_t __p0) {
   15259   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15260   int16x8_t __ret;
   15261   __ret = -__rev0;
   15262   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15263   return __ret;
   15264 }
   15265 #endif
   15266 
   15267 #ifdef __LITTLE_ENDIAN__
   15268 __ai int8x8_t vneg_s8(int8x8_t __p0) {
   15269   int8x8_t __ret;
   15270   __ret = -__p0;
   15271   return __ret;
   15272 }
   15273 #else
   15274 __ai int8x8_t vneg_s8(int8x8_t __p0) {
   15275   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15276   int8x8_t __ret;
   15277   __ret = -__rev0;
   15278   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15279   return __ret;
   15280 }
   15281 #endif
   15282 
   15283 #ifdef __LITTLE_ENDIAN__
   15284 __ai float32x2_t vneg_f32(float32x2_t __p0) {
   15285   float32x2_t __ret;
   15286   __ret = -__p0;
   15287   return __ret;
   15288 }
   15289 #else
   15290 __ai float32x2_t vneg_f32(float32x2_t __p0) {
   15291   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15292   float32x2_t __ret;
   15293   __ret = -__rev0;
   15294   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15295   return __ret;
   15296 }
   15297 #endif
   15298 
   15299 #ifdef __LITTLE_ENDIAN__
   15300 __ai int32x2_t vneg_s32(int32x2_t __p0) {
   15301   int32x2_t __ret;
   15302   __ret = -__p0;
   15303   return __ret;
   15304 }
   15305 #else
   15306 __ai int32x2_t vneg_s32(int32x2_t __p0) {
   15307   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15308   int32x2_t __ret;
   15309   __ret = -__rev0;
   15310   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15311   return __ret;
   15312 }
   15313 #endif
   15314 
   15315 #ifdef __LITTLE_ENDIAN__
   15316 __ai int16x4_t vneg_s16(int16x4_t __p0) {
   15317   int16x4_t __ret;
   15318   __ret = -__p0;
   15319   return __ret;
   15320 }
   15321 #else
   15322 __ai int16x4_t vneg_s16(int16x4_t __p0) {
   15323   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15324   int16x4_t __ret;
   15325   __ret = -__rev0;
   15326   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15327   return __ret;
   15328 }
   15329 #endif
   15330 
   15331 #ifdef __LITTLE_ENDIAN__
   15332 __ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   15333   uint8x16_t __ret;
   15334   __ret = __p0 | ~__p1;
   15335   return __ret;
   15336 }
   15337 #else
   15338 __ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   15339   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15340   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15341   uint8x16_t __ret;
   15342   __ret = __rev0 | ~__rev1;
   15343   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15344   return __ret;
   15345 }
   15346 #endif
   15347 
   15348 #ifdef __LITTLE_ENDIAN__
   15349 __ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   15350   uint32x4_t __ret;
   15351   __ret = __p0 | ~__p1;
   15352   return __ret;
   15353 }
   15354 #else
   15355 __ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   15356   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15357   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15358   uint32x4_t __ret;
   15359   __ret = __rev0 | ~__rev1;
   15360   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15361   return __ret;
   15362 }
   15363 #endif
   15364 
   15365 #ifdef __LITTLE_ENDIAN__
   15366 __ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   15367   uint64x2_t __ret;
   15368   __ret = __p0 | ~__p1;
   15369   return __ret;
   15370 }
   15371 #else
   15372 __ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   15373   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15374   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   15375   uint64x2_t __ret;
   15376   __ret = __rev0 | ~__rev1;
   15377   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15378   return __ret;
   15379 }
   15380 #endif
   15381 
   15382 #ifdef __LITTLE_ENDIAN__
   15383 __ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   15384   uint16x8_t __ret;
   15385   __ret = __p0 | ~__p1;
   15386   return __ret;
   15387 }
   15388 #else
   15389 __ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   15390   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15391   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15392   uint16x8_t __ret;
   15393   __ret = __rev0 | ~__rev1;
   15394   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15395   return __ret;
   15396 }
   15397 #endif
   15398 
   15399 #ifdef __LITTLE_ENDIAN__
   15400 __ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
   15401   int8x16_t __ret;
   15402   __ret = __p0 | ~__p1;
   15403   return __ret;
   15404 }
   15405 #else
   15406 __ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
   15407   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15408   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15409   int8x16_t __ret;
   15410   __ret = __rev0 | ~__rev1;
   15411   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15412   return __ret;
   15413 }
   15414 #endif
   15415 
   15416 #ifdef __LITTLE_ENDIAN__
   15417 __ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
   15418   int32x4_t __ret;
   15419   __ret = __p0 | ~__p1;
   15420   return __ret;
   15421 }
   15422 #else
   15423 __ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
   15424   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15425   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15426   int32x4_t __ret;
   15427   __ret = __rev0 | ~__rev1;
   15428   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15429   return __ret;
   15430 }
   15431 #endif
   15432 
   15433 #ifdef __LITTLE_ENDIAN__
   15434 __ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
   15435   int64x2_t __ret;
   15436   __ret = __p0 | ~__p1;
   15437   return __ret;
   15438 }
   15439 #else
   15440 __ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
   15441   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15442   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   15443   int64x2_t __ret;
   15444   __ret = __rev0 | ~__rev1;
   15445   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15446   return __ret;
   15447 }
   15448 #endif
   15449 
   15450 #ifdef __LITTLE_ENDIAN__
   15451 __ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
   15452   int16x8_t __ret;
   15453   __ret = __p0 | ~__p1;
   15454   return __ret;
   15455 }
   15456 #else
   15457 __ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
   15458   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15459   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15460   int16x8_t __ret;
   15461   __ret = __rev0 | ~__rev1;
   15462   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15463   return __ret;
   15464 }
   15465 #endif
   15466 
   15467 #ifdef __LITTLE_ENDIAN__
   15468 __ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
   15469   uint8x8_t __ret;
   15470   __ret = __p0 | ~__p1;
   15471   return __ret;
   15472 }
   15473 #else
   15474 __ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
   15475   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15476   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15477   uint8x8_t __ret;
   15478   __ret = __rev0 | ~__rev1;
   15479   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15480   return __ret;
   15481 }
   15482 #endif
   15483 
   15484 #ifdef __LITTLE_ENDIAN__
   15485 __ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
   15486   uint32x2_t __ret;
   15487   __ret = __p0 | ~__p1;
   15488   return __ret;
   15489 }
   15490 #else
   15491 __ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
   15492   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15493   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   15494   uint32x2_t __ret;
   15495   __ret = __rev0 | ~__rev1;
   15496   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15497   return __ret;
   15498 }
   15499 #endif
   15500 
   15501 #ifdef __LITTLE_ENDIAN__
   15502 __ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
   15503   uint64x1_t __ret;
   15504   __ret = __p0 | ~__p1;
   15505   return __ret;
   15506 }
   15507 #else
   15508 __ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
   15509   uint64x1_t __ret;
   15510   __ret = __p0 | ~__p1;
   15511   return __ret;
   15512 }
   15513 #endif
   15514 
   15515 #ifdef __LITTLE_ENDIAN__
   15516 __ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
   15517   uint16x4_t __ret;
   15518   __ret = __p0 | ~__p1;
   15519   return __ret;
   15520 }
   15521 #else
   15522 __ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
   15523   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15524   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15525   uint16x4_t __ret;
   15526   __ret = __rev0 | ~__rev1;
   15527   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15528   return __ret;
   15529 }
   15530 #endif
   15531 
   15532 #ifdef __LITTLE_ENDIAN__
   15533 __ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
   15534   int8x8_t __ret;
   15535   __ret = __p0 | ~__p1;
   15536   return __ret;
   15537 }
   15538 #else
   15539 __ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
   15540   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15541   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15542   int8x8_t __ret;
   15543   __ret = __rev0 | ~__rev1;
   15544   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15545   return __ret;
   15546 }
   15547 #endif
   15548 
   15549 #ifdef __LITTLE_ENDIAN__
   15550 __ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
   15551   int32x2_t __ret;
   15552   __ret = __p0 | ~__p1;
   15553   return __ret;
   15554 }
   15555 #else
   15556 __ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
   15557   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15558   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   15559   int32x2_t __ret;
   15560   __ret = __rev0 | ~__rev1;
   15561   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15562   return __ret;
   15563 }
   15564 #endif
   15565 
   15566 #ifdef __LITTLE_ENDIAN__
   15567 __ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
   15568   int64x1_t __ret;
   15569   __ret = __p0 | ~__p1;
   15570   return __ret;
   15571 }
   15572 #else
   15573 __ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
   15574   int64x1_t __ret;
   15575   __ret = __p0 | ~__p1;
   15576   return __ret;
   15577 }
   15578 #endif
   15579 
   15580 #ifdef __LITTLE_ENDIAN__
   15581 __ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
   15582   int16x4_t __ret;
   15583   __ret = __p0 | ~__p1;
   15584   return __ret;
   15585 }
   15586 #else
   15587 __ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
   15588   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15589   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15590   int16x4_t __ret;
   15591   __ret = __rev0 | ~__rev1;
   15592   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15593   return __ret;
   15594 }
   15595 #endif
   15596 
   15597 #ifdef __LITTLE_ENDIAN__
   15598 __ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   15599   uint8x16_t __ret;
   15600   __ret = __p0 | __p1;
   15601   return __ret;
   15602 }
   15603 #else
   15604 __ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   15605   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15606   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15607   uint8x16_t __ret;
   15608   __ret = __rev0 | __rev1;
   15609   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15610   return __ret;
   15611 }
   15612 #endif
   15613 
   15614 #ifdef __LITTLE_ENDIAN__
   15615 __ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   15616   uint32x4_t __ret;
   15617   __ret = __p0 | __p1;
   15618   return __ret;
   15619 }
   15620 #else
   15621 __ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   15622   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15623   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15624   uint32x4_t __ret;
   15625   __ret = __rev0 | __rev1;
   15626   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15627   return __ret;
   15628 }
   15629 #endif
   15630 
   15631 #ifdef __LITTLE_ENDIAN__
   15632 __ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   15633   uint64x2_t __ret;
   15634   __ret = __p0 | __p1;
   15635   return __ret;
   15636 }
   15637 #else
   15638 __ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   15639   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15640   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   15641   uint64x2_t __ret;
   15642   __ret = __rev0 | __rev1;
   15643   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15644   return __ret;
   15645 }
   15646 #endif
   15647 
   15648 #ifdef __LITTLE_ENDIAN__
   15649 __ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   15650   uint16x8_t __ret;
   15651   __ret = __p0 | __p1;
   15652   return __ret;
   15653 }
   15654 #else
   15655 __ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   15656   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15657   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15658   uint16x8_t __ret;
   15659   __ret = __rev0 | __rev1;
   15660   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15661   return __ret;
   15662 }
   15663 #endif
   15664 
   15665 #ifdef __LITTLE_ENDIAN__
   15666 __ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
   15667   int8x16_t __ret;
   15668   __ret = __p0 | __p1;
   15669   return __ret;
   15670 }
   15671 #else
   15672 __ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
   15673   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15674   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15675   int8x16_t __ret;
   15676   __ret = __rev0 | __rev1;
   15677   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15678   return __ret;
   15679 }
   15680 #endif
   15681 
   15682 #ifdef __LITTLE_ENDIAN__
   15683 __ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
   15684   int32x4_t __ret;
   15685   __ret = __p0 | __p1;
   15686   return __ret;
   15687 }
   15688 #else
   15689 __ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
   15690   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15691   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15692   int32x4_t __ret;
   15693   __ret = __rev0 | __rev1;
   15694   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15695   return __ret;
   15696 }
   15697 #endif
   15698 
   15699 #ifdef __LITTLE_ENDIAN__
   15700 __ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
   15701   int64x2_t __ret;
   15702   __ret = __p0 | __p1;
   15703   return __ret;
   15704 }
   15705 #else
   15706 __ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
   15707   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15708   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   15709   int64x2_t __ret;
   15710   __ret = __rev0 | __rev1;
   15711   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15712   return __ret;
   15713 }
   15714 #endif
   15715 
   15716 #ifdef __LITTLE_ENDIAN__
   15717 __ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
   15718   int16x8_t __ret;
   15719   __ret = __p0 | __p1;
   15720   return __ret;
   15721 }
   15722 #else
   15723 __ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
   15724   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15725   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15726   int16x8_t __ret;
   15727   __ret = __rev0 | __rev1;
   15728   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15729   return __ret;
   15730 }
   15731 #endif
   15732 
   15733 #ifdef __LITTLE_ENDIAN__
   15734 __ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
   15735   uint8x8_t __ret;
   15736   __ret = __p0 | __p1;
   15737   return __ret;
   15738 }
   15739 #else
   15740 __ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
   15741   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15742   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15743   uint8x8_t __ret;
   15744   __ret = __rev0 | __rev1;
   15745   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15746   return __ret;
   15747 }
   15748 #endif
   15749 
   15750 #ifdef __LITTLE_ENDIAN__
   15751 __ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
   15752   uint32x2_t __ret;
   15753   __ret = __p0 | __p1;
   15754   return __ret;
   15755 }
   15756 #else
   15757 __ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
   15758   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15759   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   15760   uint32x2_t __ret;
   15761   __ret = __rev0 | __rev1;
   15762   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15763   return __ret;
   15764 }
   15765 #endif
   15766 
   15767 #ifdef __LITTLE_ENDIAN__
   15768 __ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
   15769   uint64x1_t __ret;
   15770   __ret = __p0 | __p1;
   15771   return __ret;
   15772 }
   15773 #else
   15774 __ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
   15775   uint64x1_t __ret;
   15776   __ret = __p0 | __p1;
   15777   return __ret;
   15778 }
   15779 #endif
   15780 
   15781 #ifdef __LITTLE_ENDIAN__
   15782 __ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
   15783   uint16x4_t __ret;
   15784   __ret = __p0 | __p1;
   15785   return __ret;
   15786 }
   15787 #else
   15788 __ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
   15789   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15790   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15791   uint16x4_t __ret;
   15792   __ret = __rev0 | __rev1;
   15793   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15794   return __ret;
   15795 }
   15796 #endif
   15797 
   15798 #ifdef __LITTLE_ENDIAN__
   15799 __ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
   15800   int8x8_t __ret;
   15801   __ret = __p0 | __p1;
   15802   return __ret;
   15803 }
   15804 #else
   15805 __ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
   15806   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15807   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15808   int8x8_t __ret;
   15809   __ret = __rev0 | __rev1;
   15810   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15811   return __ret;
   15812 }
   15813 #endif
   15814 
   15815 #ifdef __LITTLE_ENDIAN__
   15816 __ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
   15817   int32x2_t __ret;
   15818   __ret = __p0 | __p1;
   15819   return __ret;
   15820 }
   15821 #else
   15822 __ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
   15823   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15824   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   15825   int32x2_t __ret;
   15826   __ret = __rev0 | __rev1;
   15827   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15828   return __ret;
   15829 }
   15830 #endif
   15831 
   15832 #ifdef __LITTLE_ENDIAN__
   15833 __ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
   15834   int64x1_t __ret;
   15835   __ret = __p0 | __p1;
   15836   return __ret;
   15837 }
   15838 #else
   15839 __ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
   15840   int64x1_t __ret;
   15841   __ret = __p0 | __p1;
   15842   return __ret;
   15843 }
   15844 #endif
   15845 
   15846 #ifdef __LITTLE_ENDIAN__
   15847 __ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
   15848   int16x4_t __ret;
   15849   __ret = __p0 | __p1;
   15850   return __ret;
   15851 }
   15852 #else
   15853 __ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
   15854   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15855   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15856   int16x4_t __ret;
   15857   __ret = __rev0 | __rev1;
   15858   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15859   return __ret;
   15860 }
   15861 #endif
   15862 
   15863 #ifdef __LITTLE_ENDIAN__
   15864 __ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
   15865   uint16x8_t __ret;
   15866   __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   15867   return __ret;
   15868 }
   15869 #else
   15870 __ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
   15871   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15872   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15873   uint16x8_t __ret;
   15874   __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   15875   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15876   return __ret;
   15877 }
   15878 #endif
   15879 
   15880 #ifdef __LITTLE_ENDIAN__
   15881 __ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
   15882   uint64x2_t __ret;
   15883   __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   15884   return __ret;
   15885 }
   15886 #else
   15887 __ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
   15888   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15889   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15890   uint64x2_t __ret;
   15891   __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   15892   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15893   return __ret;
   15894 }
   15895 #endif
   15896 
   15897 #ifdef __LITTLE_ENDIAN__
   15898 __ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
   15899   uint32x4_t __ret;
   15900   __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   15901   return __ret;
   15902 }
   15903 #else
   15904 __ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
   15905   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15906   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15907   uint32x4_t __ret;
   15908   __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   15909   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15910   return __ret;
   15911 }
   15912 #endif
   15913 
   15914 #ifdef __LITTLE_ENDIAN__
   15915 __ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
   15916   int16x8_t __ret;
   15917   __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   15918   return __ret;
   15919 }
   15920 #else
   15921 __ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
   15922   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   15923   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   15924   int16x8_t __ret;
   15925   __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   15926   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   15927   return __ret;
   15928 }
   15929 #endif
   15930 
   15931 #ifdef __LITTLE_ENDIAN__
   15932 __ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
   15933   int64x2_t __ret;
   15934   __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
   15935   return __ret;
   15936 }
   15937 #else
   15938 __ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
   15939   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   15940   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   15941   int64x2_t __ret;
   15942   __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
   15943   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   15944   return __ret;
   15945 }
   15946 #endif
   15947 
   15948 #ifdef __LITTLE_ENDIAN__
   15949 __ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
   15950   int32x4_t __ret;
   15951   __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   15952   return __ret;
   15953 }
   15954 #else
   15955 __ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
   15956   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15957   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15958   int32x4_t __ret;
   15959   __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   15960   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15961   return __ret;
   15962 }
   15963 #endif
   15964 
   15965 #ifdef __LITTLE_ENDIAN__
   15966 __ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
   15967   uint16x4_t __ret;
   15968   __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   15969   return __ret;
   15970 }
   15971 #else
   15972 __ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
   15973   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   15974   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   15975   uint16x4_t __ret;
   15976   __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   15977   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   15978   return __ret;
   15979 }
   15980 #endif
   15981 
   15982 #ifdef __LITTLE_ENDIAN__
   15983 __ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
   15984   uint64x1_t __ret;
   15985   __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   15986   return __ret;
   15987 }
   15988 #else
   15989 __ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
   15990   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   15991   uint64x1_t __ret;
   15992   __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 19);
   15993   return __ret;
   15994 }
   15995 #endif
   15996 
   15997 #ifdef __LITTLE_ENDIAN__
   15998 __ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
   15999   uint32x2_t __ret;
   16000   __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   16001   return __ret;
   16002 }
   16003 #else
   16004 __ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
   16005   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16006   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16007   uint32x2_t __ret;
   16008   __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   16009   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16010   return __ret;
   16011 }
   16012 #endif
   16013 
   16014 #ifdef __LITTLE_ENDIAN__
   16015 __ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
   16016   int16x4_t __ret;
   16017   __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   16018   return __ret;
   16019 }
   16020 #else
   16021 __ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
   16022   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16023   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16024   int16x4_t __ret;
   16025   __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   16026   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16027   return __ret;
   16028 }
   16029 #endif
   16030 
   16031 #ifdef __LITTLE_ENDIAN__
   16032 __ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
   16033   int64x1_t __ret;
   16034   __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   16035   return __ret;
   16036 }
   16037 #else
   16038 __ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
   16039   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16040   int64x1_t __ret;
   16041   __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 3);
   16042   return __ret;
   16043 }
   16044 #endif
   16045 
   16046 #ifdef __LITTLE_ENDIAN__
   16047 __ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
   16048   int32x2_t __ret;
   16049   __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   16050   return __ret;
   16051 }
   16052 #else
   16053 __ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
   16054   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16055   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16056   int32x2_t __ret;
   16057   __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   16058   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16059   return __ret;
   16060 }
   16061 #endif
   16062 
   16063 #ifdef __LITTLE_ENDIAN__
   16064 __ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   16065   uint8x8_t __ret;
   16066   __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   16067   return __ret;
   16068 }
   16069 #else
   16070 __ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   16071   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16072   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16073   uint8x8_t __ret;
   16074   __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   16075   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16076   return __ret;
   16077 }
   16078 #endif
   16079 
   16080 #ifdef __LITTLE_ENDIAN__
   16081 __ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   16082   uint32x2_t __ret;
   16083   __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   16084   return __ret;
   16085 }
   16086 #else
   16087 __ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   16088   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16089   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16090   uint32x2_t __ret;
   16091   __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   16092   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16093   return __ret;
   16094 }
   16095 #endif
   16096 
   16097 #ifdef __LITTLE_ENDIAN__
   16098 __ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   16099   uint16x4_t __ret;
   16100   __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   16101   return __ret;
   16102 }
   16103 #else
   16104 __ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   16105   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16106   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16107   uint16x4_t __ret;
   16108   __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   16109   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16110   return __ret;
   16111 }
   16112 #endif
   16113 
   16114 #ifdef __LITTLE_ENDIAN__
   16115 __ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
   16116   int8x8_t __ret;
   16117   __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   16118   return __ret;
   16119 }
   16120 #else
   16121 __ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
   16122   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16123   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16124   int8x8_t __ret;
   16125   __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   16126   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16127   return __ret;
   16128 }
   16129 #endif
   16130 
   16131 #ifdef __LITTLE_ENDIAN__
   16132 __ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
   16133   float32x2_t __ret;
   16134   __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   16135   return __ret;
   16136 }
   16137 #else
   16138 __ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
   16139   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16140   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16141   float32x2_t __ret;
   16142   __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   16143   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16144   return __ret;
   16145 }
   16146 #endif
   16147 
   16148 #ifdef __LITTLE_ENDIAN__
   16149 __ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
   16150   int32x2_t __ret;
   16151   __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   16152   return __ret;
   16153 }
   16154 #else
   16155 __ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
   16156   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16157   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16158   int32x2_t __ret;
   16159   __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   16160   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16161   return __ret;
   16162 }
   16163 #endif
   16164 
   16165 #ifdef __LITTLE_ENDIAN__
   16166 __ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
   16167   int16x4_t __ret;
   16168   __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   16169   return __ret;
   16170 }
   16171 #else
   16172 __ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
   16173   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16174   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16175   int16x4_t __ret;
   16176   __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   16177   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16178   return __ret;
   16179 }
   16180 #endif
   16181 
   16182 #ifdef __LITTLE_ENDIAN__
   16183 __ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
   16184   uint16x8_t __ret;
   16185   __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 49);
   16186   return __ret;
   16187 }
   16188 #else
   16189 __ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
   16190   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16191   uint16x8_t __ret;
   16192   __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 49);
   16193   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16194   return __ret;
   16195 }
   16196 #endif
   16197 
   16198 #ifdef __LITTLE_ENDIAN__
   16199 __ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
   16200   uint64x2_t __ret;
   16201   __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 51);
   16202   return __ret;
   16203 }
   16204 #else
   16205 __ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
   16206   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16207   uint64x2_t __ret;
   16208   __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 51);
   16209   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16210   return __ret;
   16211 }
   16212 #endif
   16213 
   16214 #ifdef __LITTLE_ENDIAN__
   16215 __ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
   16216   uint32x4_t __ret;
   16217   __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 50);
   16218   return __ret;
   16219 }
   16220 #else
   16221 __ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
   16222   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16223   uint32x4_t __ret;
   16224   __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 50);
   16225   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16226   return __ret;
   16227 }
   16228 #endif
   16229 
   16230 #ifdef __LITTLE_ENDIAN__
   16231 __ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
   16232   int16x8_t __ret;
   16233   __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 33);
   16234   return __ret;
   16235 }
   16236 #else
   16237 __ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
   16238   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16239   int16x8_t __ret;
   16240   __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 33);
   16241   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16242   return __ret;
   16243 }
   16244 #endif
   16245 
   16246 #ifdef __LITTLE_ENDIAN__
   16247 __ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
   16248   int64x2_t __ret;
   16249   __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 35);
   16250   return __ret;
   16251 }
   16252 #else
   16253 __ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
   16254   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16255   int64x2_t __ret;
   16256   __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 35);
   16257   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16258   return __ret;
   16259 }
   16260 #endif
   16261 
   16262 #ifdef __LITTLE_ENDIAN__
   16263 __ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
   16264   int32x4_t __ret;
   16265   __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 34);
   16266   return __ret;
   16267 }
   16268 #else
   16269 __ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
   16270   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16271   int32x4_t __ret;
   16272   __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 34);
   16273   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16274   return __ret;
   16275 }
   16276 #endif
   16277 
   16278 #ifdef __LITTLE_ENDIAN__
   16279 __ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
   16280   uint16x4_t __ret;
   16281   __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 17);
   16282   return __ret;
   16283 }
   16284 #else
   16285 __ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
   16286   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16287   uint16x4_t __ret;
   16288   __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 17);
   16289   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16290   return __ret;
   16291 }
   16292 #endif
   16293 
   16294 #ifdef __LITTLE_ENDIAN__
   16295 __ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
   16296   uint64x1_t __ret;
   16297   __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 19);
   16298   return __ret;
   16299 }
   16300 #else
   16301 __ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
   16302   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16303   uint64x1_t __ret;
   16304   __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 19);
   16305   return __ret;
   16306 }
   16307 #endif
   16308 
   16309 #ifdef __LITTLE_ENDIAN__
   16310 __ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
   16311   uint32x2_t __ret;
   16312   __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 18);
   16313   return __ret;
   16314 }
   16315 #else
   16316 __ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
   16317   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16318   uint32x2_t __ret;
   16319   __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 18);
   16320   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16321   return __ret;
   16322 }
   16323 #endif
   16324 
   16325 #ifdef __LITTLE_ENDIAN__
   16326 __ai int16x4_t vpaddl_s8(int8x8_t __p0) {
   16327   int16x4_t __ret;
   16328   __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 1);
   16329   return __ret;
   16330 }
   16331 #else
   16332 __ai int16x4_t vpaddl_s8(int8x8_t __p0) {
   16333   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16334   int16x4_t __ret;
   16335   __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 1);
   16336   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16337   return __ret;
   16338 }
   16339 #endif
   16340 
   16341 #ifdef __LITTLE_ENDIAN__
   16342 __ai int64x1_t vpaddl_s32(int32x2_t __p0) {
   16343   int64x1_t __ret;
   16344   __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 3);
   16345   return __ret;
   16346 }
   16347 #else
   16348 __ai int64x1_t vpaddl_s32(int32x2_t __p0) {
   16349   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16350   int64x1_t __ret;
   16351   __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 3);
   16352   return __ret;
   16353 }
   16354 #endif
   16355 
   16356 #ifdef __LITTLE_ENDIAN__
   16357 __ai int32x2_t vpaddl_s16(int16x4_t __p0) {
   16358   int32x2_t __ret;
   16359   __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 2);
   16360   return __ret;
   16361 }
   16362 #else
   16363 __ai int32x2_t vpaddl_s16(int16x4_t __p0) {
   16364   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16365   int32x2_t __ret;
   16366   __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 2);
   16367   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16368   return __ret;
   16369 }
   16370 #endif
   16371 
   16372 #ifdef __LITTLE_ENDIAN__
   16373 __ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
   16374   uint8x8_t __ret;
   16375   __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   16376   return __ret;
   16377 }
   16378 #else
   16379 __ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
   16380   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16381   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16382   uint8x8_t __ret;
   16383   __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   16384   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16385   return __ret;
   16386 }
   16387 #endif
   16388 
   16389 #ifdef __LITTLE_ENDIAN__
   16390 __ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
   16391   uint32x2_t __ret;
   16392   __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   16393   return __ret;
   16394 }
   16395 #else
   16396 __ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
   16397   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16398   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16399   uint32x2_t __ret;
   16400   __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   16401   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16402   return __ret;
   16403 }
   16404 #endif
   16405 
   16406 #ifdef __LITTLE_ENDIAN__
   16407 __ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
   16408   uint16x4_t __ret;
   16409   __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   16410   return __ret;
   16411 }
   16412 #else
   16413 __ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
   16414   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16415   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16416   uint16x4_t __ret;
   16417   __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   16418   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16419   return __ret;
   16420 }
   16421 #endif
   16422 
   16423 #ifdef __LITTLE_ENDIAN__
   16424 __ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
   16425   int8x8_t __ret;
   16426   __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   16427   return __ret;
   16428 }
   16429 #else
   16430 __ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
   16431   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16432   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16433   int8x8_t __ret;
   16434   __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   16435   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16436   return __ret;
   16437 }
   16438 #endif
   16439 
   16440 #ifdef __LITTLE_ENDIAN__
   16441 __ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
   16442   float32x2_t __ret;
   16443   __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   16444   return __ret;
   16445 }
   16446 #else
   16447 __ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
   16448   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16449   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16450   float32x2_t __ret;
   16451   __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   16452   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16453   return __ret;
   16454 }
   16455 #endif
   16456 
   16457 #ifdef __LITTLE_ENDIAN__
   16458 __ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
   16459   int32x2_t __ret;
   16460   __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   16461   return __ret;
   16462 }
   16463 #else
   16464 __ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
   16465   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16466   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16467   int32x2_t __ret;
   16468   __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   16469   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16470   return __ret;
   16471 }
   16472 #endif
   16473 
   16474 #ifdef __LITTLE_ENDIAN__
   16475 __ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
   16476   int16x4_t __ret;
   16477   __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   16478   return __ret;
   16479 }
   16480 #else
   16481 __ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
   16482   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16483   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16484   int16x4_t __ret;
   16485   __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   16486   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16487   return __ret;
   16488 }
   16489 #endif
   16490 
   16491 #ifdef __LITTLE_ENDIAN__
   16492 __ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
   16493   uint8x8_t __ret;
   16494   __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   16495   return __ret;
   16496 }
   16497 #else
   16498 __ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
   16499   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16500   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16501   uint8x8_t __ret;
   16502   __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   16503   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16504   return __ret;
   16505 }
   16506 #endif
   16507 
   16508 #ifdef __LITTLE_ENDIAN__
   16509 __ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
   16510   uint32x2_t __ret;
   16511   __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   16512   return __ret;
   16513 }
   16514 #else
   16515 __ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
   16516   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16517   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16518   uint32x2_t __ret;
   16519   __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   16520   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16521   return __ret;
   16522 }
   16523 #endif
   16524 
   16525 #ifdef __LITTLE_ENDIAN__
   16526 __ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
   16527   uint16x4_t __ret;
   16528   __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   16529   return __ret;
   16530 }
   16531 #else
   16532 __ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
   16533   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16534   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16535   uint16x4_t __ret;
   16536   __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   16537   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16538   return __ret;
   16539 }
   16540 #endif
   16541 
   16542 #ifdef __LITTLE_ENDIAN__
   16543 __ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
   16544   int8x8_t __ret;
   16545   __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   16546   return __ret;
   16547 }
   16548 #else
   16549 __ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
   16550   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16551   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16552   int8x8_t __ret;
   16553   __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   16554   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16555   return __ret;
   16556 }
   16557 #endif
   16558 
   16559 #ifdef __LITTLE_ENDIAN__
   16560 __ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
   16561   float32x2_t __ret;
   16562   __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   16563   return __ret;
   16564 }
   16565 #else
   16566 __ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
   16567   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16568   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16569   float32x2_t __ret;
   16570   __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   16571   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16572   return __ret;
   16573 }
   16574 #endif
   16575 
   16576 #ifdef __LITTLE_ENDIAN__
   16577 __ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
   16578   int32x2_t __ret;
   16579   __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   16580   return __ret;
   16581 }
   16582 #else
   16583 __ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
   16584   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16585   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16586   int32x2_t __ret;
   16587   __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   16588   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16589   return __ret;
   16590 }
   16591 #endif
   16592 
   16593 #ifdef __LITTLE_ENDIAN__
   16594 __ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
   16595   int16x4_t __ret;
   16596   __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   16597   return __ret;
   16598 }
   16599 #else
   16600 __ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
   16601   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16602   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16603   int16x4_t __ret;
   16604   __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   16605   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16606   return __ret;
   16607 }
   16608 #endif
   16609 
   16610 #ifdef __LITTLE_ENDIAN__
   16611 __ai int8x16_t vqabsq_s8(int8x16_t __p0) {
   16612   int8x16_t __ret;
   16613   __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 32);
   16614   return __ret;
   16615 }
   16616 #else
   16617 __ai int8x16_t vqabsq_s8(int8x16_t __p0) {
   16618   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16619   int8x16_t __ret;
   16620   __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 32);
   16621   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16622   return __ret;
   16623 }
   16624 #endif
   16625 
   16626 #ifdef __LITTLE_ENDIAN__
   16627 __ai int32x4_t vqabsq_s32(int32x4_t __p0) {
   16628   int32x4_t __ret;
   16629   __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 34);
   16630   return __ret;
   16631 }
   16632 #else
   16633 __ai int32x4_t vqabsq_s32(int32x4_t __p0) {
   16634   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16635   int32x4_t __ret;
   16636   __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 34);
   16637   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16638   return __ret;
   16639 }
   16640 #endif
   16641 
   16642 #ifdef __LITTLE_ENDIAN__
   16643 __ai int16x8_t vqabsq_s16(int16x8_t __p0) {
   16644   int16x8_t __ret;
   16645   __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 33);
   16646   return __ret;
   16647 }
   16648 #else
   16649 __ai int16x8_t vqabsq_s16(int16x8_t __p0) {
   16650   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16651   int16x8_t __ret;
   16652   __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 33);
   16653   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16654   return __ret;
   16655 }
   16656 #endif
   16657 
   16658 #ifdef __LITTLE_ENDIAN__
   16659 __ai int8x8_t vqabs_s8(int8x8_t __p0) {
   16660   int8x8_t __ret;
   16661   __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 0);
   16662   return __ret;
   16663 }
   16664 #else
   16665 __ai int8x8_t vqabs_s8(int8x8_t __p0) {
   16666   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16667   int8x8_t __ret;
   16668   __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 0);
   16669   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16670   return __ret;
   16671 }
   16672 #endif
   16673 
   16674 #ifdef __LITTLE_ENDIAN__
   16675 __ai int32x2_t vqabs_s32(int32x2_t __p0) {
   16676   int32x2_t __ret;
   16677   __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 2);
   16678   return __ret;
   16679 }
   16680 #else
   16681 __ai int32x2_t vqabs_s32(int32x2_t __p0) {
   16682   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16683   int32x2_t __ret;
   16684   __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 2);
   16685   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16686   return __ret;
   16687 }
   16688 #endif
   16689 
   16690 #ifdef __LITTLE_ENDIAN__
   16691 __ai int16x4_t vqabs_s16(int16x4_t __p0) {
   16692   int16x4_t __ret;
   16693   __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 1);
   16694   return __ret;
   16695 }
   16696 #else
   16697 __ai int16x4_t vqabs_s16(int16x4_t __p0) {
   16698   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16699   int16x4_t __ret;
   16700   __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 1);
   16701   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16702   return __ret;
   16703 }
   16704 #endif
   16705 
   16706 #ifdef __LITTLE_ENDIAN__
   16707 __ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   16708   uint8x16_t __ret;
   16709   __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   16710   return __ret;
   16711 }
   16712 #else
   16713 __ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   16714   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16715   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16716   uint8x16_t __ret;
   16717   __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   16718   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16719   return __ret;
   16720 }
   16721 #endif
   16722 
   16723 #ifdef __LITTLE_ENDIAN__
   16724 __ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   16725   uint32x4_t __ret;
   16726   __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   16727   return __ret;
   16728 }
   16729 #else
   16730 __ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   16731   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16732   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16733   uint32x4_t __ret;
   16734   __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   16735   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16736   return __ret;
   16737 }
   16738 #endif
   16739 
   16740 #ifdef __LITTLE_ENDIAN__
   16741 __ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   16742   uint64x2_t __ret;
   16743   __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   16744   return __ret;
   16745 }
   16746 #else
   16747 __ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   16748   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16749   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16750   uint64x2_t __ret;
   16751   __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   16752   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16753   return __ret;
   16754 }
   16755 #endif
   16756 
   16757 #ifdef __LITTLE_ENDIAN__
   16758 __ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   16759   uint16x8_t __ret;
   16760   __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   16761   return __ret;
   16762 }
   16763 #else
   16764 __ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   16765   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16766   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16767   uint16x8_t __ret;
   16768   __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   16769   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16770   return __ret;
   16771 }
   16772 #endif
   16773 
   16774 #ifdef __LITTLE_ENDIAN__
   16775 __ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   16776   int8x16_t __ret;
   16777   __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   16778   return __ret;
   16779 }
   16780 #else
   16781 __ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   16782   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16783   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16784   int8x16_t __ret;
   16785   __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   16786   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   16787   return __ret;
   16788 }
   16789 #endif
   16790 
   16791 #ifdef __LITTLE_ENDIAN__
   16792 __ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   16793   int32x4_t __ret;
   16794   __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   16795   return __ret;
   16796 }
   16797 #else
   16798 __ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   16799   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16800   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16801   int32x4_t __ret;
   16802   __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   16803   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16804   return __ret;
   16805 }
   16806 __ai int32x4_t __noswap_vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   16807   int32x4_t __ret;
   16808   __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   16809   return __ret;
   16810 }
   16811 #endif
   16812 
   16813 #ifdef __LITTLE_ENDIAN__
   16814 __ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
   16815   int64x2_t __ret;
   16816   __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
   16817   return __ret;
   16818 }
   16819 #else
   16820 __ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
   16821   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16822   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16823   int64x2_t __ret;
   16824   __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
   16825   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16826   return __ret;
   16827 }
   16828 #endif
   16829 
   16830 #ifdef __LITTLE_ENDIAN__
   16831 __ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   16832   int16x8_t __ret;
   16833   __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   16834   return __ret;
   16835 }
   16836 #else
   16837 __ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   16838   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16839   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16840   int16x8_t __ret;
   16841   __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   16842   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16843   return __ret;
   16844 }
   16845 __ai int16x8_t __noswap_vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   16846   int16x8_t __ret;
   16847   __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   16848   return __ret;
   16849 }
   16850 #endif
   16851 
   16852 #ifdef __LITTLE_ENDIAN__
   16853 __ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   16854   uint8x8_t __ret;
   16855   __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   16856   return __ret;
   16857 }
   16858 #else
   16859 __ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   16860   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16861   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16862   uint8x8_t __ret;
   16863   __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   16864   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16865   return __ret;
   16866 }
   16867 #endif
   16868 
   16869 #ifdef __LITTLE_ENDIAN__
   16870 __ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   16871   uint32x2_t __ret;
   16872   __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   16873   return __ret;
   16874 }
   16875 #else
   16876 __ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   16877   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16878   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16879   uint32x2_t __ret;
   16880   __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   16881   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16882   return __ret;
   16883 }
   16884 #endif
   16885 
   16886 #ifdef __LITTLE_ENDIAN__
   16887 __ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
   16888   uint64x1_t __ret;
   16889   __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   16890   return __ret;
   16891 }
   16892 #else
   16893 __ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
   16894   uint64x1_t __ret;
   16895   __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   16896   return __ret;
   16897 }
   16898 #endif
   16899 
   16900 #ifdef __LITTLE_ENDIAN__
   16901 __ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   16902   uint16x4_t __ret;
   16903   __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   16904   return __ret;
   16905 }
   16906 #else
   16907 __ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   16908   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16909   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16910   uint16x4_t __ret;
   16911   __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   16912   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16913   return __ret;
   16914 }
   16915 #endif
   16916 
   16917 #ifdef __LITTLE_ENDIAN__
   16918 __ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
   16919   int8x8_t __ret;
   16920   __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   16921   return __ret;
   16922 }
   16923 #else
   16924 __ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
   16925   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   16926   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   16927   int8x8_t __ret;
   16928   __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   16929   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   16930   return __ret;
   16931 }
   16932 #endif
   16933 
   16934 #ifdef __LITTLE_ENDIAN__
   16935 __ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
   16936   int32x2_t __ret;
   16937   __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   16938   return __ret;
   16939 }
   16940 #else
   16941 __ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
   16942   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   16943   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   16944   int32x2_t __ret;
   16945   __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   16946   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   16947   return __ret;
   16948 }
   16949 __ai int32x2_t __noswap_vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
   16950   int32x2_t __ret;
   16951   __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   16952   return __ret;
   16953 }
   16954 #endif
   16955 
   16956 #ifdef __LITTLE_ENDIAN__
   16957 __ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
   16958   int64x1_t __ret;
   16959   __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   16960   return __ret;
   16961 }
   16962 #else
   16963 __ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
   16964   int64x1_t __ret;
   16965   __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   16966   return __ret;
   16967 }
   16968 #endif
   16969 
   16970 #ifdef __LITTLE_ENDIAN__
   16971 __ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
   16972   int16x4_t __ret;
   16973   __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   16974   return __ret;
   16975 }
   16976 #else
   16977 __ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
   16978   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   16979   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   16980   int16x4_t __ret;
   16981   __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   16982   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   16983   return __ret;
   16984 }
   16985 __ai int16x4_t __noswap_vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
   16986   int16x4_t __ret;
   16987   __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   16988   return __ret;
   16989 }
   16990 #endif
   16991 
   16992 #ifdef __LITTLE_ENDIAN__
   16993 __ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   16994   int64x2_t __ret;
   16995   __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
   16996   return __ret;
   16997 }
   16998 #else
   16999 __ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   17000   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17001   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   17002   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   17003   int64x2_t __ret;
   17004   __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
   17005   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17006   return __ret;
   17007 }
   17008 __ai int64x2_t __noswap_vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   17009   int64x2_t __ret;
   17010   __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
   17011   return __ret;
   17012 }
   17013 #endif
   17014 
   17015 #ifdef __LITTLE_ENDIAN__
   17016 __ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   17017   int32x4_t __ret;
   17018   __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
   17019   return __ret;
   17020 }
   17021 #else
   17022 __ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   17023   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17024   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   17025   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   17026   int32x4_t __ret;
   17027   __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
   17028   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17029   return __ret;
   17030 }
   17031 __ai int32x4_t __noswap_vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   17032   int32x4_t __ret;
   17033   __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
   17034   return __ret;
   17035 }
   17036 #endif
   17037 
   17038 #ifdef __LITTLE_ENDIAN__
   17039 #define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   17040   int64x2_t __s0 = __p0; \
   17041   int32x2_t __s1 = __p1; \
   17042   int32x2_t __s2 = __p2; \
   17043   int64x2_t __ret; \
   17044   __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   17045   __ret; \
   17046 })
   17047 #else
   17048 #define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   17049   int64x2_t __s0 = __p0; \
   17050   int32x2_t __s1 = __p1; \
   17051   int32x2_t __s2 = __p2; \
   17052   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   17053   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   17054   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   17055   int64x2_t __ret; \
   17056   __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   17057   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   17058   __ret; \
   17059 })
   17060 #endif
   17061 
   17062 #ifdef __LITTLE_ENDIAN__
   17063 #define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   17064   int32x4_t __s0 = __p0; \
   17065   int16x4_t __s1 = __p1; \
   17066   int16x4_t __s2 = __p2; \
   17067   int32x4_t __ret; \
   17068   __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   17069   __ret; \
   17070 })
   17071 #else
   17072 #define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   17073   int32x4_t __s0 = __p0; \
   17074   int16x4_t __s1 = __p1; \
   17075   int16x4_t __s2 = __p2; \
   17076   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   17077   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   17078   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   17079   int32x4_t __ret; \
   17080   __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   17081   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   17082   __ret; \
   17083 })
   17084 #endif
   17085 
   17086 #ifdef __LITTLE_ENDIAN__
   17087 __ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   17088   int64x2_t __ret;
   17089   __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
   17090   return __ret;
   17091 }
   17092 #else
   17093 __ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   17094   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17095   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   17096   int64x2_t __ret;
   17097   __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
   17098   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17099   return __ret;
   17100 }
   17101 __ai int64x2_t __noswap_vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   17102   int64x2_t __ret;
   17103   __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
   17104   return __ret;
   17105 }
   17106 #endif
   17107 
   17108 #ifdef __LITTLE_ENDIAN__
   17109 __ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   17110   int32x4_t __ret;
   17111   __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
   17112   return __ret;
   17113 }
   17114 #else
   17115 __ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   17116   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17117   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   17118   int32x4_t __ret;
   17119   __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
   17120   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17121   return __ret;
   17122 }
   17123 __ai int32x4_t __noswap_vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   17124   int32x4_t __ret;
   17125   __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
   17126   return __ret;
   17127 }
   17128 #endif
   17129 
   17130 #ifdef __LITTLE_ENDIAN__
   17131 __ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   17132   int64x2_t __ret;
   17133   __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
   17134   return __ret;
   17135 }
   17136 #else
   17137 __ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   17138   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17139   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   17140   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   17141   int64x2_t __ret;
   17142   __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
   17143   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17144   return __ret;
   17145 }
   17146 __ai int64x2_t __noswap_vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   17147   int64x2_t __ret;
   17148   __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
   17149   return __ret;
   17150 }
   17151 #endif
   17152 
   17153 #ifdef __LITTLE_ENDIAN__
   17154 __ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   17155   int32x4_t __ret;
   17156   __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
   17157   return __ret;
   17158 }
   17159 #else
   17160 __ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   17161   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17162   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   17163   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   17164   int32x4_t __ret;
   17165   __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
   17166   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17167   return __ret;
   17168 }
   17169 __ai int32x4_t __noswap_vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   17170   int32x4_t __ret;
   17171   __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
   17172   return __ret;
   17173 }
   17174 #endif
   17175 
   17176 #ifdef __LITTLE_ENDIAN__
   17177 #define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   17178   int64x2_t __s0 = __p0; \
   17179   int32x2_t __s1 = __p1; \
   17180   int32x2_t __s2 = __p2; \
   17181   int64x2_t __ret; \
   17182   __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   17183   __ret; \
   17184 })
   17185 #else
   17186 #define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   17187   int64x2_t __s0 = __p0; \
   17188   int32x2_t __s1 = __p1; \
   17189   int32x2_t __s2 = __p2; \
   17190   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   17191   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   17192   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   17193   int64x2_t __ret; \
   17194   __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   17195   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   17196   __ret; \
   17197 })
   17198 #endif
   17199 
   17200 #ifdef __LITTLE_ENDIAN__
   17201 #define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   17202   int32x4_t __s0 = __p0; \
   17203   int16x4_t __s1 = __p1; \
   17204   int16x4_t __s2 = __p2; \
   17205   int32x4_t __ret; \
   17206   __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   17207   __ret; \
   17208 })
   17209 #else
   17210 #define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   17211   int32x4_t __s0 = __p0; \
   17212   int16x4_t __s1 = __p1; \
   17213   int16x4_t __s2 = __p2; \
   17214   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   17215   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   17216   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   17217   int32x4_t __ret; \
   17218   __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   17219   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   17220   __ret; \
   17221 })
   17222 #endif
   17223 
   17224 #ifdef __LITTLE_ENDIAN__
   17225 __ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   17226   int64x2_t __ret;
   17227   __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
   17228   return __ret;
   17229 }
   17230 #else
   17231 __ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   17232   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17233   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   17234   int64x2_t __ret;
   17235   __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
   17236   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17237   return __ret;
   17238 }
   17239 __ai int64x2_t __noswap_vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   17240   int64x2_t __ret;
   17241   __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
   17242   return __ret;
   17243 }
   17244 #endif
   17245 
   17246 #ifdef __LITTLE_ENDIAN__
   17247 __ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   17248   int32x4_t __ret;
   17249   __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
   17250   return __ret;
   17251 }
   17252 #else
   17253 __ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   17254   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17255   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   17256   int32x4_t __ret;
   17257   __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
   17258   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17259   return __ret;
   17260 }
   17261 __ai int32x4_t __noswap_vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   17262   int32x4_t __ret;
   17263   __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
   17264   return __ret;
   17265 }
   17266 #endif
   17267 
   17268 #ifdef __LITTLE_ENDIAN__
   17269 __ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
   17270   int32x4_t __ret;
   17271   __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   17272   return __ret;
   17273 }
   17274 #else
   17275 __ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
   17276   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17277   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   17278   int32x4_t __ret;
   17279   __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   17280   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17281   return __ret;
   17282 }
   17283 __ai int32x4_t __noswap_vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
   17284   int32x4_t __ret;
   17285   __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   17286   return __ret;
   17287 }
   17288 #endif
   17289 
   17290 #ifdef __LITTLE_ENDIAN__
   17291 __ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
   17292   int16x8_t __ret;
   17293   __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   17294   return __ret;
   17295 }
   17296 #else
   17297 __ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
   17298   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   17299   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   17300   int16x8_t __ret;
   17301   __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   17302   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   17303   return __ret;
   17304 }
   17305 __ai int16x8_t __noswap_vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
   17306   int16x8_t __ret;
   17307   __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   17308   return __ret;
   17309 }
   17310 #endif
   17311 
   17312 #ifdef __LITTLE_ENDIAN__
   17313 __ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
   17314   int32x2_t __ret;
   17315   __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   17316   return __ret;
   17317 }
   17318 #else
   17319 __ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
   17320   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17321   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   17322   int32x2_t __ret;
   17323   __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   17324   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17325   return __ret;
   17326 }
   17327 __ai int32x2_t __noswap_vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
   17328   int32x2_t __ret;
   17329   __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   17330   return __ret;
   17331 }
   17332 #endif
   17333 
   17334 #ifdef __LITTLE_ENDIAN__
   17335 __ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
   17336   int16x4_t __ret;
   17337   __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   17338   return __ret;
   17339 }
   17340 #else
   17341 __ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
   17342   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17343   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   17344   int16x4_t __ret;
   17345   __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   17346   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17347   return __ret;
   17348 }
   17349 __ai int16x4_t __noswap_vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
   17350   int16x4_t __ret;
   17351   __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   17352   return __ret;
   17353 }
   17354 #endif
   17355 
   17356 #ifdef __LITTLE_ENDIAN__
   17357 #define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   17358   int32x4_t __s0 = __p0; \
   17359   int32x2_t __s1 = __p1; \
   17360   int32x4_t __ret; \
   17361   __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   17362   __ret; \
   17363 })
   17364 #else
   17365 #define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   17366   int32x4_t __s0 = __p0; \
   17367   int32x2_t __s1 = __p1; \
   17368   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   17369   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   17370   int32x4_t __ret; \
   17371   __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   17372   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   17373   __ret; \
   17374 })
   17375 #endif
   17376 
   17377 #ifdef __LITTLE_ENDIAN__
   17378 #define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   17379   int16x8_t __s0 = __p0; \
   17380   int16x4_t __s1 = __p1; \
   17381   int16x8_t __ret; \
   17382   __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
   17383   __ret; \
   17384 })
   17385 #else
   17386 #define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   17387   int16x8_t __s0 = __p0; \
   17388   int16x4_t __s1 = __p1; \
   17389   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   17390   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   17391   int16x8_t __ret; \
   17392   __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
   17393   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   17394   __ret; \
   17395 })
   17396 #endif
   17397 
   17398 #ifdef __LITTLE_ENDIAN__
   17399 #define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   17400   int32x2_t __s0 = __p0; \
   17401   int32x2_t __s1 = __p1; \
   17402   int32x2_t __ret; \
   17403   __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   17404   __ret; \
   17405 })
   17406 #else
   17407 #define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   17408   int32x2_t __s0 = __p0; \
   17409   int32x2_t __s1 = __p1; \
   17410   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   17411   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   17412   int32x2_t __ret; \
   17413   __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   17414   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   17415   __ret; \
   17416 })
   17417 #endif
   17418 
   17419 #ifdef __LITTLE_ENDIAN__
   17420 #define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   17421   int16x4_t __s0 = __p0; \
   17422   int16x4_t __s1 = __p1; \
   17423   int16x4_t __ret; \
   17424   __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   17425   __ret; \
   17426 })
   17427 #else
   17428 #define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   17429   int16x4_t __s0 = __p0; \
   17430   int16x4_t __s1 = __p1; \
   17431   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   17432   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   17433   int16x4_t __ret; \
   17434   __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   17435   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   17436   __ret; \
   17437 })
   17438 #endif
   17439 
   17440 #ifdef __LITTLE_ENDIAN__
   17441 __ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
   17442   int32x4_t __ret;
   17443   __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
   17444   return __ret;
   17445 }
   17446 #else
   17447 __ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
   17448   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17449   int32x4_t __ret;
   17450   __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
   17451   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17452   return __ret;
   17453 }
   17454 #endif
   17455 
   17456 #ifdef __LITTLE_ENDIAN__
   17457 __ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
   17458   int16x8_t __ret;
   17459   __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
   17460   return __ret;
   17461 }
   17462 #else
   17463 __ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
   17464   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   17465   int16x8_t __ret;
   17466   __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
   17467   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   17468   return __ret;
   17469 }
   17470 #endif
   17471 
   17472 #ifdef __LITTLE_ENDIAN__
   17473 __ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
   17474   int32x2_t __ret;
   17475   __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
   17476   return __ret;
   17477 }
   17478 #else
   17479 __ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
   17480   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17481   int32x2_t __ret;
   17482   __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
   17483   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17484   return __ret;
   17485 }
   17486 #endif
   17487 
   17488 #ifdef __LITTLE_ENDIAN__
   17489 __ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
   17490   int16x4_t __ret;
   17491   __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
   17492   return __ret;
   17493 }
   17494 #else
   17495 __ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
   17496   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17497   int16x4_t __ret;
   17498   __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
   17499   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17500   return __ret;
   17501 }
   17502 #endif
   17503 
   17504 #ifdef __LITTLE_ENDIAN__
   17505 __ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
   17506   int64x2_t __ret;
   17507   __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
   17508   return __ret;
   17509 }
   17510 #else
   17511 __ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
   17512   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17513   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   17514   int64x2_t __ret;
   17515   __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
   17516   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17517   return __ret;
   17518 }
   17519 __ai int64x2_t __noswap_vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
   17520   int64x2_t __ret;
   17521   __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
   17522   return __ret;
   17523 }
   17524 #endif
   17525 
   17526 #ifdef __LITTLE_ENDIAN__
   17527 __ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
   17528   int32x4_t __ret;
   17529   __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
   17530   return __ret;
   17531 }
   17532 #else
   17533 __ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
   17534   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17535   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   17536   int32x4_t __ret;
   17537   __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
   17538   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17539   return __ret;
   17540 }
   17541 __ai int32x4_t __noswap_vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
   17542   int32x4_t __ret;
   17543   __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
   17544   return __ret;
   17545 }
   17546 #endif
   17547 
   17548 #ifdef __LITTLE_ENDIAN__
   17549 #define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   17550   int32x2_t __s0 = __p0; \
   17551   int32x2_t __s1 = __p1; \
   17552   int64x2_t __ret; \
   17553   __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   17554   __ret; \
   17555 })
   17556 #else
   17557 #define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   17558   int32x2_t __s0 = __p0; \
   17559   int32x2_t __s1 = __p1; \
   17560   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   17561   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   17562   int64x2_t __ret; \
   17563   __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   17564   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   17565   __ret; \
   17566 })
   17567 #endif
   17568 
   17569 #ifdef __LITTLE_ENDIAN__
   17570 #define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   17571   int16x4_t __s0 = __p0; \
   17572   int16x4_t __s1 = __p1; \
   17573   int32x4_t __ret; \
   17574   __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   17575   __ret; \
   17576 })
   17577 #else
   17578 #define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   17579   int16x4_t __s0 = __p0; \
   17580   int16x4_t __s1 = __p1; \
   17581   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   17582   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   17583   int32x4_t __ret; \
   17584   __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   17585   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   17586   __ret; \
   17587 })
   17588 #endif
   17589 
   17590 #ifdef __LITTLE_ENDIAN__
   17591 __ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
   17592   int64x2_t __ret;
   17593   __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
   17594   return __ret;
   17595 }
   17596 #else
   17597 __ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
   17598   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17599   int64x2_t __ret;
   17600   __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
   17601   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17602   return __ret;
   17603 }
   17604 __ai int64x2_t __noswap_vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
   17605   int64x2_t __ret;
   17606   __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
   17607   return __ret;
   17608 }
   17609 #endif
   17610 
   17611 #ifdef __LITTLE_ENDIAN__
   17612 __ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
   17613   int32x4_t __ret;
   17614   __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
   17615   return __ret;
   17616 }
   17617 #else
   17618 __ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
   17619   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17620   int32x4_t __ret;
   17621   __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
   17622   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17623   return __ret;
   17624 }
   17625 __ai int32x4_t __noswap_vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
   17626   int32x4_t __ret;
   17627   __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
   17628   return __ret;
   17629 }
   17630 #endif
   17631 
   17632 #ifdef __LITTLE_ENDIAN__
   17633 __ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
   17634   uint16x4_t __ret;
   17635   __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
   17636   return __ret;
   17637 }
   17638 #else
   17639 __ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
   17640   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17641   uint16x4_t __ret;
   17642   __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 17);
   17643   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17644   return __ret;
   17645 }
   17646 __ai uint16x4_t __noswap_vqmovn_u32(uint32x4_t __p0) {
   17647   uint16x4_t __ret;
   17648   __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
   17649   return __ret;
   17650 }
   17651 #endif
   17652 
   17653 #ifdef __LITTLE_ENDIAN__
   17654 __ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
   17655   uint32x2_t __ret;
   17656   __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
   17657   return __ret;
   17658 }
   17659 #else
   17660 __ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
   17661   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17662   uint32x2_t __ret;
   17663   __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 18);
   17664   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17665   return __ret;
   17666 }
   17667 __ai uint32x2_t __noswap_vqmovn_u64(uint64x2_t __p0) {
   17668   uint32x2_t __ret;
   17669   __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
   17670   return __ret;
   17671 }
   17672 #endif
   17673 
   17674 #ifdef __LITTLE_ENDIAN__
   17675 __ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
   17676   uint8x8_t __ret;
   17677   __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
   17678   return __ret;
   17679 }
   17680 #else
   17681 __ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
   17682   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   17683   uint8x8_t __ret;
   17684   __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 16);
   17685   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   17686   return __ret;
   17687 }
   17688 __ai uint8x8_t __noswap_vqmovn_u16(uint16x8_t __p0) {
   17689   uint8x8_t __ret;
   17690   __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
   17691   return __ret;
   17692 }
   17693 #endif
   17694 
   17695 #ifdef __LITTLE_ENDIAN__
   17696 __ai int16x4_t vqmovn_s32(int32x4_t __p0) {
   17697   int16x4_t __ret;
   17698   __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
   17699   return __ret;
   17700 }
   17701 #else
   17702 __ai int16x4_t vqmovn_s32(int32x4_t __p0) {
   17703   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17704   int16x4_t __ret;
   17705   __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 1);
   17706   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17707   return __ret;
   17708 }
   17709 __ai int16x4_t __noswap_vqmovn_s32(int32x4_t __p0) {
   17710   int16x4_t __ret;
   17711   __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
   17712   return __ret;
   17713 }
   17714 #endif
   17715 
   17716 #ifdef __LITTLE_ENDIAN__
   17717 __ai int32x2_t vqmovn_s64(int64x2_t __p0) {
   17718   int32x2_t __ret;
   17719   __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
   17720   return __ret;
   17721 }
   17722 #else
   17723 __ai int32x2_t vqmovn_s64(int64x2_t __p0) {
   17724   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17725   int32x2_t __ret;
   17726   __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 2);
   17727   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17728   return __ret;
   17729 }
   17730 __ai int32x2_t __noswap_vqmovn_s64(int64x2_t __p0) {
   17731   int32x2_t __ret;
   17732   __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
   17733   return __ret;
   17734 }
   17735 #endif
   17736 
   17737 #ifdef __LITTLE_ENDIAN__
   17738 __ai int8x8_t vqmovn_s16(int16x8_t __p0) {
   17739   int8x8_t __ret;
   17740   __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
   17741   return __ret;
   17742 }
   17743 #else
   17744 __ai int8x8_t vqmovn_s16(int16x8_t __p0) {
   17745   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   17746   int8x8_t __ret;
   17747   __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 0);
   17748   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   17749   return __ret;
   17750 }
   17751 __ai int8x8_t __noswap_vqmovn_s16(int16x8_t __p0) {
   17752   int8x8_t __ret;
   17753   __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
   17754   return __ret;
   17755 }
   17756 #endif
   17757 
   17758 #ifdef __LITTLE_ENDIAN__
   17759 __ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
   17760   uint16x4_t __ret;
   17761   __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
   17762   return __ret;
   17763 }
   17764 #else
   17765 __ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
   17766   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17767   uint16x4_t __ret;
   17768   __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 17);
   17769   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17770   return __ret;
   17771 }
   17772 __ai uint16x4_t __noswap_vqmovun_s32(int32x4_t __p0) {
   17773   uint16x4_t __ret;
   17774   __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
   17775   return __ret;
   17776 }
   17777 #endif
   17778 
   17779 #ifdef __LITTLE_ENDIAN__
   17780 __ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
   17781   uint32x2_t __ret;
   17782   __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
   17783   return __ret;
   17784 }
   17785 #else
   17786 __ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
   17787   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17788   uint32x2_t __ret;
   17789   __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 18);
   17790   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17791   return __ret;
   17792 }
   17793 __ai uint32x2_t __noswap_vqmovun_s64(int64x2_t __p0) {
   17794   uint32x2_t __ret;
   17795   __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
   17796   return __ret;
   17797 }
   17798 #endif
   17799 
   17800 #ifdef __LITTLE_ENDIAN__
   17801 __ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
   17802   uint8x8_t __ret;
   17803   __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
   17804   return __ret;
   17805 }
   17806 #else
   17807 __ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
   17808   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   17809   uint8x8_t __ret;
   17810   __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 16);
   17811   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   17812   return __ret;
   17813 }
   17814 __ai uint8x8_t __noswap_vqmovun_s16(int16x8_t __p0) {
   17815   uint8x8_t __ret;
   17816   __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
   17817   return __ret;
   17818 }
   17819 #endif
   17820 
   17821 #ifdef __LITTLE_ENDIAN__
   17822 __ai int8x16_t vqnegq_s8(int8x16_t __p0) {
   17823   int8x16_t __ret;
   17824   __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 32);
   17825   return __ret;
   17826 }
   17827 #else
   17828 __ai int8x16_t vqnegq_s8(int8x16_t __p0) {
   17829   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   17830   int8x16_t __ret;
   17831   __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 32);
   17832   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   17833   return __ret;
   17834 }
   17835 #endif
   17836 
   17837 #ifdef __LITTLE_ENDIAN__
   17838 __ai int32x4_t vqnegq_s32(int32x4_t __p0) {
   17839   int32x4_t __ret;
   17840   __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 34);
   17841   return __ret;
   17842 }
   17843 #else
   17844 __ai int32x4_t vqnegq_s32(int32x4_t __p0) {
   17845   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17846   int32x4_t __ret;
   17847   __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 34);
   17848   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17849   return __ret;
   17850 }
   17851 #endif
   17852 
   17853 #ifdef __LITTLE_ENDIAN__
   17854 __ai int16x8_t vqnegq_s16(int16x8_t __p0) {
   17855   int16x8_t __ret;
   17856   __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 33);
   17857   return __ret;
   17858 }
   17859 #else
   17860 __ai int16x8_t vqnegq_s16(int16x8_t __p0) {
   17861   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   17862   int16x8_t __ret;
   17863   __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 33);
   17864   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   17865   return __ret;
   17866 }
   17867 #endif
   17868 
   17869 #ifdef __LITTLE_ENDIAN__
   17870 __ai int8x8_t vqneg_s8(int8x8_t __p0) {
   17871   int8x8_t __ret;
   17872   __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 0);
   17873   return __ret;
   17874 }
   17875 #else
   17876 __ai int8x8_t vqneg_s8(int8x8_t __p0) {
   17877   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   17878   int8x8_t __ret;
   17879   __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 0);
   17880   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   17881   return __ret;
   17882 }
   17883 #endif
   17884 
   17885 #ifdef __LITTLE_ENDIAN__
   17886 __ai int32x2_t vqneg_s32(int32x2_t __p0) {
   17887   int32x2_t __ret;
   17888   __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 2);
   17889   return __ret;
   17890 }
   17891 #else
   17892 __ai int32x2_t vqneg_s32(int32x2_t __p0) {
   17893   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17894   int32x2_t __ret;
   17895   __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 2);
   17896   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17897   return __ret;
   17898 }
   17899 #endif
   17900 
   17901 #ifdef __LITTLE_ENDIAN__
   17902 __ai int16x4_t vqneg_s16(int16x4_t __p0) {
   17903   int16x4_t __ret;
   17904   __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 1);
   17905   return __ret;
   17906 }
   17907 #else
   17908 __ai int16x4_t vqneg_s16(int16x4_t __p0) {
   17909   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17910   int16x4_t __ret;
   17911   __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 1);
   17912   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17913   return __ret;
   17914 }
   17915 #endif
   17916 
   17917 #ifdef __LITTLE_ENDIAN__
   17918 __ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
   17919   int32x4_t __ret;
   17920   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   17921   return __ret;
   17922 }
   17923 #else
   17924 __ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
   17925   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17926   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   17927   int32x4_t __ret;
   17928   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   17929   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17930   return __ret;
   17931 }
   17932 __ai int32x4_t __noswap_vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
   17933   int32x4_t __ret;
   17934   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   17935   return __ret;
   17936 }
   17937 #endif
   17938 
   17939 #ifdef __LITTLE_ENDIAN__
   17940 __ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
   17941   int16x8_t __ret;
   17942   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   17943   return __ret;
   17944 }
   17945 #else
   17946 __ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
   17947   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   17948   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   17949   int16x8_t __ret;
   17950   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   17951   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   17952   return __ret;
   17953 }
   17954 __ai int16x8_t __noswap_vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
   17955   int16x8_t __ret;
   17956   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   17957   return __ret;
   17958 }
   17959 #endif
   17960 
   17961 #ifdef __LITTLE_ENDIAN__
   17962 __ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
   17963   int32x2_t __ret;
   17964   __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   17965   return __ret;
   17966 }
   17967 #else
   17968 __ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
   17969   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   17970   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   17971   int32x2_t __ret;
   17972   __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   17973   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   17974   return __ret;
   17975 }
   17976 __ai int32x2_t __noswap_vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
   17977   int32x2_t __ret;
   17978   __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   17979   return __ret;
   17980 }
   17981 #endif
   17982 
   17983 #ifdef __LITTLE_ENDIAN__
   17984 __ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
   17985   int16x4_t __ret;
   17986   __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   17987   return __ret;
   17988 }
   17989 #else
   17990 __ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
   17991   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   17992   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   17993   int16x4_t __ret;
   17994   __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   17995   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   17996   return __ret;
   17997 }
   17998 __ai int16x4_t __noswap_vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
   17999   int16x4_t __ret;
   18000   __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   18001   return __ret;
   18002 }
   18003 #endif
   18004 
   18005 #ifdef __LITTLE_ENDIAN__
   18006 #define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   18007   int32x4_t __s0 = __p0; \
   18008   int32x2_t __s1 = __p1; \
   18009   int32x4_t __ret; \
   18010   __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   18011   __ret; \
   18012 })
   18013 #else
   18014 #define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   18015   int32x4_t __s0 = __p0; \
   18016   int32x2_t __s1 = __p1; \
   18017   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   18018   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   18019   int32x4_t __ret; \
   18020   __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   18021   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   18022   __ret; \
   18023 })
   18024 #endif
   18025 
   18026 #ifdef __LITTLE_ENDIAN__
   18027 #define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   18028   int16x8_t __s0 = __p0; \
   18029   int16x4_t __s1 = __p1; \
   18030   int16x8_t __ret; \
   18031   __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
   18032   __ret; \
   18033 })
   18034 #else
   18035 #define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   18036   int16x8_t __s0 = __p0; \
   18037   int16x4_t __s1 = __p1; \
   18038   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   18039   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   18040   int16x8_t __ret; \
   18041   __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
   18042   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   18043   __ret; \
   18044 })
   18045 #endif
   18046 
   18047 #ifdef __LITTLE_ENDIAN__
   18048 #define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   18049   int32x2_t __s0 = __p0; \
   18050   int32x2_t __s1 = __p1; \
   18051   int32x2_t __ret; \
   18052   __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   18053   __ret; \
   18054 })
   18055 #else
   18056 #define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   18057   int32x2_t __s0 = __p0; \
   18058   int32x2_t __s1 = __p1; \
   18059   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   18060   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   18061   int32x2_t __ret; \
   18062   __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   18063   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   18064   __ret; \
   18065 })
   18066 #endif
   18067 
   18068 #ifdef __LITTLE_ENDIAN__
   18069 #define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   18070   int16x4_t __s0 = __p0; \
   18071   int16x4_t __s1 = __p1; \
   18072   int16x4_t __ret; \
   18073   __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   18074   __ret; \
   18075 })
   18076 #else
   18077 #define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   18078   int16x4_t __s0 = __p0; \
   18079   int16x4_t __s1 = __p1; \
   18080   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   18081   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   18082   int16x4_t __ret; \
   18083   __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   18084   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   18085   __ret; \
   18086 })
   18087 #endif
   18088 
   18089 #ifdef __LITTLE_ENDIAN__
   18090 __ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
   18091   int32x4_t __ret;
   18092   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
   18093   return __ret;
   18094 }
   18095 #else
   18096 __ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
   18097   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18098   int32x4_t __ret;
   18099   __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
   18100   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18101   return __ret;
   18102 }
   18103 #endif
   18104 
   18105 #ifdef __LITTLE_ENDIAN__
   18106 __ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
   18107   int16x8_t __ret;
   18108   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
   18109   return __ret;
   18110 }
   18111 #else
   18112 __ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
   18113   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   18114   int16x8_t __ret;
   18115   __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
   18116   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   18117   return __ret;
   18118 }
   18119 #endif
   18120 
   18121 #ifdef __LITTLE_ENDIAN__
   18122 __ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
   18123   int32x2_t __ret;
   18124   __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
   18125   return __ret;
   18126 }
   18127 #else
   18128 __ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
   18129   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   18130   int32x2_t __ret;
   18131   __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
   18132   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   18133   return __ret;
   18134 }
   18135 #endif
   18136 
   18137 #ifdef __LITTLE_ENDIAN__
   18138 __ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
   18139   int16x4_t __ret;
   18140   __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
   18141   return __ret;
   18142 }
   18143 #else
   18144 __ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
   18145   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18146   int16x4_t __ret;
   18147   __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
   18148   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18149   return __ret;
   18150 }
   18151 #endif
   18152 
   18153 #ifdef __LITTLE_ENDIAN__
   18154 __ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
   18155   uint8x16_t __ret;
   18156   __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   18157   return __ret;
   18158 }
   18159 #else
   18160 __ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
   18161   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18162   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18163   uint8x16_t __ret;
   18164   __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   18165   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18166   return __ret;
   18167 }
   18168 #endif
   18169 
   18170 #ifdef __LITTLE_ENDIAN__
   18171 __ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
   18172   uint32x4_t __ret;
   18173   __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   18174   return __ret;
   18175 }
   18176 #else
   18177 __ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
   18178   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18179   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   18180   uint32x4_t __ret;
   18181   __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   18182   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18183   return __ret;
   18184 }
   18185 #endif
   18186 
   18187 #ifdef __LITTLE_ENDIAN__
   18188 __ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
   18189   uint64x2_t __ret;
   18190   __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   18191   return __ret;
   18192 }
   18193 #else
   18194 __ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
   18195   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   18196   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   18197   uint64x2_t __ret;
   18198   __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   18199   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   18200   return __ret;
   18201 }
   18202 #endif
   18203 
   18204 #ifdef __LITTLE_ENDIAN__
   18205 __ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
   18206   uint16x8_t __ret;
   18207   __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   18208   return __ret;
   18209 }
   18210 #else
   18211 __ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
   18212   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   18213   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   18214   uint16x8_t __ret;
   18215   __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   18216   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   18217   return __ret;
   18218 }
   18219 #endif
   18220 
   18221 #ifdef __LITTLE_ENDIAN__
   18222 __ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
   18223   int8x16_t __ret;
   18224   __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   18225   return __ret;
   18226 }
   18227 #else
   18228 __ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
   18229   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18230   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18231   int8x16_t __ret;
   18232   __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   18233   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18234   return __ret;
   18235 }
   18236 #endif
   18237 
   18238 #ifdef __LITTLE_ENDIAN__
   18239 __ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
   18240   int32x4_t __ret;
   18241   __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   18242   return __ret;
   18243 }
   18244 #else
   18245 __ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
   18246   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18247   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   18248   int32x4_t __ret;
   18249   __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   18250   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18251   return __ret;
   18252 }
   18253 #endif
   18254 
   18255 #ifdef __LITTLE_ENDIAN__
   18256 __ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
   18257   int64x2_t __ret;
   18258   __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
   18259   return __ret;
   18260 }
   18261 #else
   18262 __ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
   18263   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   18264   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   18265   int64x2_t __ret;
   18266   __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
   18267   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   18268   return __ret;
   18269 }
   18270 #endif
   18271 
   18272 #ifdef __LITTLE_ENDIAN__
   18273 __ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
   18274   int16x8_t __ret;
   18275   __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   18276   return __ret;
   18277 }
   18278 #else
   18279 __ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
   18280   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   18281   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   18282   int16x8_t __ret;
   18283   __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   18284   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   18285   return __ret;
   18286 }
   18287 #endif
   18288 
   18289 #ifdef __LITTLE_ENDIAN__
   18290 __ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
   18291   uint8x8_t __ret;
   18292   __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   18293   return __ret;
   18294 }
   18295 #else
   18296 __ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
   18297   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   18298   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   18299   uint8x8_t __ret;
   18300   __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   18301   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   18302   return __ret;
   18303 }
   18304 #endif
   18305 
   18306 #ifdef __LITTLE_ENDIAN__
   18307 __ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
   18308   uint32x2_t __ret;
   18309   __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   18310   return __ret;
   18311 }
   18312 #else
   18313 __ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
   18314   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   18315   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   18316   uint32x2_t __ret;
   18317   __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   18318   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   18319   return __ret;
   18320 }
   18321 #endif
   18322 
   18323 #ifdef __LITTLE_ENDIAN__
   18324 __ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
   18325   uint64x1_t __ret;
   18326   __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   18327   return __ret;
   18328 }
   18329 #else
   18330 __ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
   18331   uint64x1_t __ret;
   18332   __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   18333   return __ret;
   18334 }
   18335 #endif
   18336 
   18337 #ifdef __LITTLE_ENDIAN__
   18338 __ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
   18339   uint16x4_t __ret;
   18340   __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   18341   return __ret;
   18342 }
   18343 #else
   18344 __ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
   18345   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18346   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   18347   uint16x4_t __ret;
   18348   __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   18349   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18350   return __ret;
   18351 }
   18352 #endif
   18353 
   18354 #ifdef __LITTLE_ENDIAN__
   18355 __ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
   18356   int8x8_t __ret;
   18357   __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   18358   return __ret;
   18359 }
   18360 #else
   18361 __ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
   18362   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   18363   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   18364   int8x8_t __ret;
   18365   __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   18366   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   18367   return __ret;
   18368 }
   18369 #endif
   18370 
   18371 #ifdef __LITTLE_ENDIAN__
   18372 __ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
   18373   int32x2_t __ret;
   18374   __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   18375   return __ret;
   18376 }
   18377 #else
   18378 __ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
   18379   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   18380   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   18381   int32x2_t __ret;
   18382   __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   18383   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   18384   return __ret;
   18385 }
   18386 #endif
   18387 
   18388 #ifdef __LITTLE_ENDIAN__
   18389 __ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
   18390   int64x1_t __ret;
   18391   __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   18392   return __ret;
   18393 }
   18394 #else
   18395 __ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
   18396   int64x1_t __ret;
   18397   __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   18398   return __ret;
   18399 }
   18400 #endif
   18401 
   18402 #ifdef __LITTLE_ENDIAN__
   18403 __ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
   18404   int16x4_t __ret;
   18405   __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   18406   return __ret;
   18407 }
   18408 #else
   18409 __ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
   18410   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18411   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   18412   int16x4_t __ret;
   18413   __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   18414   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18415   return __ret;
   18416 }
   18417 #endif
   18418 
   18419 #ifdef __LITTLE_ENDIAN__
   18420 #define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
   18421   uint32x4_t __s0 = __p0; \
   18422   uint16x4_t __ret; \
   18423   __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
   18424   __ret; \
   18425 })
   18426 #else
   18427 #define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
   18428   uint32x4_t __s0 = __p0; \
   18429   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   18430   uint16x4_t __ret; \
   18431   __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
   18432   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   18433   __ret; \
   18434 })
   18435 #define __noswap_vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
   18436   uint32x4_t __s0 = __p0; \
   18437   uint16x4_t __ret; \
   18438   __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
   18439   __ret; \
   18440 })
   18441 #endif
   18442 
   18443 #ifdef __LITTLE_ENDIAN__
   18444 #define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
   18445   uint64x2_t __s0 = __p0; \
   18446   uint32x2_t __ret; \
   18447   __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
   18448   __ret; \
   18449 })
   18450 #else
   18451 #define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
   18452   uint64x2_t __s0 = __p0; \
   18453   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   18454   uint32x2_t __ret; \
   18455   __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
   18456   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   18457   __ret; \
   18458 })
   18459 #define __noswap_vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
   18460   uint64x2_t __s0 = __p0; \
   18461   uint32x2_t __ret; \
   18462   __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
   18463   __ret; \
   18464 })
   18465 #endif
   18466 
   18467 #ifdef __LITTLE_ENDIAN__
   18468 #define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
   18469   uint16x8_t __s0 = __p0; \
   18470   uint8x8_t __ret; \
   18471   __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
   18472   __ret; \
   18473 })
   18474 #else
   18475 #define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
   18476   uint16x8_t __s0 = __p0; \
   18477   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   18478   uint8x8_t __ret; \
   18479   __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
   18480   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   18481   __ret; \
   18482 })
   18483 #define __noswap_vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
   18484   uint16x8_t __s0 = __p0; \
   18485   uint8x8_t __ret; \
   18486   __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
   18487   __ret; \
   18488 })
   18489 #endif
   18490 
   18491 #ifdef __LITTLE_ENDIAN__
   18492 #define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
   18493   int32x4_t __s0 = __p0; \
   18494   int16x4_t __ret; \
   18495   __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
   18496   __ret; \
   18497 })
   18498 #else
   18499 #define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
   18500   int32x4_t __s0 = __p0; \
   18501   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   18502   int16x4_t __ret; \
   18503   __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
   18504   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   18505   __ret; \
   18506 })
   18507 #define __noswap_vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
   18508   int32x4_t __s0 = __p0; \
   18509   int16x4_t __ret; \
   18510   __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
   18511   __ret; \
   18512 })
   18513 #endif
   18514 
   18515 #ifdef __LITTLE_ENDIAN__
   18516 #define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
   18517   int64x2_t __s0 = __p0; \
   18518   int32x2_t __ret; \
   18519   __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
   18520   __ret; \
   18521 })
   18522 #else
   18523 #define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
   18524   int64x2_t __s0 = __p0; \
   18525   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   18526   int32x2_t __ret; \
   18527   __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
   18528   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   18529   __ret; \
   18530 })
   18531 #define __noswap_vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
   18532   int64x2_t __s0 = __p0; \
   18533   int32x2_t __ret; \
   18534   __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
   18535   __ret; \
   18536 })
   18537 #endif
   18538 
   18539 #ifdef __LITTLE_ENDIAN__
   18540 #define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
   18541   int16x8_t __s0 = __p0; \
   18542   int8x8_t __ret; \
   18543   __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
   18544   __ret; \
   18545 })
   18546 #else
   18547 #define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
   18548   int16x8_t __s0 = __p0; \
   18549   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   18550   int8x8_t __ret; \
   18551   __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
   18552   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   18553   __ret; \
   18554 })
   18555 #define __noswap_vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
   18556   int16x8_t __s0 = __p0; \
   18557   int8x8_t __ret; \
   18558   __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
   18559   __ret; \
   18560 })
   18561 #endif
   18562 
   18563 #ifdef __LITTLE_ENDIAN__
   18564 #define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
   18565   int32x4_t __s0 = __p0; \
   18566   uint16x4_t __ret; \
   18567   __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
   18568   __ret; \
   18569 })
   18570 #else
   18571 #define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
   18572   int32x4_t __s0 = __p0; \
   18573   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   18574   uint16x4_t __ret; \
   18575   __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 17); \
   18576   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   18577   __ret; \
   18578 })
   18579 #define __noswap_vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
   18580   int32x4_t __s0 = __p0; \
   18581   uint16x4_t __ret; \
   18582   __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
   18583   __ret; \
   18584 })
   18585 #endif
   18586 
   18587 #ifdef __LITTLE_ENDIAN__
   18588 #define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
   18589   int64x2_t __s0 = __p0; \
   18590   uint32x2_t __ret; \
   18591   __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
   18592   __ret; \
   18593 })
   18594 #else
   18595 #define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
   18596   int64x2_t __s0 = __p0; \
   18597   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   18598   uint32x2_t __ret; \
   18599   __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 18); \
   18600   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   18601   __ret; \
   18602 })
   18603 #define __noswap_vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
   18604   int64x2_t __s0 = __p0; \
   18605   uint32x2_t __ret; \
   18606   __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
   18607   __ret; \
   18608 })
   18609 #endif
   18610 
   18611 #ifdef __LITTLE_ENDIAN__
   18612 #define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
   18613   int16x8_t __s0 = __p0; \
   18614   uint8x8_t __ret; \
   18615   __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
   18616   __ret; \
   18617 })
   18618 #else
   18619 #define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
   18620   int16x8_t __s0 = __p0; \
   18621   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   18622   uint8x8_t __ret; \
   18623   __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 16); \
   18624   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   18625   __ret; \
   18626 })
   18627 #define __noswap_vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
   18628   int16x8_t __s0 = __p0; \
   18629   uint8x8_t __ret; \
   18630   __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
   18631   __ret; \
   18632 })
   18633 #endif
   18634 
   18635 #ifdef __LITTLE_ENDIAN__
   18636 __ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
   18637   uint8x16_t __ret;
   18638   __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   18639   return __ret;
   18640 }
   18641 #else
   18642 __ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
   18643   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18644   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18645   uint8x16_t __ret;
   18646   __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   18647   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18648   return __ret;
   18649 }
   18650 #endif
   18651 
   18652 #ifdef __LITTLE_ENDIAN__
   18653 __ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
   18654   uint32x4_t __ret;
   18655   __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   18656   return __ret;
   18657 }
   18658 #else
   18659 __ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
   18660   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18661   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   18662   uint32x4_t __ret;
   18663   __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   18664   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18665   return __ret;
   18666 }
   18667 #endif
   18668 
   18669 #ifdef __LITTLE_ENDIAN__
   18670 __ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
   18671   uint64x2_t __ret;
   18672   __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   18673   return __ret;
   18674 }
   18675 #else
   18676 __ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
   18677   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   18678   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   18679   uint64x2_t __ret;
   18680   __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   18681   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   18682   return __ret;
   18683 }
   18684 #endif
   18685 
   18686 #ifdef __LITTLE_ENDIAN__
   18687 __ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
   18688   uint16x8_t __ret;
   18689   __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   18690   return __ret;
   18691 }
   18692 #else
   18693 __ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
   18694   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   18695   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   18696   uint16x8_t __ret;
   18697   __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   18698   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   18699   return __ret;
   18700 }
   18701 #endif
   18702 
   18703 #ifdef __LITTLE_ENDIAN__
   18704 __ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
   18705   int8x16_t __ret;
   18706   __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   18707   return __ret;
   18708 }
   18709 #else
   18710 __ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
   18711   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18712   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18713   int8x16_t __ret;
   18714   __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   18715   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   18716   return __ret;
   18717 }
   18718 #endif
   18719 
   18720 #ifdef __LITTLE_ENDIAN__
   18721 __ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
   18722   int32x4_t __ret;
   18723   __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   18724   return __ret;
   18725 }
   18726 #else
   18727 __ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
   18728   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18729   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   18730   int32x4_t __ret;
   18731   __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   18732   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18733   return __ret;
   18734 }
   18735 #endif
   18736 
   18737 #ifdef __LITTLE_ENDIAN__
   18738 __ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
   18739   int64x2_t __ret;
   18740   __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
   18741   return __ret;
   18742 }
   18743 #else
   18744 __ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
   18745   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   18746   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   18747   int64x2_t __ret;
   18748   __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
   18749   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   18750   return __ret;
   18751 }
   18752 #endif
   18753 
   18754 #ifdef __LITTLE_ENDIAN__
   18755 __ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
   18756   int16x8_t __ret;
   18757   __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   18758   return __ret;
   18759 }
   18760 #else
   18761 __ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
   18762   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   18763   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   18764   int16x8_t __ret;
   18765   __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   18766   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   18767   return __ret;
   18768 }
   18769 #endif
   18770 
   18771 #ifdef __LITTLE_ENDIAN__
   18772 __ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
   18773   uint8x8_t __ret;
   18774   __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   18775   return __ret;
   18776 }
   18777 #else
   18778 __ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
   18779   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   18780   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   18781   uint8x8_t __ret;
   18782   __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   18783   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   18784   return __ret;
   18785 }
   18786 #endif
   18787 
   18788 #ifdef __LITTLE_ENDIAN__
   18789 __ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
   18790   uint32x2_t __ret;
   18791   __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   18792   return __ret;
   18793 }
   18794 #else
   18795 __ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
   18796   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   18797   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   18798   uint32x2_t __ret;
   18799   __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   18800   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   18801   return __ret;
   18802 }
   18803 #endif
   18804 
   18805 #ifdef __LITTLE_ENDIAN__
   18806 __ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
   18807   uint64x1_t __ret;
   18808   __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   18809   return __ret;
   18810 }
   18811 #else
   18812 __ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
   18813   uint64x1_t __ret;
   18814   __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   18815   return __ret;
   18816 }
   18817 #endif
   18818 
   18819 #ifdef __LITTLE_ENDIAN__
   18820 __ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
   18821   uint16x4_t __ret;
   18822   __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   18823   return __ret;
   18824 }
   18825 #else
   18826 __ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
   18827   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18828   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   18829   uint16x4_t __ret;
   18830   __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   18831   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18832   return __ret;
   18833 }
   18834 #endif
   18835 
   18836 #ifdef __LITTLE_ENDIAN__
   18837 __ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
   18838   int8x8_t __ret;
   18839   __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   18840   return __ret;
   18841 }
   18842 #else
   18843 __ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
   18844   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   18845   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   18846   int8x8_t __ret;
   18847   __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   18848   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   18849   return __ret;
   18850 }
   18851 #endif
   18852 
   18853 #ifdef __LITTLE_ENDIAN__
   18854 __ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
   18855   int32x2_t __ret;
   18856   __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   18857   return __ret;
   18858 }
   18859 #else
   18860 __ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
   18861   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   18862   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   18863   int32x2_t __ret;
   18864   __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   18865   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   18866   return __ret;
   18867 }
   18868 #endif
   18869 
   18870 #ifdef __LITTLE_ENDIAN__
   18871 __ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
   18872   int64x1_t __ret;
   18873   __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   18874   return __ret;
   18875 }
   18876 #else
   18877 __ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
   18878   int64x1_t __ret;
   18879   __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   18880   return __ret;
   18881 }
   18882 #endif
   18883 
   18884 #ifdef __LITTLE_ENDIAN__
   18885 __ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
   18886   int16x4_t __ret;
   18887   __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   18888   return __ret;
   18889 }
   18890 #else
   18891 __ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
   18892   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   18893   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   18894   int16x4_t __ret;
   18895   __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   18896   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   18897   return __ret;
   18898 }
   18899 #endif
   18900 
   18901 #ifdef __LITTLE_ENDIAN__
   18902 #define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
   18903   uint8x16_t __s0 = __p0; \
   18904   uint8x16_t __ret; \
   18905   __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 48); \
   18906   __ret; \
   18907 })
   18908 #else
   18909 #define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
   18910   uint8x16_t __s0 = __p0; \
   18911   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   18912   uint8x16_t __ret; \
   18913   __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 48); \
   18914   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   18915   __ret; \
   18916 })
   18917 #endif
   18918 
   18919 #ifdef __LITTLE_ENDIAN__
   18920 #define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
   18921   uint32x4_t __s0 = __p0; \
   18922   uint32x4_t __ret; \
   18923   __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 50); \
   18924   __ret; \
   18925 })
   18926 #else
   18927 #define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
   18928   uint32x4_t __s0 = __p0; \
   18929   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   18930   uint32x4_t __ret; \
   18931   __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 50); \
   18932   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   18933   __ret; \
   18934 })
   18935 #endif
   18936 
   18937 #ifdef __LITTLE_ENDIAN__
   18938 #define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
   18939   uint64x2_t __s0 = __p0; \
   18940   uint64x2_t __ret; \
   18941   __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 51); \
   18942   __ret; \
   18943 })
   18944 #else
   18945 #define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
   18946   uint64x2_t __s0 = __p0; \
   18947   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   18948   uint64x2_t __ret; \
   18949   __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 51); \
   18950   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   18951   __ret; \
   18952 })
   18953 #endif
   18954 
   18955 #ifdef __LITTLE_ENDIAN__
   18956 #define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
   18957   uint16x8_t __s0 = __p0; \
   18958   uint16x8_t __ret; \
   18959   __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 49); \
   18960   __ret; \
   18961 })
   18962 #else
   18963 #define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
   18964   uint16x8_t __s0 = __p0; \
   18965   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   18966   uint16x8_t __ret; \
   18967   __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 49); \
   18968   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   18969   __ret; \
   18970 })
   18971 #endif
   18972 
   18973 #ifdef __LITTLE_ENDIAN__
   18974 #define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
   18975   int8x16_t __s0 = __p0; \
   18976   int8x16_t __ret; \
   18977   __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 32); \
   18978   __ret; \
   18979 })
   18980 #else
   18981 #define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
   18982   int8x16_t __s0 = __p0; \
   18983   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   18984   int8x16_t __ret; \
   18985   __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 32); \
   18986   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   18987   __ret; \
   18988 })
   18989 #endif
   18990 
   18991 #ifdef __LITTLE_ENDIAN__
   18992 #define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
   18993   int32x4_t __s0 = __p0; \
   18994   int32x4_t __ret; \
   18995   __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 34); \
   18996   __ret; \
   18997 })
   18998 #else
   18999 #define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
   19000   int32x4_t __s0 = __p0; \
   19001   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   19002   int32x4_t __ret; \
   19003   __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 34); \
   19004   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   19005   __ret; \
   19006 })
   19007 #endif
   19008 
   19009 #ifdef __LITTLE_ENDIAN__
   19010 #define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
   19011   int64x2_t __s0 = __p0; \
   19012   int64x2_t __ret; \
   19013   __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 35); \
   19014   __ret; \
   19015 })
   19016 #else
   19017 #define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
   19018   int64x2_t __s0 = __p0; \
   19019   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   19020   int64x2_t __ret; \
   19021   __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 35); \
   19022   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   19023   __ret; \
   19024 })
   19025 #endif
   19026 
   19027 #ifdef __LITTLE_ENDIAN__
   19028 #define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
   19029   int16x8_t __s0 = __p0; \
   19030   int16x8_t __ret; \
   19031   __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 33); \
   19032   __ret; \
   19033 })
   19034 #else
   19035 #define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
   19036   int16x8_t __s0 = __p0; \
   19037   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   19038   int16x8_t __ret; \
   19039   __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 33); \
   19040   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   19041   __ret; \
   19042 })
   19043 #endif
   19044 
   19045 #ifdef __LITTLE_ENDIAN__
   19046 #define vqshl_n_u8(__p0, __p1) __extension__ ({ \
   19047   uint8x8_t __s0 = __p0; \
   19048   uint8x8_t __ret; \
   19049   __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 16); \
   19050   __ret; \
   19051 })
   19052 #else
   19053 #define vqshl_n_u8(__p0, __p1) __extension__ ({ \
   19054   uint8x8_t __s0 = __p0; \
   19055   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   19056   uint8x8_t __ret; \
   19057   __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 16); \
   19058   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   19059   __ret; \
   19060 })
   19061 #endif
   19062 
   19063 #ifdef __LITTLE_ENDIAN__
   19064 #define vqshl_n_u32(__p0, __p1) __extension__ ({ \
   19065   uint32x2_t __s0 = __p0; \
   19066   uint32x2_t __ret; \
   19067   __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 18); \
   19068   __ret; \
   19069 })
   19070 #else
   19071 #define vqshl_n_u32(__p0, __p1) __extension__ ({ \
   19072   uint32x2_t __s0 = __p0; \
   19073   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   19074   uint32x2_t __ret; \
   19075   __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 18); \
   19076   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   19077   __ret; \
   19078 })
   19079 #endif
   19080 
   19081 #ifdef __LITTLE_ENDIAN__
   19082 #define vqshl_n_u64(__p0, __p1) __extension__ ({ \
   19083   uint64x1_t __s0 = __p0; \
   19084   uint64x1_t __ret; \
   19085   __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
   19086   __ret; \
   19087 })
   19088 #else
   19089 #define vqshl_n_u64(__p0, __p1) __extension__ ({ \
   19090   uint64x1_t __s0 = __p0; \
   19091   uint64x1_t __ret; \
   19092   __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
   19093   __ret; \
   19094 })
   19095 #endif
   19096 
   19097 #ifdef __LITTLE_ENDIAN__
   19098 #define vqshl_n_u16(__p0, __p1) __extension__ ({ \
   19099   uint16x4_t __s0 = __p0; \
   19100   uint16x4_t __ret; \
   19101   __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 17); \
   19102   __ret; \
   19103 })
   19104 #else
   19105 #define vqshl_n_u16(__p0, __p1) __extension__ ({ \
   19106   uint16x4_t __s0 = __p0; \
   19107   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   19108   uint16x4_t __ret; \
   19109   __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 17); \
   19110   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   19111   __ret; \
   19112 })
   19113 #endif
   19114 
   19115 #ifdef __LITTLE_ENDIAN__
   19116 #define vqshl_n_s8(__p0, __p1) __extension__ ({ \
   19117   int8x8_t __s0 = __p0; \
   19118   int8x8_t __ret; \
   19119   __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 0); \
   19120   __ret; \
   19121 })
   19122 #else
   19123 #define vqshl_n_s8(__p0, __p1) __extension__ ({ \
   19124   int8x8_t __s0 = __p0; \
   19125   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   19126   int8x8_t __ret; \
   19127   __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 0); \
   19128   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   19129   __ret; \
   19130 })
   19131 #endif
   19132 
   19133 #ifdef __LITTLE_ENDIAN__
   19134 #define vqshl_n_s32(__p0, __p1) __extension__ ({ \
   19135   int32x2_t __s0 = __p0; \
   19136   int32x2_t __ret; \
   19137   __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 2); \
   19138   __ret; \
   19139 })
   19140 #else
   19141 #define vqshl_n_s32(__p0, __p1) __extension__ ({ \
   19142   int32x2_t __s0 = __p0; \
   19143   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   19144   int32x2_t __ret; \
   19145   __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 2); \
   19146   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   19147   __ret; \
   19148 })
   19149 #endif
   19150 
   19151 #ifdef __LITTLE_ENDIAN__
   19152 #define vqshl_n_s64(__p0, __p1) __extension__ ({ \
   19153   int64x1_t __s0 = __p0; \
   19154   int64x1_t __ret; \
   19155   __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
   19156   __ret; \
   19157 })
   19158 #else
   19159 #define vqshl_n_s64(__p0, __p1) __extension__ ({ \
   19160   int64x1_t __s0 = __p0; \
   19161   int64x1_t __ret; \
   19162   __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
   19163   __ret; \
   19164 })
   19165 #endif
   19166 
   19167 #ifdef __LITTLE_ENDIAN__
   19168 #define vqshl_n_s16(__p0, __p1) __extension__ ({ \
   19169   int16x4_t __s0 = __p0; \
   19170   int16x4_t __ret; \
   19171   __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 1); \
   19172   __ret; \
   19173 })
   19174 #else
   19175 #define vqshl_n_s16(__p0, __p1) __extension__ ({ \
   19176   int16x4_t __s0 = __p0; \
   19177   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   19178   int16x4_t __ret; \
   19179   __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 1); \
   19180   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   19181   __ret; \
   19182 })
   19183 #endif
   19184 
   19185 #ifdef __LITTLE_ENDIAN__
   19186 #define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
   19187   int8x16_t __s0 = __p0; \
   19188   uint8x16_t __ret; \
   19189   __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 48); \
   19190   __ret; \
   19191 })
   19192 #else
   19193 #define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
   19194   int8x16_t __s0 = __p0; \
   19195   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   19196   uint8x16_t __ret; \
   19197   __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 48); \
   19198   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   19199   __ret; \
   19200 })
   19201 #endif
   19202 
   19203 #ifdef __LITTLE_ENDIAN__
   19204 #define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
   19205   int32x4_t __s0 = __p0; \
   19206   uint32x4_t __ret; \
   19207   __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 50); \
   19208   __ret; \
   19209 })
   19210 #else
   19211 #define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
   19212   int32x4_t __s0 = __p0; \
   19213   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   19214   uint32x4_t __ret; \
   19215   __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 50); \
   19216   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   19217   __ret; \
   19218 })
   19219 #endif
   19220 
   19221 #ifdef __LITTLE_ENDIAN__
   19222 #define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
   19223   int64x2_t __s0 = __p0; \
   19224   uint64x2_t __ret; \
   19225   __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 51); \
   19226   __ret; \
   19227 })
   19228 #else
   19229 #define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
   19230   int64x2_t __s0 = __p0; \
   19231   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   19232   uint64x2_t __ret; \
   19233   __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 51); \
   19234   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   19235   __ret; \
   19236 })
   19237 #endif
   19238 
   19239 #ifdef __LITTLE_ENDIAN__
   19240 #define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
   19241   int16x8_t __s0 = __p0; \
   19242   uint16x8_t __ret; \
   19243   __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 49); \
   19244   __ret; \
   19245 })
   19246 #else
   19247 #define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
   19248   int16x8_t __s0 = __p0; \
   19249   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   19250   uint16x8_t __ret; \
   19251   __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 49); \
   19252   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   19253   __ret; \
   19254 })
   19255 #endif
   19256 
   19257 #ifdef __LITTLE_ENDIAN__
   19258 #define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
   19259   int8x8_t __s0 = __p0; \
   19260   uint8x8_t __ret; \
   19261   __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 16); \
   19262   __ret; \
   19263 })
   19264 #else
   19265 #define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
   19266   int8x8_t __s0 = __p0; \
   19267   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   19268   uint8x8_t __ret; \
   19269   __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 16); \
   19270   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   19271   __ret; \
   19272 })
   19273 #endif
   19274 
   19275 #ifdef __LITTLE_ENDIAN__
   19276 #define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
   19277   int32x2_t __s0 = __p0; \
   19278   uint32x2_t __ret; \
   19279   __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 18); \
   19280   __ret; \
   19281 })
   19282 #else
   19283 #define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
   19284   int32x2_t __s0 = __p0; \
   19285   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   19286   uint32x2_t __ret; \
   19287   __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 18); \
   19288   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   19289   __ret; \
   19290 })
   19291 #endif
   19292 
   19293 #ifdef __LITTLE_ENDIAN__
   19294 #define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
   19295   int64x1_t __s0 = __p0; \
   19296   uint64x1_t __ret; \
   19297   __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
   19298   __ret; \
   19299 })
   19300 #else
   19301 #define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
   19302   int64x1_t __s0 = __p0; \
   19303   uint64x1_t __ret; \
   19304   __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
   19305   __ret; \
   19306 })
   19307 #endif
   19308 
   19309 #ifdef __LITTLE_ENDIAN__
   19310 #define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
   19311   int16x4_t __s0 = __p0; \
   19312   uint16x4_t __ret; \
   19313   __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 17); \
   19314   __ret; \
   19315 })
   19316 #else
   19317 #define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
   19318   int16x4_t __s0 = __p0; \
   19319   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   19320   uint16x4_t __ret; \
   19321   __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 17); \
   19322   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   19323   __ret; \
   19324 })
   19325 #endif
   19326 
   19327 #ifdef __LITTLE_ENDIAN__
   19328 #define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
   19329   uint32x4_t __s0 = __p0; \
   19330   uint16x4_t __ret; \
   19331   __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
   19332   __ret; \
   19333 })
   19334 #else
   19335 #define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
   19336   uint32x4_t __s0 = __p0; \
   19337   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   19338   uint16x4_t __ret; \
   19339   __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 17); \
   19340   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   19341   __ret; \
   19342 })
   19343 #define __noswap_vqshrn_n_u32(__p0, __p1) __extension__ ({ \
   19344   uint32x4_t __s0 = __p0; \
   19345   uint16x4_t __ret; \
   19346   __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
   19347   __ret; \
   19348 })
   19349 #endif
   19350 
   19351 #ifdef __LITTLE_ENDIAN__
   19352 #define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
   19353   uint64x2_t __s0 = __p0; \
   19354   uint32x2_t __ret; \
   19355   __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
   19356   __ret; \
   19357 })
   19358 #else
   19359 #define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
   19360   uint64x2_t __s0 = __p0; \
   19361   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   19362   uint32x2_t __ret; \
   19363   __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 18); \
   19364   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   19365   __ret; \
   19366 })
   19367 #define __noswap_vqshrn_n_u64(__p0, __p1) __extension__ ({ \
   19368   uint64x2_t __s0 = __p0; \
   19369   uint32x2_t __ret; \
   19370   __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
   19371   __ret; \
   19372 })
   19373 #endif
   19374 
   19375 #ifdef __LITTLE_ENDIAN__
   19376 #define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
   19377   uint16x8_t __s0 = __p0; \
   19378   uint8x8_t __ret; \
   19379   __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
   19380   __ret; \
   19381 })
   19382 #else
   19383 #define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
   19384   uint16x8_t __s0 = __p0; \
   19385   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   19386   uint8x8_t __ret; \
   19387   __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 16); \
   19388   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   19389   __ret; \
   19390 })
   19391 #define __noswap_vqshrn_n_u16(__p0, __p1) __extension__ ({ \
   19392   uint16x8_t __s0 = __p0; \
   19393   uint8x8_t __ret; \
   19394   __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
   19395   __ret; \
   19396 })
   19397 #endif
   19398 
   19399 #ifdef __LITTLE_ENDIAN__
   19400 #define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
   19401   int32x4_t __s0 = __p0; \
   19402   int16x4_t __ret; \
   19403   __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
   19404   __ret; \
   19405 })
   19406 #else
   19407 #define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
   19408   int32x4_t __s0 = __p0; \
   19409   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   19410   int16x4_t __ret; \
   19411   __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 1); \
   19412   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   19413   __ret; \
   19414 })
   19415 #define __noswap_vqshrn_n_s32(__p0, __p1) __extension__ ({ \
   19416   int32x4_t __s0 = __p0; \
   19417   int16x4_t __ret; \
   19418   __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
   19419   __ret; \
   19420 })
   19421 #endif
   19422 
   19423 #ifdef __LITTLE_ENDIAN__
   19424 #define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
   19425   int64x2_t __s0 = __p0; \
   19426   int32x2_t __ret; \
   19427   __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
   19428   __ret; \
   19429 })
   19430 #else
   19431 #define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
   19432   int64x2_t __s0 = __p0; \
   19433   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   19434   int32x2_t __ret; \
   19435   __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 2); \
   19436   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   19437   __ret; \
   19438 })
   19439 #define __noswap_vqshrn_n_s64(__p0, __p1) __extension__ ({ \
   19440   int64x2_t __s0 = __p0; \
   19441   int32x2_t __ret; \
   19442   __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
   19443   __ret; \
   19444 })
   19445 #endif
   19446 
   19447 #ifdef __LITTLE_ENDIAN__
   19448 #define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
   19449   int16x8_t __s0 = __p0; \
   19450   int8x8_t __ret; \
   19451   __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
   19452   __ret; \
   19453 })
   19454 #else
   19455 #define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
   19456   int16x8_t __s0 = __p0; \
   19457   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   19458   int8x8_t __ret; \
   19459   __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 0); \
   19460   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   19461   __ret; \
   19462 })
   19463 #define __noswap_vqshrn_n_s16(__p0, __p1) __extension__ ({ \
   19464   int16x8_t __s0 = __p0; \
   19465   int8x8_t __ret; \
   19466   __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
   19467   __ret; \
   19468 })
   19469 #endif
   19470 
   19471 #ifdef __LITTLE_ENDIAN__
   19472 #define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
   19473   int32x4_t __s0 = __p0; \
   19474   uint16x4_t __ret; \
   19475   __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
   19476   __ret; \
   19477 })
   19478 #else
   19479 #define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
   19480   int32x4_t __s0 = __p0; \
   19481   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   19482   uint16x4_t __ret; \
   19483   __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 17); \
   19484   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   19485   __ret; \
   19486 })
   19487 #define __noswap_vqshrun_n_s32(__p0, __p1) __extension__ ({ \
   19488   int32x4_t __s0 = __p0; \
   19489   uint16x4_t __ret; \
   19490   __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
   19491   __ret; \
   19492 })
   19493 #endif
   19494 
   19495 #ifdef __LITTLE_ENDIAN__
   19496 #define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
   19497   int64x2_t __s0 = __p0; \
   19498   uint32x2_t __ret; \
   19499   __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
   19500   __ret; \
   19501 })
   19502 #else
   19503 #define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
   19504   int64x2_t __s0 = __p0; \
   19505   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   19506   uint32x2_t __ret; \
   19507   __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 18); \
   19508   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   19509   __ret; \
   19510 })
   19511 #define __noswap_vqshrun_n_s64(__p0, __p1) __extension__ ({ \
   19512   int64x2_t __s0 = __p0; \
   19513   uint32x2_t __ret; \
   19514   __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
   19515   __ret; \
   19516 })
   19517 #endif
   19518 
   19519 #ifdef __LITTLE_ENDIAN__
   19520 #define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
   19521   int16x8_t __s0 = __p0; \
   19522   uint8x8_t __ret; \
   19523   __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
   19524   __ret; \
   19525 })
   19526 #else
   19527 #define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
   19528   int16x8_t __s0 = __p0; \
   19529   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   19530   uint8x8_t __ret; \
   19531   __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 16); \
   19532   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   19533   __ret; \
   19534 })
   19535 #define __noswap_vqshrun_n_s16(__p0, __p1) __extension__ ({ \
   19536   int16x8_t __s0 = __p0; \
   19537   uint8x8_t __ret; \
   19538   __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
   19539   __ret; \
   19540 })
   19541 #endif
   19542 
   19543 #ifdef __LITTLE_ENDIAN__
   19544 __ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   19545   uint8x16_t __ret;
   19546   __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   19547   return __ret;
   19548 }
   19549 #else
   19550 __ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   19551   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   19552   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   19553   uint8x16_t __ret;
   19554   __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   19555   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   19556   return __ret;
   19557 }
   19558 #endif
   19559 
   19560 #ifdef __LITTLE_ENDIAN__
   19561 __ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   19562   uint32x4_t __ret;
   19563   __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   19564   return __ret;
   19565 }
   19566 #else
   19567 __ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   19568   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   19569   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   19570   uint32x4_t __ret;
   19571   __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   19572   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   19573   return __ret;
   19574 }
   19575 #endif
   19576 
   19577 #ifdef __LITTLE_ENDIAN__
   19578 __ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   19579   uint64x2_t __ret;
   19580   __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   19581   return __ret;
   19582 }
   19583 #else
   19584 __ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   19585   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   19586   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   19587   uint64x2_t __ret;
   19588   __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   19589   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   19590   return __ret;
   19591 }
   19592 #endif
   19593 
   19594 #ifdef __LITTLE_ENDIAN__
   19595 __ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   19596   uint16x8_t __ret;
   19597   __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   19598   return __ret;
   19599 }
   19600 #else
   19601 __ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   19602   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   19603   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   19604   uint16x8_t __ret;
   19605   __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   19606   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   19607   return __ret;
   19608 }
   19609 #endif
   19610 
   19611 #ifdef __LITTLE_ENDIAN__
   19612 __ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
   19613   int8x16_t __ret;
   19614   __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   19615   return __ret;
   19616 }
   19617 #else
   19618 __ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
   19619   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   19620   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   19621   int8x16_t __ret;
   19622   __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   19623   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   19624   return __ret;
   19625 }
   19626 #endif
   19627 
   19628 #ifdef __LITTLE_ENDIAN__
   19629 __ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
   19630   int32x4_t __ret;
   19631   __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   19632   return __ret;
   19633 }
   19634 #else
   19635 __ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
   19636   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   19637   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   19638   int32x4_t __ret;
   19639   __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   19640   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   19641   return __ret;
   19642 }
   19643 __ai int32x4_t __noswap_vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
   19644   int32x4_t __ret;
   19645   __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   19646   return __ret;
   19647 }
   19648 #endif
   19649 
   19650 #ifdef __LITTLE_ENDIAN__
   19651 __ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
   19652   int64x2_t __ret;
   19653   __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
   19654   return __ret;
   19655 }
   19656 #else
   19657 __ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
   19658   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   19659   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   19660   int64x2_t __ret;
   19661   __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
   19662   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   19663   return __ret;
   19664 }
   19665 #endif
   19666 
   19667 #ifdef __LITTLE_ENDIAN__
   19668 __ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
   19669   int16x8_t __ret;
   19670   __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   19671   return __ret;
   19672 }
   19673 #else
   19674 __ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
   19675   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   19676   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   19677   int16x8_t __ret;
   19678   __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   19679   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   19680   return __ret;
   19681 }
   19682 __ai int16x8_t __noswap_vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
   19683   int16x8_t __ret;
   19684   __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   19685   return __ret;
   19686 }
   19687 #endif
   19688 
   19689 #ifdef __LITTLE_ENDIAN__
   19690 __ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
   19691   uint8x8_t __ret;
   19692   __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   19693   return __ret;
   19694 }
   19695 #else
   19696 __ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
   19697   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   19698   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   19699   uint8x8_t __ret;
   19700   __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   19701   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   19702   return __ret;
   19703 }
   19704 #endif
   19705 
   19706 #ifdef __LITTLE_ENDIAN__
   19707 __ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
   19708   uint32x2_t __ret;
   19709   __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   19710   return __ret;
   19711 }
   19712 #else
   19713 __ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
   19714   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   19715   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   19716   uint32x2_t __ret;
   19717   __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   19718   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   19719   return __ret;
   19720 }
   19721 #endif
   19722 
   19723 #ifdef __LITTLE_ENDIAN__
   19724 __ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
   19725   uint64x1_t __ret;
   19726   __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   19727   return __ret;
   19728 }
   19729 #else
   19730 __ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
   19731   uint64x1_t __ret;
   19732   __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   19733   return __ret;
   19734 }
   19735 #endif
   19736 
   19737 #ifdef __LITTLE_ENDIAN__
   19738 __ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
   19739   uint16x4_t __ret;
   19740   __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   19741   return __ret;
   19742 }
   19743 #else
   19744 __ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
   19745   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   19746   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   19747   uint16x4_t __ret;
   19748   __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   19749   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   19750   return __ret;
   19751 }
   19752 #endif
   19753 
   19754 #ifdef __LITTLE_ENDIAN__
   19755 __ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
   19756   int8x8_t __ret;
   19757   __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   19758   return __ret;
   19759 }
   19760 #else
   19761 __ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
   19762   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   19763   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   19764   int8x8_t __ret;
   19765   __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   19766   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   19767   return __ret;
   19768 }
   19769 #endif
   19770 
   19771 #ifdef __LITTLE_ENDIAN__
   19772 __ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
   19773   int32x2_t __ret;
   19774   __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   19775   return __ret;
   19776 }
   19777 #else
   19778 __ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
   19779   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   19780   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   19781   int32x2_t __ret;
   19782   __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   19783   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   19784   return __ret;
   19785 }
   19786 __ai int32x2_t __noswap_vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
   19787   int32x2_t __ret;
   19788   __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   19789   return __ret;
   19790 }
   19791 #endif
   19792 
   19793 #ifdef __LITTLE_ENDIAN__
   19794 __ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
   19795   int64x1_t __ret;
   19796   __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   19797   return __ret;
   19798 }
   19799 #else
   19800 __ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
   19801   int64x1_t __ret;
   19802   __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   19803   return __ret;
   19804 }
   19805 #endif
   19806 
   19807 #ifdef __LITTLE_ENDIAN__
   19808 __ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
   19809   int16x4_t __ret;
   19810   __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   19811   return __ret;
   19812 }
   19813 #else
   19814 __ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
   19815   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   19816   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   19817   int16x4_t __ret;
   19818   __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   19819   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   19820   return __ret;
   19821 }
   19822 __ai int16x4_t __noswap_vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
   19823   int16x4_t __ret;
   19824   __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   19825   return __ret;
   19826 }
   19827 #endif
   19828 
   19829 #ifdef __LITTLE_ENDIAN__
   19830 __ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   19831   uint16x4_t __ret;
   19832   __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
   19833   return __ret;
   19834 }
   19835 #else
   19836 __ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   19837   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   19838   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   19839   uint16x4_t __ret;
   19840   __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
   19841   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   19842   return __ret;
   19843 }
   19844 __ai uint16x4_t __noswap_vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   19845   uint16x4_t __ret;
   19846   __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
   19847   return __ret;
   19848 }
   19849 #endif
   19850 
   19851 #ifdef __LITTLE_ENDIAN__
   19852 __ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   19853   uint32x2_t __ret;
   19854   __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
   19855   return __ret;
   19856 }
   19857 #else
   19858 __ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   19859   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   19860   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   19861   uint32x2_t __ret;
   19862   __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
   19863   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   19864   return __ret;
   19865 }
   19866 __ai uint32x2_t __noswap_vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   19867   uint32x2_t __ret;
   19868   __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
   19869   return __ret;
   19870 }
   19871 #endif
   19872 
   19873 #ifdef __LITTLE_ENDIAN__
   19874 __ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   19875   uint8x8_t __ret;
   19876   __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
   19877   return __ret;
   19878 }
   19879 #else
   19880 __ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   19881   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   19882   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   19883   uint8x8_t __ret;
   19884   __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
   19885   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   19886   return __ret;
   19887 }
   19888 __ai uint8x8_t __noswap_vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   19889   uint8x8_t __ret;
   19890   __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
   19891   return __ret;
   19892 }
   19893 #endif
   19894 
   19895 #ifdef __LITTLE_ENDIAN__
   19896 __ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
   19897   int16x4_t __ret;
   19898   __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
   19899   return __ret;
   19900 }
   19901 #else
   19902 __ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
   19903   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   19904   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   19905   int16x4_t __ret;
   19906   __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
   19907   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   19908   return __ret;
   19909 }
   19910 __ai int16x4_t __noswap_vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
   19911   int16x4_t __ret;
   19912   __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
   19913   return __ret;
   19914 }
   19915 #endif
   19916 
   19917 #ifdef __LITTLE_ENDIAN__
   19918 __ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
   19919   int32x2_t __ret;
   19920   __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
   19921   return __ret;
   19922 }
   19923 #else
   19924 __ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
   19925   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   19926   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   19927   int32x2_t __ret;
   19928   __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
   19929   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   19930   return __ret;
   19931 }
   19932 __ai int32x2_t __noswap_vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
   19933   int32x2_t __ret;
   19934   __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
   19935   return __ret;
   19936 }
   19937 #endif
   19938 
   19939 #ifdef __LITTLE_ENDIAN__
   19940 __ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
   19941   int8x8_t __ret;
   19942   __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
   19943   return __ret;
   19944 }
   19945 #else
   19946 __ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
   19947   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   19948   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   19949   int8x8_t __ret;
   19950   __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
   19951   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   19952   return __ret;
   19953 }
   19954 __ai int8x8_t __noswap_vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
   19955   int8x8_t __ret;
   19956   __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
   19957   return __ret;
   19958 }
   19959 #endif
   19960 
   19961 #ifdef __LITTLE_ENDIAN__
   19962 __ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
   19963   uint32x4_t __ret;
   19964   __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 50);
   19965   return __ret;
   19966 }
   19967 #else
   19968 __ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
   19969   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   19970   uint32x4_t __ret;
   19971   __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 50);
   19972   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   19973   return __ret;
   19974 }
   19975 #endif
   19976 
   19977 #ifdef __LITTLE_ENDIAN__
   19978 __ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
   19979   float32x4_t __ret;
   19980   __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 41);
   19981   return __ret;
   19982 }
   19983 #else
   19984 __ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
   19985   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   19986   float32x4_t __ret;
   19987   __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 41);
   19988   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   19989   return __ret;
   19990 }
   19991 #endif
   19992 
   19993 #ifdef __LITTLE_ENDIAN__
   19994 __ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
   19995   uint32x2_t __ret;
   19996   __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 18);
   19997   return __ret;
   19998 }
   19999 #else
   20000 __ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
   20001   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20002   uint32x2_t __ret;
   20003   __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 18);
   20004   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20005   return __ret;
   20006 }
   20007 #endif
   20008 
   20009 #ifdef __LITTLE_ENDIAN__
   20010 __ai float32x2_t vrecpe_f32(float32x2_t __p0) {
   20011   float32x2_t __ret;
   20012   __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 9);
   20013   return __ret;
   20014 }
   20015 #else
   20016 __ai float32x2_t vrecpe_f32(float32x2_t __p0) {
   20017   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20018   float32x2_t __ret;
   20019   __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 9);
   20020   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20021   return __ret;
   20022 }
   20023 #endif
   20024 
   20025 #ifdef __LITTLE_ENDIAN__
   20026 __ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
   20027   float32x4_t __ret;
   20028   __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   20029   return __ret;
   20030 }
   20031 #else
   20032 __ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
   20033   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20034   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   20035   float32x4_t __ret;
   20036   __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   20037   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20038   return __ret;
   20039 }
   20040 #endif
   20041 
   20042 #ifdef __LITTLE_ENDIAN__
   20043 __ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
   20044   float32x2_t __ret;
   20045   __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   20046   return __ret;
   20047 }
   20048 #else
   20049 __ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
   20050   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20051   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   20052   float32x2_t __ret;
   20053   __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   20054   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20055   return __ret;
   20056 }
   20057 #endif
   20058 
   20059 #ifdef __LITTLE_ENDIAN__
   20060 __ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
   20061   poly8x8_t __ret;
   20062   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
   20063   return __ret;
   20064 }
   20065 #else
   20066 __ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
   20067   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20068   poly8x8_t __ret;
   20069   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
   20070   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20071   return __ret;
   20072 }
   20073 #endif
   20074 
   20075 #ifdef __LITTLE_ENDIAN__
   20076 __ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
   20077   poly8x16_t __ret;
   20078   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
   20079   return __ret;
   20080 }
   20081 #else
   20082 __ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
   20083   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20084   poly8x16_t __ret;
   20085   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
   20086   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20087   return __ret;
   20088 }
   20089 #endif
   20090 
   20091 #ifdef __LITTLE_ENDIAN__
   20092 __ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
   20093   uint8x16_t __ret;
   20094   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
   20095   return __ret;
   20096 }
   20097 #else
   20098 __ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
   20099   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20100   uint8x16_t __ret;
   20101   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
   20102   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20103   return __ret;
   20104 }
   20105 #endif
   20106 
   20107 #ifdef __LITTLE_ENDIAN__
   20108 __ai int8x16_t vrev16q_s8(int8x16_t __p0) {
   20109   int8x16_t __ret;
   20110   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
   20111   return __ret;
   20112 }
   20113 #else
   20114 __ai int8x16_t vrev16q_s8(int8x16_t __p0) {
   20115   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20116   int8x16_t __ret;
   20117   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
   20118   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20119   return __ret;
   20120 }
   20121 #endif
   20122 
   20123 #ifdef __LITTLE_ENDIAN__
   20124 __ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
   20125   uint8x8_t __ret;
   20126   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
   20127   return __ret;
   20128 }
   20129 #else
   20130 __ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
   20131   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20132   uint8x8_t __ret;
   20133   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
   20134   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20135   return __ret;
   20136 }
   20137 #endif
   20138 
   20139 #ifdef __LITTLE_ENDIAN__
   20140 __ai int8x8_t vrev16_s8(int8x8_t __p0) {
   20141   int8x8_t __ret;
   20142   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
   20143   return __ret;
   20144 }
   20145 #else
   20146 __ai int8x8_t vrev16_s8(int8x8_t __p0) {
   20147   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20148   int8x8_t __ret;
   20149   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
   20150   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20151   return __ret;
   20152 }
   20153 #endif
   20154 
   20155 #ifdef __LITTLE_ENDIAN__
   20156 __ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
   20157   poly8x8_t __ret;
   20158   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
   20159   return __ret;
   20160 }
   20161 #else
   20162 __ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
   20163   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20164   poly8x8_t __ret;
   20165   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
   20166   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20167   return __ret;
   20168 }
   20169 #endif
   20170 
   20171 #ifdef __LITTLE_ENDIAN__
   20172 __ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
   20173   poly16x4_t __ret;
   20174   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
   20175   return __ret;
   20176 }
   20177 #else
   20178 __ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
   20179   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20180   poly16x4_t __ret;
   20181   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
   20182   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20183   return __ret;
   20184 }
   20185 #endif
   20186 
   20187 #ifdef __LITTLE_ENDIAN__
   20188 __ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
   20189   poly8x16_t __ret;
   20190   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
   20191   return __ret;
   20192 }
   20193 #else
   20194 __ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
   20195   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20196   poly8x16_t __ret;
   20197   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
   20198   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20199   return __ret;
   20200 }
   20201 #endif
   20202 
   20203 #ifdef __LITTLE_ENDIAN__
   20204 __ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
   20205   poly16x8_t __ret;
   20206   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
   20207   return __ret;
   20208 }
   20209 #else
   20210 __ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
   20211   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20212   poly16x8_t __ret;
   20213   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
   20214   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20215   return __ret;
   20216 }
   20217 #endif
   20218 
   20219 #ifdef __LITTLE_ENDIAN__
   20220 __ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
   20221   uint8x16_t __ret;
   20222   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
   20223   return __ret;
   20224 }
   20225 #else
   20226 __ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
   20227   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20228   uint8x16_t __ret;
   20229   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
   20230   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20231   return __ret;
   20232 }
   20233 #endif
   20234 
   20235 #ifdef __LITTLE_ENDIAN__
   20236 __ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
   20237   uint16x8_t __ret;
   20238   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
   20239   return __ret;
   20240 }
   20241 #else
   20242 __ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
   20243   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20244   uint16x8_t __ret;
   20245   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
   20246   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20247   return __ret;
   20248 }
   20249 #endif
   20250 
   20251 #ifdef __LITTLE_ENDIAN__
   20252 __ai int8x16_t vrev32q_s8(int8x16_t __p0) {
   20253   int8x16_t __ret;
   20254   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
   20255   return __ret;
   20256 }
   20257 #else
   20258 __ai int8x16_t vrev32q_s8(int8x16_t __p0) {
   20259   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20260   int8x16_t __ret;
   20261   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
   20262   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20263   return __ret;
   20264 }
   20265 #endif
   20266 
   20267 #ifdef __LITTLE_ENDIAN__
   20268 __ai int16x8_t vrev32q_s16(int16x8_t __p0) {
   20269   int16x8_t __ret;
   20270   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
   20271   return __ret;
   20272 }
   20273 #else
   20274 __ai int16x8_t vrev32q_s16(int16x8_t __p0) {
   20275   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20276   int16x8_t __ret;
   20277   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
   20278   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20279   return __ret;
   20280 }
   20281 #endif
   20282 
   20283 #ifdef __LITTLE_ENDIAN__
   20284 __ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
   20285   uint8x8_t __ret;
   20286   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
   20287   return __ret;
   20288 }
   20289 #else
   20290 __ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
   20291   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20292   uint8x8_t __ret;
   20293   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
   20294   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20295   return __ret;
   20296 }
   20297 #endif
   20298 
   20299 #ifdef __LITTLE_ENDIAN__
   20300 __ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
   20301   uint16x4_t __ret;
   20302   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
   20303   return __ret;
   20304 }
   20305 #else
   20306 __ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
   20307   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20308   uint16x4_t __ret;
   20309   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
   20310   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20311   return __ret;
   20312 }
   20313 #endif
   20314 
   20315 #ifdef __LITTLE_ENDIAN__
   20316 __ai int8x8_t vrev32_s8(int8x8_t __p0) {
   20317   int8x8_t __ret;
   20318   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
   20319   return __ret;
   20320 }
   20321 #else
   20322 __ai int8x8_t vrev32_s8(int8x8_t __p0) {
   20323   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20324   int8x8_t __ret;
   20325   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
   20326   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20327   return __ret;
   20328 }
   20329 #endif
   20330 
   20331 #ifdef __LITTLE_ENDIAN__
   20332 __ai int16x4_t vrev32_s16(int16x4_t __p0) {
   20333   int16x4_t __ret;
   20334   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
   20335   return __ret;
   20336 }
   20337 #else
   20338 __ai int16x4_t vrev32_s16(int16x4_t __p0) {
   20339   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20340   int16x4_t __ret;
   20341   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
   20342   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20343   return __ret;
   20344 }
   20345 #endif
   20346 
   20347 #ifdef __LITTLE_ENDIAN__
   20348 __ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
   20349   poly8x8_t __ret;
   20350   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20351   return __ret;
   20352 }
   20353 #else
   20354 __ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
   20355   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20356   poly8x8_t __ret;
   20357   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
   20358   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20359   return __ret;
   20360 }
   20361 #endif
   20362 
   20363 #ifdef __LITTLE_ENDIAN__
   20364 __ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
   20365   poly16x4_t __ret;
   20366   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20367   return __ret;
   20368 }
   20369 #else
   20370 __ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
   20371   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20372   poly16x4_t __ret;
   20373   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
   20374   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20375   return __ret;
   20376 }
   20377 #endif
   20378 
   20379 #ifdef __LITTLE_ENDIAN__
   20380 __ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
   20381   poly8x16_t __ret;
   20382   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
   20383   return __ret;
   20384 }
   20385 #else
   20386 __ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
   20387   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20388   poly8x16_t __ret;
   20389   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
   20390   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20391   return __ret;
   20392 }
   20393 #endif
   20394 
   20395 #ifdef __LITTLE_ENDIAN__
   20396 __ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
   20397   poly16x8_t __ret;
   20398   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
   20399   return __ret;
   20400 }
   20401 #else
   20402 __ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
   20403   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20404   poly16x8_t __ret;
   20405   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
   20406   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20407   return __ret;
   20408 }
   20409 #endif
   20410 
   20411 #ifdef __LITTLE_ENDIAN__
   20412 __ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
   20413   uint8x16_t __ret;
   20414   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
   20415   return __ret;
   20416 }
   20417 #else
   20418 __ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
   20419   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20420   uint8x16_t __ret;
   20421   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
   20422   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20423   return __ret;
   20424 }
   20425 #endif
   20426 
   20427 #ifdef __LITTLE_ENDIAN__
   20428 __ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
   20429   uint32x4_t __ret;
   20430   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
   20431   return __ret;
   20432 }
   20433 #else
   20434 __ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
   20435   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20436   uint32x4_t __ret;
   20437   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
   20438   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20439   return __ret;
   20440 }
   20441 #endif
   20442 
   20443 #ifdef __LITTLE_ENDIAN__
   20444 __ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
   20445   uint16x8_t __ret;
   20446   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
   20447   return __ret;
   20448 }
   20449 #else
   20450 __ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
   20451   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20452   uint16x8_t __ret;
   20453   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
   20454   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20455   return __ret;
   20456 }
   20457 #endif
   20458 
   20459 #ifdef __LITTLE_ENDIAN__
   20460 __ai int8x16_t vrev64q_s8(int8x16_t __p0) {
   20461   int8x16_t __ret;
   20462   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
   20463   return __ret;
   20464 }
   20465 #else
   20466 __ai int8x16_t vrev64q_s8(int8x16_t __p0) {
   20467   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20468   int8x16_t __ret;
   20469   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
   20470   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20471   return __ret;
   20472 }
   20473 #endif
   20474 
   20475 #ifdef __LITTLE_ENDIAN__
   20476 __ai float32x4_t vrev64q_f32(float32x4_t __p0) {
   20477   float32x4_t __ret;
   20478   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
   20479   return __ret;
   20480 }
   20481 #else
   20482 __ai float32x4_t vrev64q_f32(float32x4_t __p0) {
   20483   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20484   float32x4_t __ret;
   20485   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
   20486   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20487   return __ret;
   20488 }
   20489 #endif
   20490 
   20491 #ifdef __LITTLE_ENDIAN__
   20492 __ai int32x4_t vrev64q_s32(int32x4_t __p0) {
   20493   int32x4_t __ret;
   20494   __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
   20495   return __ret;
   20496 }
   20497 #else
   20498 __ai int32x4_t vrev64q_s32(int32x4_t __p0) {
   20499   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20500   int32x4_t __ret;
   20501   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
   20502   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20503   return __ret;
   20504 }
   20505 #endif
   20506 
   20507 #ifdef __LITTLE_ENDIAN__
   20508 __ai int16x8_t vrev64q_s16(int16x8_t __p0) {
   20509   int16x8_t __ret;
   20510   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
   20511   return __ret;
   20512 }
   20513 #else
   20514 __ai int16x8_t vrev64q_s16(int16x8_t __p0) {
   20515   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20516   int16x8_t __ret;
   20517   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
   20518   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20519   return __ret;
   20520 }
   20521 #endif
   20522 
   20523 #ifdef __LITTLE_ENDIAN__
   20524 __ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
   20525   uint8x8_t __ret;
   20526   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20527   return __ret;
   20528 }
   20529 #else
   20530 __ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
   20531   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20532   uint8x8_t __ret;
   20533   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
   20534   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20535   return __ret;
   20536 }
   20537 #endif
   20538 
   20539 #ifdef __LITTLE_ENDIAN__
   20540 __ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
   20541   uint32x2_t __ret;
   20542   __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
   20543   return __ret;
   20544 }
   20545 #else
   20546 __ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
   20547   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20548   uint32x2_t __ret;
   20549   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
   20550   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20551   return __ret;
   20552 }
   20553 #endif
   20554 
   20555 #ifdef __LITTLE_ENDIAN__
   20556 __ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
   20557   uint16x4_t __ret;
   20558   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20559   return __ret;
   20560 }
   20561 #else
   20562 __ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
   20563   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20564   uint16x4_t __ret;
   20565   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
   20566   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20567   return __ret;
   20568 }
   20569 #endif
   20570 
   20571 #ifdef __LITTLE_ENDIAN__
   20572 __ai int8x8_t vrev64_s8(int8x8_t __p0) {
   20573   int8x8_t __ret;
   20574   __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20575   return __ret;
   20576 }
   20577 #else
   20578 __ai int8x8_t vrev64_s8(int8x8_t __p0) {
   20579   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20580   int8x8_t __ret;
   20581   __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
   20582   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20583   return __ret;
   20584 }
   20585 #endif
   20586 
   20587 #ifdef __LITTLE_ENDIAN__
   20588 __ai float32x2_t vrev64_f32(float32x2_t __p0) {
   20589   float32x2_t __ret;
   20590   __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
   20591   return __ret;
   20592 }
   20593 #else
   20594 __ai float32x2_t vrev64_f32(float32x2_t __p0) {
   20595   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20596   float32x2_t __ret;
   20597   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
   20598   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20599   return __ret;
   20600 }
   20601 #endif
   20602 
   20603 #ifdef __LITTLE_ENDIAN__
   20604 __ai int32x2_t vrev64_s32(int32x2_t __p0) {
   20605   int32x2_t __ret;
   20606   __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
   20607   return __ret;
   20608 }
   20609 #else
   20610 __ai int32x2_t vrev64_s32(int32x2_t __p0) {
   20611   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20612   int32x2_t __ret;
   20613   __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
   20614   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20615   return __ret;
   20616 }
   20617 #endif
   20618 
   20619 #ifdef __LITTLE_ENDIAN__
   20620 __ai int16x4_t vrev64_s16(int16x4_t __p0) {
   20621   int16x4_t __ret;
   20622   __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20623   return __ret;
   20624 }
   20625 #else
   20626 __ai int16x4_t vrev64_s16(int16x4_t __p0) {
   20627   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20628   int16x4_t __ret;
   20629   __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
   20630   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20631   return __ret;
   20632 }
   20633 #endif
   20634 
   20635 #ifdef __LITTLE_ENDIAN__
   20636 __ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   20637   uint8x16_t __ret;
   20638   __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   20639   return __ret;
   20640 }
   20641 #else
   20642 __ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   20643   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20644   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20645   uint8x16_t __ret;
   20646   __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   20647   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20648   return __ret;
   20649 }
   20650 #endif
   20651 
   20652 #ifdef __LITTLE_ENDIAN__
   20653 __ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   20654   uint32x4_t __ret;
   20655   __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   20656   return __ret;
   20657 }
   20658 #else
   20659 __ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   20660   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20661   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   20662   uint32x4_t __ret;
   20663   __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   20664   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20665   return __ret;
   20666 }
   20667 #endif
   20668 
   20669 #ifdef __LITTLE_ENDIAN__
   20670 __ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   20671   uint16x8_t __ret;
   20672   __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   20673   return __ret;
   20674 }
   20675 #else
   20676 __ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   20677   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20678   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   20679   uint16x8_t __ret;
   20680   __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   20681   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20682   return __ret;
   20683 }
   20684 #endif
   20685 
   20686 #ifdef __LITTLE_ENDIAN__
   20687 __ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   20688   int8x16_t __ret;
   20689   __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   20690   return __ret;
   20691 }
   20692 #else
   20693 __ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   20694   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20695   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20696   int8x16_t __ret;
   20697   __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   20698   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20699   return __ret;
   20700 }
   20701 #endif
   20702 
   20703 #ifdef __LITTLE_ENDIAN__
   20704 __ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   20705   int32x4_t __ret;
   20706   __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   20707   return __ret;
   20708 }
   20709 #else
   20710 __ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   20711   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20712   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   20713   int32x4_t __ret;
   20714   __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   20715   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20716   return __ret;
   20717 }
   20718 #endif
   20719 
   20720 #ifdef __LITTLE_ENDIAN__
   20721 __ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   20722   int16x8_t __ret;
   20723   __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   20724   return __ret;
   20725 }
   20726 #else
   20727 __ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   20728   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20729   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   20730   int16x8_t __ret;
   20731   __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   20732   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20733   return __ret;
   20734 }
   20735 #endif
   20736 
   20737 #ifdef __LITTLE_ENDIAN__
   20738 __ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   20739   uint8x8_t __ret;
   20740   __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   20741   return __ret;
   20742 }
   20743 #else
   20744 __ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   20745   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20746   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   20747   uint8x8_t __ret;
   20748   __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   20749   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20750   return __ret;
   20751 }
   20752 #endif
   20753 
   20754 #ifdef __LITTLE_ENDIAN__
   20755 __ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   20756   uint32x2_t __ret;
   20757   __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   20758   return __ret;
   20759 }
   20760 #else
   20761 __ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   20762   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20763   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   20764   uint32x2_t __ret;
   20765   __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   20766   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20767   return __ret;
   20768 }
   20769 #endif
   20770 
   20771 #ifdef __LITTLE_ENDIAN__
   20772 __ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   20773   uint16x4_t __ret;
   20774   __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   20775   return __ret;
   20776 }
   20777 #else
   20778 __ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   20779   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20780   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   20781   uint16x4_t __ret;
   20782   __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   20783   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20784   return __ret;
   20785 }
   20786 #endif
   20787 
   20788 #ifdef __LITTLE_ENDIAN__
   20789 __ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
   20790   int8x8_t __ret;
   20791   __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   20792   return __ret;
   20793 }
   20794 #else
   20795 __ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
   20796   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20797   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   20798   int8x8_t __ret;
   20799   __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   20800   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20801   return __ret;
   20802 }
   20803 #endif
   20804 
   20805 #ifdef __LITTLE_ENDIAN__
   20806 __ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
   20807   int32x2_t __ret;
   20808   __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   20809   return __ret;
   20810 }
   20811 #else
   20812 __ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
   20813   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20814   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   20815   int32x2_t __ret;
   20816   __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   20817   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20818   return __ret;
   20819 }
   20820 #endif
   20821 
   20822 #ifdef __LITTLE_ENDIAN__
   20823 __ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
   20824   int16x4_t __ret;
   20825   __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   20826   return __ret;
   20827 }
   20828 #else
   20829 __ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
   20830   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20831   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   20832   int16x4_t __ret;
   20833   __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   20834   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20835   return __ret;
   20836 }
   20837 #endif
   20838 
   20839 #ifdef __LITTLE_ENDIAN__
   20840 __ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
   20841   uint8x16_t __ret;
   20842   __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   20843   return __ret;
   20844 }
   20845 #else
   20846 __ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
   20847   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20848   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20849   uint8x16_t __ret;
   20850   __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   20851   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20852   return __ret;
   20853 }
   20854 #endif
   20855 
   20856 #ifdef __LITTLE_ENDIAN__
   20857 __ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
   20858   uint32x4_t __ret;
   20859   __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   20860   return __ret;
   20861 }
   20862 #else
   20863 __ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
   20864   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20865   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   20866   uint32x4_t __ret;
   20867   __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   20868   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20869   return __ret;
   20870 }
   20871 #endif
   20872 
   20873 #ifdef __LITTLE_ENDIAN__
   20874 __ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
   20875   uint64x2_t __ret;
   20876   __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   20877   return __ret;
   20878 }
   20879 #else
   20880 __ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
   20881   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20882   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   20883   uint64x2_t __ret;
   20884   __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   20885   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20886   return __ret;
   20887 }
   20888 #endif
   20889 
   20890 #ifdef __LITTLE_ENDIAN__
   20891 __ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
   20892   uint16x8_t __ret;
   20893   __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   20894   return __ret;
   20895 }
   20896 #else
   20897 __ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
   20898   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20899   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   20900   uint16x8_t __ret;
   20901   __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   20902   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20903   return __ret;
   20904 }
   20905 #endif
   20906 
   20907 #ifdef __LITTLE_ENDIAN__
   20908 __ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
   20909   int8x16_t __ret;
   20910   __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   20911   return __ret;
   20912 }
   20913 #else
   20914 __ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
   20915   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20916   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20917   int8x16_t __ret;
   20918   __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   20919   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   20920   return __ret;
   20921 }
   20922 #endif
   20923 
   20924 #ifdef __LITTLE_ENDIAN__
   20925 __ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
   20926   int32x4_t __ret;
   20927   __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   20928   return __ret;
   20929 }
   20930 #else
   20931 __ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
   20932   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   20933   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   20934   int32x4_t __ret;
   20935   __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   20936   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   20937   return __ret;
   20938 }
   20939 #endif
   20940 
   20941 #ifdef __LITTLE_ENDIAN__
   20942 __ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
   20943   int64x2_t __ret;
   20944   __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
   20945   return __ret;
   20946 }
   20947 #else
   20948 __ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
   20949   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   20950   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   20951   int64x2_t __ret;
   20952   __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
   20953   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   20954   return __ret;
   20955 }
   20956 #endif
   20957 
   20958 #ifdef __LITTLE_ENDIAN__
   20959 __ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
   20960   int16x8_t __ret;
   20961   __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   20962   return __ret;
   20963 }
   20964 #else
   20965 __ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
   20966   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20967   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   20968   int16x8_t __ret;
   20969   __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   20970   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20971   return __ret;
   20972 }
   20973 #endif
   20974 
   20975 #ifdef __LITTLE_ENDIAN__
   20976 __ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
   20977   uint8x8_t __ret;
   20978   __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   20979   return __ret;
   20980 }
   20981 #else
   20982 __ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
   20983   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   20984   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   20985   uint8x8_t __ret;
   20986   __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   20987   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   20988   return __ret;
   20989 }
   20990 #endif
   20991 
   20992 #ifdef __LITTLE_ENDIAN__
   20993 __ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
   20994   uint32x2_t __ret;
   20995   __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   20996   return __ret;
   20997 }
   20998 #else
   20999 __ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
   21000   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   21001   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   21002   uint32x2_t __ret;
   21003   __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   21004   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   21005   return __ret;
   21006 }
   21007 #endif
   21008 
   21009 #ifdef __LITTLE_ENDIAN__
   21010 __ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
   21011   uint64x1_t __ret;
   21012   __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   21013   return __ret;
   21014 }
   21015 #else
   21016 __ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
   21017   uint64x1_t __ret;
   21018   __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   21019   return __ret;
   21020 }
   21021 #endif
   21022 
   21023 #ifdef __LITTLE_ENDIAN__
   21024 __ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
   21025   uint16x4_t __ret;
   21026   __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   21027   return __ret;
   21028 }
   21029 #else
   21030 __ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
   21031   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   21032   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   21033   uint16x4_t __ret;
   21034   __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   21035   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   21036   return __ret;
   21037 }
   21038 #endif
   21039 
   21040 #ifdef __LITTLE_ENDIAN__
   21041 __ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
   21042   int8x8_t __ret;
   21043   __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   21044   return __ret;
   21045 }
   21046 #else
   21047 __ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
   21048   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   21049   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   21050   int8x8_t __ret;
   21051   __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   21052   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   21053   return __ret;
   21054 }
   21055 #endif
   21056 
   21057 #ifdef __LITTLE_ENDIAN__
   21058 __ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
   21059   int32x2_t __ret;
   21060   __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   21061   return __ret;
   21062 }
   21063 #else
   21064 __ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
   21065   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   21066   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   21067   int32x2_t __ret;
   21068   __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   21069   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   21070   return __ret;
   21071 }
   21072 #endif
   21073 
   21074 #ifdef __LITTLE_ENDIAN__
   21075 __ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
   21076   int64x1_t __ret;
   21077   __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   21078   return __ret;
   21079 }
   21080 #else
   21081 __ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
   21082   int64x1_t __ret;
   21083   __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   21084   return __ret;
   21085 }
   21086 #endif
   21087 
   21088 #ifdef __LITTLE_ENDIAN__
   21089 __ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
   21090   int16x4_t __ret;
   21091   __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   21092   return __ret;
   21093 }
   21094 #else
   21095 __ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
   21096   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   21097   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   21098   int16x4_t __ret;
   21099   __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   21100   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   21101   return __ret;
   21102 }
   21103 #endif
   21104 
   21105 #ifdef __LITTLE_ENDIAN__
   21106 #define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
   21107   uint8x16_t __s0 = __p0; \
   21108   uint8x16_t __ret; \
   21109   __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 48); \
   21110   __ret; \
   21111 })
   21112 #else
   21113 #define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
   21114   uint8x16_t __s0 = __p0; \
   21115   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21116   uint8x16_t __ret; \
   21117   __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 48); \
   21118   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21119   __ret; \
   21120 })
   21121 #endif
   21122 
   21123 #ifdef __LITTLE_ENDIAN__
   21124 #define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
   21125   uint32x4_t __s0 = __p0; \
   21126   uint32x4_t __ret; \
   21127   __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 50); \
   21128   __ret; \
   21129 })
   21130 #else
   21131 #define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
   21132   uint32x4_t __s0 = __p0; \
   21133   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21134   uint32x4_t __ret; \
   21135   __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 50); \
   21136   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21137   __ret; \
   21138 })
   21139 #endif
   21140 
   21141 #ifdef __LITTLE_ENDIAN__
   21142 #define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
   21143   uint64x2_t __s0 = __p0; \
   21144   uint64x2_t __ret; \
   21145   __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 51); \
   21146   __ret; \
   21147 })
   21148 #else
   21149 #define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
   21150   uint64x2_t __s0 = __p0; \
   21151   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21152   uint64x2_t __ret; \
   21153   __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 51); \
   21154   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21155   __ret; \
   21156 })
   21157 #endif
   21158 
   21159 #ifdef __LITTLE_ENDIAN__
   21160 #define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
   21161   uint16x8_t __s0 = __p0; \
   21162   uint16x8_t __ret; \
   21163   __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 49); \
   21164   __ret; \
   21165 })
   21166 #else
   21167 #define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
   21168   uint16x8_t __s0 = __p0; \
   21169   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21170   uint16x8_t __ret; \
   21171   __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 49); \
   21172   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21173   __ret; \
   21174 })
   21175 #endif
   21176 
   21177 #ifdef __LITTLE_ENDIAN__
   21178 #define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
   21179   int8x16_t __s0 = __p0; \
   21180   int8x16_t __ret; \
   21181   __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 32); \
   21182   __ret; \
   21183 })
   21184 #else
   21185 #define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
   21186   int8x16_t __s0 = __p0; \
   21187   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21188   int8x16_t __ret; \
   21189   __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 32); \
   21190   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21191   __ret; \
   21192 })
   21193 #endif
   21194 
   21195 #ifdef __LITTLE_ENDIAN__
   21196 #define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
   21197   int32x4_t __s0 = __p0; \
   21198   int32x4_t __ret; \
   21199   __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 34); \
   21200   __ret; \
   21201 })
   21202 #else
   21203 #define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
   21204   int32x4_t __s0 = __p0; \
   21205   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21206   int32x4_t __ret; \
   21207   __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 34); \
   21208   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21209   __ret; \
   21210 })
   21211 #endif
   21212 
   21213 #ifdef __LITTLE_ENDIAN__
   21214 #define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
   21215   int64x2_t __s0 = __p0; \
   21216   int64x2_t __ret; \
   21217   __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 35); \
   21218   __ret; \
   21219 })
   21220 #else
   21221 #define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
   21222   int64x2_t __s0 = __p0; \
   21223   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21224   int64x2_t __ret; \
   21225   __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 35); \
   21226   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21227   __ret; \
   21228 })
   21229 #endif
   21230 
   21231 #ifdef __LITTLE_ENDIAN__
   21232 #define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
   21233   int16x8_t __s0 = __p0; \
   21234   int16x8_t __ret; \
   21235   __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 33); \
   21236   __ret; \
   21237 })
   21238 #else
   21239 #define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
   21240   int16x8_t __s0 = __p0; \
   21241   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21242   int16x8_t __ret; \
   21243   __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 33); \
   21244   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21245   __ret; \
   21246 })
   21247 #endif
   21248 
   21249 #ifdef __LITTLE_ENDIAN__
   21250 #define vrshr_n_u8(__p0, __p1) __extension__ ({ \
   21251   uint8x8_t __s0 = __p0; \
   21252   uint8x8_t __ret; \
   21253   __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 16); \
   21254   __ret; \
   21255 })
   21256 #else
   21257 #define vrshr_n_u8(__p0, __p1) __extension__ ({ \
   21258   uint8x8_t __s0 = __p0; \
   21259   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21260   uint8x8_t __ret; \
   21261   __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 16); \
   21262   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21263   __ret; \
   21264 })
   21265 #endif
   21266 
   21267 #ifdef __LITTLE_ENDIAN__
   21268 #define vrshr_n_u32(__p0, __p1) __extension__ ({ \
   21269   uint32x2_t __s0 = __p0; \
   21270   uint32x2_t __ret; \
   21271   __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 18); \
   21272   __ret; \
   21273 })
   21274 #else
   21275 #define vrshr_n_u32(__p0, __p1) __extension__ ({ \
   21276   uint32x2_t __s0 = __p0; \
   21277   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21278   uint32x2_t __ret; \
   21279   __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 18); \
   21280   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21281   __ret; \
   21282 })
   21283 #endif
   21284 
   21285 #ifdef __LITTLE_ENDIAN__
   21286 #define vrshr_n_u64(__p0, __p1) __extension__ ({ \
   21287   uint64x1_t __s0 = __p0; \
   21288   uint64x1_t __ret; \
   21289   __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
   21290   __ret; \
   21291 })
   21292 #else
   21293 #define vrshr_n_u64(__p0, __p1) __extension__ ({ \
   21294   uint64x1_t __s0 = __p0; \
   21295   uint64x1_t __ret; \
   21296   __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
   21297   __ret; \
   21298 })
   21299 #endif
   21300 
   21301 #ifdef __LITTLE_ENDIAN__
   21302 #define vrshr_n_u16(__p0, __p1) __extension__ ({ \
   21303   uint16x4_t __s0 = __p0; \
   21304   uint16x4_t __ret; \
   21305   __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 17); \
   21306   __ret; \
   21307 })
   21308 #else
   21309 #define vrshr_n_u16(__p0, __p1) __extension__ ({ \
   21310   uint16x4_t __s0 = __p0; \
   21311   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21312   uint16x4_t __ret; \
   21313   __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 17); \
   21314   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21315   __ret; \
   21316 })
   21317 #endif
   21318 
   21319 #ifdef __LITTLE_ENDIAN__
   21320 #define vrshr_n_s8(__p0, __p1) __extension__ ({ \
   21321   int8x8_t __s0 = __p0; \
   21322   int8x8_t __ret; \
   21323   __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 0); \
   21324   __ret; \
   21325 })
   21326 #else
   21327 #define vrshr_n_s8(__p0, __p1) __extension__ ({ \
   21328   int8x8_t __s0 = __p0; \
   21329   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21330   int8x8_t __ret; \
   21331   __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 0); \
   21332   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21333   __ret; \
   21334 })
   21335 #endif
   21336 
   21337 #ifdef __LITTLE_ENDIAN__
   21338 #define vrshr_n_s32(__p0, __p1) __extension__ ({ \
   21339   int32x2_t __s0 = __p0; \
   21340   int32x2_t __ret; \
   21341   __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 2); \
   21342   __ret; \
   21343 })
   21344 #else
   21345 #define vrshr_n_s32(__p0, __p1) __extension__ ({ \
   21346   int32x2_t __s0 = __p0; \
   21347   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21348   int32x2_t __ret; \
   21349   __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 2); \
   21350   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21351   __ret; \
   21352 })
   21353 #endif
   21354 
   21355 #ifdef __LITTLE_ENDIAN__
   21356 #define vrshr_n_s64(__p0, __p1) __extension__ ({ \
   21357   int64x1_t __s0 = __p0; \
   21358   int64x1_t __ret; \
   21359   __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
   21360   __ret; \
   21361 })
   21362 #else
   21363 #define vrshr_n_s64(__p0, __p1) __extension__ ({ \
   21364   int64x1_t __s0 = __p0; \
   21365   int64x1_t __ret; \
   21366   __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
   21367   __ret; \
   21368 })
   21369 #endif
   21370 
   21371 #ifdef __LITTLE_ENDIAN__
   21372 #define vrshr_n_s16(__p0, __p1) __extension__ ({ \
   21373   int16x4_t __s0 = __p0; \
   21374   int16x4_t __ret; \
   21375   __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 1); \
   21376   __ret; \
   21377 })
   21378 #else
   21379 #define vrshr_n_s16(__p0, __p1) __extension__ ({ \
   21380   int16x4_t __s0 = __p0; \
   21381   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21382   int16x4_t __ret; \
   21383   __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 1); \
   21384   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21385   __ret; \
   21386 })
   21387 #endif
   21388 
   21389 #ifdef __LITTLE_ENDIAN__
   21390 #define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
   21391   uint32x4_t __s0 = __p0; \
   21392   uint16x4_t __ret; \
   21393   __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
   21394   __ret; \
   21395 })
   21396 #else
   21397 #define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
   21398   uint32x4_t __s0 = __p0; \
   21399   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21400   uint16x4_t __ret; \
   21401   __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
   21402   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21403   __ret; \
   21404 })
   21405 #define __noswap_vrshrn_n_u32(__p0, __p1) __extension__ ({ \
   21406   uint32x4_t __s0 = __p0; \
   21407   uint16x4_t __ret; \
   21408   __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
   21409   __ret; \
   21410 })
   21411 #endif
   21412 
   21413 #ifdef __LITTLE_ENDIAN__
   21414 #define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
   21415   uint64x2_t __s0 = __p0; \
   21416   uint32x2_t __ret; \
   21417   __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
   21418   __ret; \
   21419 })
   21420 #else
   21421 #define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
   21422   uint64x2_t __s0 = __p0; \
   21423   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21424   uint32x2_t __ret; \
   21425   __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
   21426   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21427   __ret; \
   21428 })
   21429 #define __noswap_vrshrn_n_u64(__p0, __p1) __extension__ ({ \
   21430   uint64x2_t __s0 = __p0; \
   21431   uint32x2_t __ret; \
   21432   __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
   21433   __ret; \
   21434 })
   21435 #endif
   21436 
   21437 #ifdef __LITTLE_ENDIAN__
   21438 #define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
   21439   uint16x8_t __s0 = __p0; \
   21440   uint8x8_t __ret; \
   21441   __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
   21442   __ret; \
   21443 })
   21444 #else
   21445 #define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
   21446   uint16x8_t __s0 = __p0; \
   21447   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21448   uint8x8_t __ret; \
   21449   __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
   21450   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21451   __ret; \
   21452 })
   21453 #define __noswap_vrshrn_n_u16(__p0, __p1) __extension__ ({ \
   21454   uint16x8_t __s0 = __p0; \
   21455   uint8x8_t __ret; \
   21456   __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
   21457   __ret; \
   21458 })
   21459 #endif
   21460 
   21461 #ifdef __LITTLE_ENDIAN__
   21462 #define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
   21463   int32x4_t __s0 = __p0; \
   21464   int16x4_t __ret; \
   21465   __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
   21466   __ret; \
   21467 })
   21468 #else
   21469 #define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
   21470   int32x4_t __s0 = __p0; \
   21471   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21472   int16x4_t __ret; \
   21473   __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
   21474   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21475   __ret; \
   21476 })
   21477 #define __noswap_vrshrn_n_s32(__p0, __p1) __extension__ ({ \
   21478   int32x4_t __s0 = __p0; \
   21479   int16x4_t __ret; \
   21480   __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
   21481   __ret; \
   21482 })
   21483 #endif
   21484 
   21485 #ifdef __LITTLE_ENDIAN__
   21486 #define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
   21487   int64x2_t __s0 = __p0; \
   21488   int32x2_t __ret; \
   21489   __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
   21490   __ret; \
   21491 })
   21492 #else
   21493 #define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
   21494   int64x2_t __s0 = __p0; \
   21495   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21496   int32x2_t __ret; \
   21497   __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
   21498   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21499   __ret; \
   21500 })
   21501 #define __noswap_vrshrn_n_s64(__p0, __p1) __extension__ ({ \
   21502   int64x2_t __s0 = __p0; \
   21503   int32x2_t __ret; \
   21504   __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
   21505   __ret; \
   21506 })
   21507 #endif
   21508 
   21509 #ifdef __LITTLE_ENDIAN__
   21510 #define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
   21511   int16x8_t __s0 = __p0; \
   21512   int8x8_t __ret; \
   21513   __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
   21514   __ret; \
   21515 })
   21516 #else
   21517 #define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
   21518   int16x8_t __s0 = __p0; \
   21519   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21520   int8x8_t __ret; \
   21521   __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
   21522   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21523   __ret; \
   21524 })
   21525 #define __noswap_vrshrn_n_s16(__p0, __p1) __extension__ ({ \
   21526   int16x8_t __s0 = __p0; \
   21527   int8x8_t __ret; \
   21528   __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
   21529   __ret; \
   21530 })
   21531 #endif
   21532 
   21533 #ifdef __LITTLE_ENDIAN__
   21534 __ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
   21535   uint32x4_t __ret;
   21536   __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 50);
   21537   return __ret;
   21538 }
   21539 #else
   21540 __ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
   21541   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   21542   uint32x4_t __ret;
   21543   __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 50);
   21544   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   21545   return __ret;
   21546 }
   21547 #endif
   21548 
   21549 #ifdef __LITTLE_ENDIAN__
   21550 __ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
   21551   float32x4_t __ret;
   21552   __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 41);
   21553   return __ret;
   21554 }
   21555 #else
   21556 __ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
   21557   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   21558   float32x4_t __ret;
   21559   __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 41);
   21560   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   21561   return __ret;
   21562 }
   21563 #endif
   21564 
   21565 #ifdef __LITTLE_ENDIAN__
   21566 __ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
   21567   uint32x2_t __ret;
   21568   __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 18);
   21569   return __ret;
   21570 }
   21571 #else
   21572 __ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
   21573   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   21574   uint32x2_t __ret;
   21575   __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 18);
   21576   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   21577   return __ret;
   21578 }
   21579 #endif
   21580 
   21581 #ifdef __LITTLE_ENDIAN__
   21582 __ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
   21583   float32x2_t __ret;
   21584   __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 9);
   21585   return __ret;
   21586 }
   21587 #else
   21588 __ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
   21589   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   21590   float32x2_t __ret;
   21591   __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 9);
   21592   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   21593   return __ret;
   21594 }
   21595 #endif
   21596 
   21597 #ifdef __LITTLE_ENDIAN__
   21598 __ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
   21599   float32x4_t __ret;
   21600   __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   21601   return __ret;
   21602 }
   21603 #else
   21604 __ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
   21605   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   21606   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   21607   float32x4_t __ret;
   21608   __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   21609   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   21610   return __ret;
   21611 }
   21612 #endif
   21613 
   21614 #ifdef __LITTLE_ENDIAN__
   21615 __ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
   21616   float32x2_t __ret;
   21617   __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   21618   return __ret;
   21619 }
   21620 #else
   21621 __ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
   21622   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   21623   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   21624   float32x2_t __ret;
   21625   __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   21626   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   21627   return __ret;
   21628 }
   21629 #endif
   21630 
   21631 #ifdef __LITTLE_ENDIAN__
   21632 #define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
   21633   uint8x16_t __s0 = __p0; \
   21634   uint8x16_t __s1 = __p1; \
   21635   uint8x16_t __ret; \
   21636   __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
   21637   __ret; \
   21638 })
   21639 #else
   21640 #define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
   21641   uint8x16_t __s0 = __p0; \
   21642   uint8x16_t __s1 = __p1; \
   21643   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21644   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21645   uint8x16_t __ret; \
   21646   __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
   21647   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21648   __ret; \
   21649 })
   21650 #endif
   21651 
   21652 #ifdef __LITTLE_ENDIAN__
   21653 #define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
   21654   uint32x4_t __s0 = __p0; \
   21655   uint32x4_t __s1 = __p1; \
   21656   uint32x4_t __ret; \
   21657   __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
   21658   __ret; \
   21659 })
   21660 #else
   21661 #define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
   21662   uint32x4_t __s0 = __p0; \
   21663   uint32x4_t __s1 = __p1; \
   21664   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21665   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   21666   uint32x4_t __ret; \
   21667   __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
   21668   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21669   __ret; \
   21670 })
   21671 #endif
   21672 
   21673 #ifdef __LITTLE_ENDIAN__
   21674 #define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
   21675   uint64x2_t __s0 = __p0; \
   21676   uint64x2_t __s1 = __p1; \
   21677   uint64x2_t __ret; \
   21678   __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
   21679   __ret; \
   21680 })
   21681 #else
   21682 #define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
   21683   uint64x2_t __s0 = __p0; \
   21684   uint64x2_t __s1 = __p1; \
   21685   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21686   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   21687   uint64x2_t __ret; \
   21688   __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
   21689   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21690   __ret; \
   21691 })
   21692 #endif
   21693 
   21694 #ifdef __LITTLE_ENDIAN__
   21695 #define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
   21696   uint16x8_t __s0 = __p0; \
   21697   uint16x8_t __s1 = __p1; \
   21698   uint16x8_t __ret; \
   21699   __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
   21700   __ret; \
   21701 })
   21702 #else
   21703 #define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
   21704   uint16x8_t __s0 = __p0; \
   21705   uint16x8_t __s1 = __p1; \
   21706   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21707   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   21708   uint16x8_t __ret; \
   21709   __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
   21710   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21711   __ret; \
   21712 })
   21713 #endif
   21714 
   21715 #ifdef __LITTLE_ENDIAN__
   21716 #define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
   21717   int8x16_t __s0 = __p0; \
   21718   int8x16_t __s1 = __p1; \
   21719   int8x16_t __ret; \
   21720   __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
   21721   __ret; \
   21722 })
   21723 #else
   21724 #define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
   21725   int8x16_t __s0 = __p0; \
   21726   int8x16_t __s1 = __p1; \
   21727   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21728   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21729   int8x16_t __ret; \
   21730   __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
   21731   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   21732   __ret; \
   21733 })
   21734 #endif
   21735 
   21736 #ifdef __LITTLE_ENDIAN__
   21737 #define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
   21738   int32x4_t __s0 = __p0; \
   21739   int32x4_t __s1 = __p1; \
   21740   int32x4_t __ret; \
   21741   __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
   21742   __ret; \
   21743 })
   21744 #else
   21745 #define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
   21746   int32x4_t __s0 = __p0; \
   21747   int32x4_t __s1 = __p1; \
   21748   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21749   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   21750   int32x4_t __ret; \
   21751   __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
   21752   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21753   __ret; \
   21754 })
   21755 #endif
   21756 
   21757 #ifdef __LITTLE_ENDIAN__
   21758 #define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
   21759   int64x2_t __s0 = __p0; \
   21760   int64x2_t __s1 = __p1; \
   21761   int64x2_t __ret; \
   21762   __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
   21763   __ret; \
   21764 })
   21765 #else
   21766 #define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
   21767   int64x2_t __s0 = __p0; \
   21768   int64x2_t __s1 = __p1; \
   21769   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21770   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   21771   int64x2_t __ret; \
   21772   __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
   21773   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21774   __ret; \
   21775 })
   21776 #endif
   21777 
   21778 #ifdef __LITTLE_ENDIAN__
   21779 #define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
   21780   int16x8_t __s0 = __p0; \
   21781   int16x8_t __s1 = __p1; \
   21782   int16x8_t __ret; \
   21783   __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
   21784   __ret; \
   21785 })
   21786 #else
   21787 #define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
   21788   int16x8_t __s0 = __p0; \
   21789   int16x8_t __s1 = __p1; \
   21790   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21791   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   21792   int16x8_t __ret; \
   21793   __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
   21794   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21795   __ret; \
   21796 })
   21797 #endif
   21798 
   21799 #ifdef __LITTLE_ENDIAN__
   21800 #define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
   21801   uint8x8_t __s0 = __p0; \
   21802   uint8x8_t __s1 = __p1; \
   21803   uint8x8_t __ret; \
   21804   __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
   21805   __ret; \
   21806 })
   21807 #else
   21808 #define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
   21809   uint8x8_t __s0 = __p0; \
   21810   uint8x8_t __s1 = __p1; \
   21811   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21812   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   21813   uint8x8_t __ret; \
   21814   __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
   21815   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21816   __ret; \
   21817 })
   21818 #endif
   21819 
   21820 #ifdef __LITTLE_ENDIAN__
   21821 #define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
   21822   uint32x2_t __s0 = __p0; \
   21823   uint32x2_t __s1 = __p1; \
   21824   uint32x2_t __ret; \
   21825   __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
   21826   __ret; \
   21827 })
   21828 #else
   21829 #define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
   21830   uint32x2_t __s0 = __p0; \
   21831   uint32x2_t __s1 = __p1; \
   21832   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21833   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   21834   uint32x2_t __ret; \
   21835   __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
   21836   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21837   __ret; \
   21838 })
   21839 #endif
   21840 
   21841 #ifdef __LITTLE_ENDIAN__
   21842 #define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
   21843   uint64x1_t __s0 = __p0; \
   21844   uint64x1_t __s1 = __p1; \
   21845   uint64x1_t __ret; \
   21846   __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   21847   __ret; \
   21848 })
   21849 #else
   21850 #define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
   21851   uint64x1_t __s0 = __p0; \
   21852   uint64x1_t __s1 = __p1; \
   21853   uint64x1_t __ret; \
   21854   __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   21855   __ret; \
   21856 })
   21857 #endif
   21858 
   21859 #ifdef __LITTLE_ENDIAN__
   21860 #define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
   21861   uint16x4_t __s0 = __p0; \
   21862   uint16x4_t __s1 = __p1; \
   21863   uint16x4_t __ret; \
   21864   __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
   21865   __ret; \
   21866 })
   21867 #else
   21868 #define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
   21869   uint16x4_t __s0 = __p0; \
   21870   uint16x4_t __s1 = __p1; \
   21871   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21872   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   21873   uint16x4_t __ret; \
   21874   __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
   21875   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21876   __ret; \
   21877 })
   21878 #endif
   21879 
   21880 #ifdef __LITTLE_ENDIAN__
   21881 #define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
   21882   int8x8_t __s0 = __p0; \
   21883   int8x8_t __s1 = __p1; \
   21884   int8x8_t __ret; \
   21885   __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
   21886   __ret; \
   21887 })
   21888 #else
   21889 #define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
   21890   int8x8_t __s0 = __p0; \
   21891   int8x8_t __s1 = __p1; \
   21892   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   21893   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   21894   int8x8_t __ret; \
   21895   __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
   21896   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   21897   __ret; \
   21898 })
   21899 #endif
   21900 
   21901 #ifdef __LITTLE_ENDIAN__
   21902 #define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
   21903   int32x2_t __s0 = __p0; \
   21904   int32x2_t __s1 = __p1; \
   21905   int32x2_t __ret; \
   21906   __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
   21907   __ret; \
   21908 })
   21909 #else
   21910 #define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
   21911   int32x2_t __s0 = __p0; \
   21912   int32x2_t __s1 = __p1; \
   21913   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   21914   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   21915   int32x2_t __ret; \
   21916   __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
   21917   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   21918   __ret; \
   21919 })
   21920 #endif
   21921 
   21922 #ifdef __LITTLE_ENDIAN__
   21923 #define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
   21924   int64x1_t __s0 = __p0; \
   21925   int64x1_t __s1 = __p1; \
   21926   int64x1_t __ret; \
   21927   __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   21928   __ret; \
   21929 })
   21930 #else
   21931 #define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
   21932   int64x1_t __s0 = __p0; \
   21933   int64x1_t __s1 = __p1; \
   21934   int64x1_t __ret; \
   21935   __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   21936   __ret; \
   21937 })
   21938 #endif
   21939 
   21940 #ifdef __LITTLE_ENDIAN__
   21941 #define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
   21942   int16x4_t __s0 = __p0; \
   21943   int16x4_t __s1 = __p1; \
   21944   int16x4_t __ret; \
   21945   __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
   21946   __ret; \
   21947 })
   21948 #else
   21949 #define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
   21950   int16x4_t __s0 = __p0; \
   21951   int16x4_t __s1 = __p1; \
   21952   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   21953   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   21954   int16x4_t __ret; \
   21955   __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
   21956   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   21957   __ret; \
   21958 })
   21959 #endif
   21960 
   21961 #ifdef __LITTLE_ENDIAN__
   21962 __ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   21963   uint16x4_t __ret;
   21964   __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
   21965   return __ret;
   21966 }
   21967 #else
   21968 __ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   21969   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   21970   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   21971   uint16x4_t __ret;
   21972   __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
   21973   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   21974   return __ret;
   21975 }
   21976 __ai uint16x4_t __noswap_vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   21977   uint16x4_t __ret;
   21978   __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
   21979   return __ret;
   21980 }
   21981 #endif
   21982 
   21983 #ifdef __LITTLE_ENDIAN__
   21984 __ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   21985   uint32x2_t __ret;
   21986   __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
   21987   return __ret;
   21988 }
   21989 #else
   21990 __ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   21991   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   21992   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   21993   uint32x2_t __ret;
   21994   __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
   21995   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   21996   return __ret;
   21997 }
   21998 __ai uint32x2_t __noswap_vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   21999   uint32x2_t __ret;
   22000   __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
   22001   return __ret;
   22002 }
   22003 #endif
   22004 
   22005 #ifdef __LITTLE_ENDIAN__
   22006 __ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   22007   uint8x8_t __ret;
   22008   __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
   22009   return __ret;
   22010 }
   22011 #else
   22012 __ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   22013   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   22014   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   22015   uint8x8_t __ret;
   22016   __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
   22017   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   22018   return __ret;
   22019 }
   22020 __ai uint8x8_t __noswap_vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   22021   uint8x8_t __ret;
   22022   __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
   22023   return __ret;
   22024 }
   22025 #endif
   22026 
   22027 #ifdef __LITTLE_ENDIAN__
   22028 __ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
   22029   int16x4_t __ret;
   22030   __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
   22031   return __ret;
   22032 }
   22033 #else
   22034 __ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
   22035   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   22036   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   22037   int16x4_t __ret;
   22038   __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
   22039   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   22040   return __ret;
   22041 }
   22042 __ai int16x4_t __noswap_vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
   22043   int16x4_t __ret;
   22044   __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
   22045   return __ret;
   22046 }
   22047 #endif
   22048 
   22049 #ifdef __LITTLE_ENDIAN__
   22050 __ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
   22051   int32x2_t __ret;
   22052   __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
   22053   return __ret;
   22054 }
   22055 #else
   22056 __ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
   22057   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   22058   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   22059   int32x2_t __ret;
   22060   __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
   22061   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   22062   return __ret;
   22063 }
   22064 __ai int32x2_t __noswap_vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
   22065   int32x2_t __ret;
   22066   __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
   22067   return __ret;
   22068 }
   22069 #endif
   22070 
   22071 #ifdef __LITTLE_ENDIAN__
   22072 __ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
   22073   int8x8_t __ret;
   22074   __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
   22075   return __ret;
   22076 }
   22077 #else
   22078 __ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
   22079   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   22080   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   22081   int8x8_t __ret;
   22082   __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
   22083   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   22084   return __ret;
   22085 }
   22086 __ai int8x8_t __noswap_vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
   22087   int8x8_t __ret;
   22088   __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
   22089   return __ret;
   22090 }
   22091 #endif
   22092 
   22093 #ifdef __LITTLE_ENDIAN__
   22094 #define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   22095   poly8_t __s0 = __p0; \
   22096   poly8x8_t __s1 = __p1; \
   22097   poly8x8_t __ret; \
   22098   __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
   22099   __ret; \
   22100 })
   22101 #else
   22102 #define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   22103   poly8_t __s0 = __p0; \
   22104   poly8x8_t __s1 = __p1; \
   22105   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   22106   poly8x8_t __ret; \
   22107   __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
   22108   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   22109   __ret; \
   22110 })
   22111 #define __noswap_vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   22112   poly8_t __s0 = __p0; \
   22113   poly8x8_t __s1 = __p1; \
   22114   poly8x8_t __ret; \
   22115   __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
   22116   __ret; \
   22117 })
   22118 #endif
   22119 
   22120 #ifdef __LITTLE_ENDIAN__
   22121 #define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   22122   poly16_t __s0 = __p0; \
   22123   poly16x4_t __s1 = __p1; \
   22124   poly16x4_t __ret; \
   22125   __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
   22126   __ret; \
   22127 })
   22128 #else
   22129 #define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   22130   poly16_t __s0 = __p0; \
   22131   poly16x4_t __s1 = __p1; \
   22132   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   22133   poly16x4_t __ret; \
   22134   __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
   22135   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   22136   __ret; \
   22137 })
   22138 #define __noswap_vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   22139   poly16_t __s0 = __p0; \
   22140   poly16x4_t __s1 = __p1; \
   22141   poly16x4_t __ret; \
   22142   __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
   22143   __ret; \
   22144 })
   22145 #endif
   22146 
   22147 #ifdef __LITTLE_ENDIAN__
   22148 #define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   22149   poly8_t __s0 = __p0; \
   22150   poly8x16_t __s1 = __p1; \
   22151   poly8x16_t __ret; \
   22152   __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
   22153   __ret; \
   22154 })
   22155 #else
   22156 #define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   22157   poly8_t __s0 = __p0; \
   22158   poly8x16_t __s1 = __p1; \
   22159   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   22160   poly8x16_t __ret; \
   22161   __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
   22162   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   22163   __ret; \
   22164 })
   22165 #define __noswap_vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   22166   poly8_t __s0 = __p0; \
   22167   poly8x16_t __s1 = __p1; \
   22168   poly8x16_t __ret; \
   22169   __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
   22170   __ret; \
   22171 })
   22172 #endif
   22173 
   22174 #ifdef __LITTLE_ENDIAN__
   22175 #define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   22176   poly16_t __s0 = __p0; \
   22177   poly16x8_t __s1 = __p1; \
   22178   poly16x8_t __ret; \
   22179   __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
   22180   __ret; \
   22181 })
   22182 #else
   22183 #define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   22184   poly16_t __s0 = __p0; \
   22185   poly16x8_t __s1 = __p1; \
   22186   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   22187   poly16x8_t __ret; \
   22188   __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
   22189   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   22190   __ret; \
   22191 })
   22192 #define __noswap_vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   22193   poly16_t __s0 = __p0; \
   22194   poly16x8_t __s1 = __p1; \
   22195   poly16x8_t __ret; \
   22196   __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
   22197   __ret; \
   22198 })
   22199 #endif
   22200 
   22201 #ifdef __LITTLE_ENDIAN__
   22202 #define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   22203   uint8_t __s0 = __p0; \
   22204   uint8x16_t __s1 = __p1; \
   22205   uint8x16_t __ret; \
   22206   __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
   22207   __ret; \
   22208 })
   22209 #else
   22210 #define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   22211   uint8_t __s0 = __p0; \
   22212   uint8x16_t __s1 = __p1; \
   22213   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   22214   uint8x16_t __ret; \
   22215   __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
   22216   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   22217   __ret; \
   22218 })
   22219 #define __noswap_vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   22220   uint8_t __s0 = __p0; \
   22221   uint8x16_t __s1 = __p1; \
   22222   uint8x16_t __ret; \
   22223   __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
   22224   __ret; \
   22225 })
   22226 #endif
   22227 
   22228 #ifdef __LITTLE_ENDIAN__
   22229 #define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   22230   uint32_t __s0 = __p0; \
   22231   uint32x4_t __s1 = __p1; \
   22232   uint32x4_t __ret; \
   22233   __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
   22234   __ret; \
   22235 })
   22236 #else
   22237 #define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   22238   uint32_t __s0 = __p0; \
   22239   uint32x4_t __s1 = __p1; \
   22240   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   22241   uint32x4_t __ret; \
   22242   __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
   22243   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   22244   __ret; \
   22245 })
   22246 #define __noswap_vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   22247   uint32_t __s0 = __p0; \
   22248   uint32x4_t __s1 = __p1; \
   22249   uint32x4_t __ret; \
   22250   __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
   22251   __ret; \
   22252 })
   22253 #endif
   22254 
   22255 #ifdef __LITTLE_ENDIAN__
   22256 #define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   22257   uint64_t __s0 = __p0; \
   22258   uint64x2_t __s1 = __p1; \
   22259   uint64x2_t __ret; \
   22260   __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
   22261   __ret; \
   22262 })
   22263 #else
   22264 #define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   22265   uint64_t __s0 = __p0; \
   22266   uint64x2_t __s1 = __p1; \
   22267   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   22268   uint64x2_t __ret; \
   22269   __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
   22270   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   22271   __ret; \
   22272 })
   22273 #define __noswap_vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   22274   uint64_t __s0 = __p0; \
   22275   uint64x2_t __s1 = __p1; \
   22276   uint64x2_t __ret; \
   22277   __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
   22278   __ret; \
   22279 })
   22280 #endif
   22281 
   22282 #ifdef __LITTLE_ENDIAN__
   22283 #define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   22284   uint16_t __s0 = __p0; \
   22285   uint16x8_t __s1 = __p1; \
   22286   uint16x8_t __ret; \
   22287   __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
   22288   __ret; \
   22289 })
   22290 #else
   22291 #define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   22292   uint16_t __s0 = __p0; \
   22293   uint16x8_t __s1 = __p1; \
   22294   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   22295   uint16x8_t __ret; \
   22296   __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
   22297   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   22298   __ret; \
   22299 })
   22300 #define __noswap_vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   22301   uint16_t __s0 = __p0; \
   22302   uint16x8_t __s1 = __p1; \
   22303   uint16x8_t __ret; \
   22304   __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
   22305   __ret; \
   22306 })
   22307 #endif
   22308 
   22309 #ifdef __LITTLE_ENDIAN__
   22310 #define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   22311   int8_t __s0 = __p0; \
   22312   int8x16_t __s1 = __p1; \
   22313   int8x16_t __ret; \
   22314   __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
   22315   __ret; \
   22316 })
   22317 #else
   22318 #define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   22319   int8_t __s0 = __p0; \
   22320   int8x16_t __s1 = __p1; \
   22321   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   22322   int8x16_t __ret; \
   22323   __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
   22324   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   22325   __ret; \
   22326 })
   22327 #define __noswap_vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   22328   int8_t __s0 = __p0; \
   22329   int8x16_t __s1 = __p1; \
   22330   int8x16_t __ret; \
   22331   __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
   22332   __ret; \
   22333 })
   22334 #endif
   22335 
   22336 #ifdef __LITTLE_ENDIAN__
   22337 #define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   22338   float32_t __s0 = __p0; \
   22339   float32x4_t __s1 = __p1; \
   22340   float32x4_t __ret; \
   22341   __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
   22342   __ret; \
   22343 })
   22344 #else
   22345 #define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   22346   float32_t __s0 = __p0; \
   22347   float32x4_t __s1 = __p1; \
   22348   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   22349   float32x4_t __ret; \
   22350   __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__rev1, __p2); \
   22351   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   22352   __ret; \
   22353 })
   22354 #define __noswap_vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   22355   float32_t __s0 = __p0; \
   22356   float32x4_t __s1 = __p1; \
   22357   float32x4_t __ret; \
   22358   __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
   22359   __ret; \
   22360 })
   22361 #endif
   22362 
   22363 #ifdef __LITTLE_ENDIAN__
   22364 #define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   22365   int32_t __s0 = __p0; \
   22366   int32x4_t __s1 = __p1; \
   22367   int32x4_t __ret; \
   22368   __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
   22369   __ret; \
   22370 })
   22371 #else
   22372 #define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   22373   int32_t __s0 = __p0; \
   22374   int32x4_t __s1 = __p1; \
   22375   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   22376   int32x4_t __ret; \
   22377   __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
   22378   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   22379   __ret; \
   22380 })
   22381 #define __noswap_vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   22382   int32_t __s0 = __p0; \
   22383   int32x4_t __s1 = __p1; \
   22384   int32x4_t __ret; \
   22385   __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
   22386   __ret; \
   22387 })
   22388 #endif
   22389 
   22390 #ifdef __LITTLE_ENDIAN__
   22391 #define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   22392   int64_t __s0 = __p0; \
   22393   int64x2_t __s1 = __p1; \
   22394   int64x2_t __ret; \
   22395   __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
   22396   __ret; \
   22397 })
   22398 #else
   22399 #define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   22400   int64_t __s0 = __p0; \
   22401   int64x2_t __s1 = __p1; \
   22402   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   22403   int64x2_t __ret; \
   22404   __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
   22405   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   22406   __ret; \
   22407 })
   22408 #define __noswap_vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   22409   int64_t __s0 = __p0; \
   22410   int64x2_t __s1 = __p1; \
   22411   int64x2_t __ret; \
   22412   __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
   22413   __ret; \
   22414 })
   22415 #endif
   22416 
   22417 #ifdef __LITTLE_ENDIAN__
   22418 #define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   22419   int16_t __s0 = __p0; \
   22420   int16x8_t __s1 = __p1; \
   22421   int16x8_t __ret; \
   22422   __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
   22423   __ret; \
   22424 })
   22425 #else
   22426 #define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   22427   int16_t __s0 = __p0; \
   22428   int16x8_t __s1 = __p1; \
   22429   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   22430   int16x8_t __ret; \
   22431   __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
   22432   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   22433   __ret; \
   22434 })
   22435 #define __noswap_vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   22436   int16_t __s0 = __p0; \
   22437   int16x8_t __s1 = __p1; \
   22438   int16x8_t __ret; \
   22439   __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
   22440   __ret; \
   22441 })
   22442 #endif
   22443 
   22444 #ifdef __LITTLE_ENDIAN__
   22445 #define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   22446   uint8_t __s0 = __p0; \
   22447   uint8x8_t __s1 = __p1; \
   22448   uint8x8_t __ret; \
   22449   __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
   22450   __ret; \
   22451 })
   22452 #else
   22453 #define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   22454   uint8_t __s0 = __p0; \
   22455   uint8x8_t __s1 = __p1; \
   22456   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   22457   uint8x8_t __ret; \
   22458   __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
   22459   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   22460   __ret; \
   22461 })
   22462 #define __noswap_vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   22463   uint8_t __s0 = __p0; \
   22464   uint8x8_t __s1 = __p1; \
   22465   uint8x8_t __ret; \
   22466   __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
   22467   __ret; \
   22468 })
   22469 #endif
   22470 
   22471 #ifdef __LITTLE_ENDIAN__
   22472 #define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   22473   uint32_t __s0 = __p0; \
   22474   uint32x2_t __s1 = __p1; \
   22475   uint32x2_t __ret; \
   22476   __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
   22477   __ret; \
   22478 })
   22479 #else
   22480 #define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   22481   uint32_t __s0 = __p0; \
   22482   uint32x2_t __s1 = __p1; \
   22483   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   22484   uint32x2_t __ret; \
   22485   __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
   22486   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   22487   __ret; \
   22488 })
   22489 #define __noswap_vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   22490   uint32_t __s0 = __p0; \
   22491   uint32x2_t __s1 = __p1; \
   22492   uint32x2_t __ret; \
   22493   __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
   22494   __ret; \
   22495 })
   22496 #endif
   22497 
   22498 #ifdef __LITTLE_ENDIAN__
   22499 #define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   22500   uint64_t __s0 = __p0; \
   22501   uint64x1_t __s1 = __p1; \
   22502   uint64x1_t __ret; \
   22503   __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
   22504   __ret; \
   22505 })
   22506 #else
   22507 #define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   22508   uint64_t __s0 = __p0; \
   22509   uint64x1_t __s1 = __p1; \
   22510   uint64x1_t __ret; \
   22511   __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
   22512   __ret; \
   22513 })
   22514 #define __noswap_vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   22515   uint64_t __s0 = __p0; \
   22516   uint64x1_t __s1 = __p1; \
   22517   uint64x1_t __ret; \
   22518   __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
   22519   __ret; \
   22520 })
   22521 #endif
   22522 
   22523 #ifdef __LITTLE_ENDIAN__
   22524 #define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   22525   uint16_t __s0 = __p0; \
   22526   uint16x4_t __s1 = __p1; \
   22527   uint16x4_t __ret; \
   22528   __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
   22529   __ret; \
   22530 })
   22531 #else
   22532 #define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   22533   uint16_t __s0 = __p0; \
   22534   uint16x4_t __s1 = __p1; \
   22535   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   22536   uint16x4_t __ret; \
   22537   __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
   22538   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   22539   __ret; \
   22540 })
   22541 #define __noswap_vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   22542   uint16_t __s0 = __p0; \
   22543   uint16x4_t __s1 = __p1; \
   22544   uint16x4_t __ret; \
   22545   __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
   22546   __ret; \
   22547 })
   22548 #endif
   22549 
   22550 #ifdef __LITTLE_ENDIAN__
   22551 #define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   22552   int8_t __s0 = __p0; \
   22553   int8x8_t __s1 = __p1; \
   22554   int8x8_t __ret; \
   22555   __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
   22556   __ret; \
   22557 })
   22558 #else
   22559 #define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   22560   int8_t __s0 = __p0; \
   22561   int8x8_t __s1 = __p1; \
   22562   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   22563   int8x8_t __ret; \
   22564   __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
   22565   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   22566   __ret; \
   22567 })
   22568 #define __noswap_vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   22569   int8_t __s0 = __p0; \
   22570   int8x8_t __s1 = __p1; \
   22571   int8x8_t __ret; \
   22572   __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
   22573   __ret; \
   22574 })
   22575 #endif
   22576 
   22577 #ifdef __LITTLE_ENDIAN__
   22578 #define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   22579   float32_t __s0 = __p0; \
   22580   float32x2_t __s1 = __p1; \
   22581   float32x2_t __ret; \
   22582   __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
   22583   __ret; \
   22584 })
   22585 #else
   22586 #define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   22587   float32_t __s0 = __p0; \
   22588   float32x2_t __s1 = __p1; \
   22589   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   22590   float32x2_t __ret; \
   22591   __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__rev1, __p2); \
   22592   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   22593   __ret; \
   22594 })
   22595 #define __noswap_vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   22596   float32_t __s0 = __p0; \
   22597   float32x2_t __s1 = __p1; \
   22598   float32x2_t __ret; \
   22599   __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
   22600   __ret; \
   22601 })
   22602 #endif
   22603 
   22604 #ifdef __LITTLE_ENDIAN__
   22605 #define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   22606   int32_t __s0 = __p0; \
   22607   int32x2_t __s1 = __p1; \
   22608   int32x2_t __ret; \
   22609   __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
   22610   __ret; \
   22611 })
   22612 #else
   22613 #define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   22614   int32_t __s0 = __p0; \
   22615   int32x2_t __s1 = __p1; \
   22616   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   22617   int32x2_t __ret; \
   22618   __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
   22619   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   22620   __ret; \
   22621 })
   22622 #define __noswap_vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   22623   int32_t __s0 = __p0; \
   22624   int32x2_t __s1 = __p1; \
   22625   int32x2_t __ret; \
   22626   __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
   22627   __ret; \
   22628 })
   22629 #endif
   22630 
   22631 #ifdef __LITTLE_ENDIAN__
   22632 #define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   22633   int64_t __s0 = __p0; \
   22634   int64x1_t __s1 = __p1; \
   22635   int64x1_t __ret; \
   22636   __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
   22637   __ret; \
   22638 })
   22639 #else
   22640 #define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   22641   int64_t __s0 = __p0; \
   22642   int64x1_t __s1 = __p1; \
   22643   int64x1_t __ret; \
   22644   __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
   22645   __ret; \
   22646 })
   22647 #define __noswap_vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   22648   int64_t __s0 = __p0; \
   22649   int64x1_t __s1 = __p1; \
   22650   int64x1_t __ret; \
   22651   __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
   22652   __ret; \
   22653 })
   22654 #endif
   22655 
   22656 #ifdef __LITTLE_ENDIAN__
   22657 #define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   22658   int16_t __s0 = __p0; \
   22659   int16x4_t __s1 = __p1; \
   22660   int16x4_t __ret; \
   22661   __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
   22662   __ret; \
   22663 })
   22664 #else
   22665 #define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   22666   int16_t __s0 = __p0; \
   22667   int16x4_t __s1 = __p1; \
   22668   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   22669   int16x4_t __ret; \
   22670   __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
   22671   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   22672   __ret; \
   22673 })
   22674 #define __noswap_vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   22675   int16_t __s0 = __p0; \
   22676   int16x4_t __s1 = __p1; \
   22677   int16x4_t __ret; \
   22678   __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
   22679   __ret; \
   22680 })
   22681 #endif
   22682 
   22683 #ifdef __LITTLE_ENDIAN__
   22684 __ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
   22685   uint8x16_t __ret;
   22686   __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   22687   return __ret;
   22688 }
   22689 #else
   22690 __ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
   22691   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   22692   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   22693   uint8x16_t __ret;
   22694   __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   22695   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   22696   return __ret;
   22697 }
   22698 #endif
   22699 
   22700 #ifdef __LITTLE_ENDIAN__
   22701 __ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
   22702   uint32x4_t __ret;
   22703   __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   22704   return __ret;
   22705 }
   22706 #else
   22707 __ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
   22708   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   22709   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   22710   uint32x4_t __ret;
   22711   __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   22712   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   22713   return __ret;
   22714 }
   22715 #endif
   22716 
   22717 #ifdef __LITTLE_ENDIAN__
   22718 __ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
   22719   uint64x2_t __ret;
   22720   __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   22721   return __ret;
   22722 }
   22723 #else
   22724 __ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
   22725   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   22726   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   22727   uint64x2_t __ret;
   22728   __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   22729   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   22730   return __ret;
   22731 }
   22732 #endif
   22733 
   22734 #ifdef __LITTLE_ENDIAN__
   22735 __ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
   22736   uint16x8_t __ret;
   22737   __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   22738   return __ret;
   22739 }
   22740 #else
   22741 __ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
   22742   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   22743   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   22744   uint16x8_t __ret;
   22745   __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   22746   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   22747   return __ret;
   22748 }
   22749 #endif
   22750 
   22751 #ifdef __LITTLE_ENDIAN__
   22752 __ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
   22753   int8x16_t __ret;
   22754   __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   22755   return __ret;
   22756 }
   22757 #else
   22758 __ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
   22759   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   22760   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   22761   int8x16_t __ret;
   22762   __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   22763   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   22764   return __ret;
   22765 }
   22766 #endif
   22767 
   22768 #ifdef __LITTLE_ENDIAN__
   22769 __ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
   22770   int32x4_t __ret;
   22771   __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   22772   return __ret;
   22773 }
   22774 #else
   22775 __ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
   22776   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   22777   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   22778   int32x4_t __ret;
   22779   __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   22780   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   22781   return __ret;
   22782 }
   22783 #endif
   22784 
   22785 #ifdef __LITTLE_ENDIAN__
   22786 __ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
   22787   int64x2_t __ret;
   22788   __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
   22789   return __ret;
   22790 }
   22791 #else
   22792 __ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
   22793   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   22794   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   22795   int64x2_t __ret;
   22796   __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
   22797   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   22798   return __ret;
   22799 }
   22800 #endif
   22801 
   22802 #ifdef __LITTLE_ENDIAN__
   22803 __ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
   22804   int16x8_t __ret;
   22805   __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   22806   return __ret;
   22807 }
   22808 #else
   22809 __ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
   22810   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   22811   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   22812   int16x8_t __ret;
   22813   __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   22814   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   22815   return __ret;
   22816 }
   22817 #endif
   22818 
   22819 #ifdef __LITTLE_ENDIAN__
   22820 __ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
   22821   uint8x8_t __ret;
   22822   __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   22823   return __ret;
   22824 }
   22825 #else
   22826 __ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
   22827   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   22828   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   22829   uint8x8_t __ret;
   22830   __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   22831   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   22832   return __ret;
   22833 }
   22834 #endif
   22835 
   22836 #ifdef __LITTLE_ENDIAN__
   22837 __ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
   22838   uint32x2_t __ret;
   22839   __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   22840   return __ret;
   22841 }
   22842 #else
   22843 __ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
   22844   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   22845   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   22846   uint32x2_t __ret;
   22847   __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   22848   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   22849   return __ret;
   22850 }
   22851 #endif
   22852 
   22853 #ifdef __LITTLE_ENDIAN__
   22854 __ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
   22855   uint64x1_t __ret;
   22856   __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   22857   return __ret;
   22858 }
   22859 #else
   22860 __ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
   22861   uint64x1_t __ret;
   22862   __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   22863   return __ret;
   22864 }
   22865 #endif
   22866 
   22867 #ifdef __LITTLE_ENDIAN__
   22868 __ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
   22869   uint16x4_t __ret;
   22870   __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   22871   return __ret;
   22872 }
   22873 #else
   22874 __ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
   22875   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   22876   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   22877   uint16x4_t __ret;
   22878   __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   22879   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   22880   return __ret;
   22881 }
   22882 #endif
   22883 
   22884 #ifdef __LITTLE_ENDIAN__
   22885 __ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
   22886   int8x8_t __ret;
   22887   __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   22888   return __ret;
   22889 }
   22890 #else
   22891 __ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
   22892   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   22893   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   22894   int8x8_t __ret;
   22895   __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   22896   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   22897   return __ret;
   22898 }
   22899 #endif
   22900 
   22901 #ifdef __LITTLE_ENDIAN__
   22902 __ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
   22903   int32x2_t __ret;
   22904   __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   22905   return __ret;
   22906 }
   22907 #else
   22908 __ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
   22909   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   22910   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   22911   int32x2_t __ret;
   22912   __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   22913   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   22914   return __ret;
   22915 }
   22916 #endif
   22917 
   22918 #ifdef __LITTLE_ENDIAN__
   22919 __ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
   22920   int64x1_t __ret;
   22921   __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   22922   return __ret;
   22923 }
   22924 #else
   22925 __ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
   22926   int64x1_t __ret;
   22927   __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   22928   return __ret;
   22929 }
   22930 #endif
   22931 
   22932 #ifdef __LITTLE_ENDIAN__
   22933 __ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
   22934   int16x4_t __ret;
   22935   __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   22936   return __ret;
   22937 }
   22938 #else
   22939 __ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
   22940   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   22941   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   22942   int16x4_t __ret;
   22943   __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   22944   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   22945   return __ret;
   22946 }
   22947 #endif
   22948 
   22949 #ifdef __LITTLE_ENDIAN__
   22950 #define vshlq_n_u8(__p0, __p1) __extension__ ({ \
   22951   uint8x16_t __s0 = __p0; \
   22952   uint8x16_t __ret; \
   22953   __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 48); \
   22954   __ret; \
   22955 })
   22956 #else
   22957 #define vshlq_n_u8(__p0, __p1) __extension__ ({ \
   22958   uint8x16_t __s0 = __p0; \
   22959   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   22960   uint8x16_t __ret; \
   22961   __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 48); \
   22962   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   22963   __ret; \
   22964 })
   22965 #endif
   22966 
   22967 #ifdef __LITTLE_ENDIAN__
   22968 #define vshlq_n_u32(__p0, __p1) __extension__ ({ \
   22969   uint32x4_t __s0 = __p0; \
   22970   uint32x4_t __ret; \
   22971   __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 50); \
   22972   __ret; \
   22973 })
   22974 #else
   22975 #define vshlq_n_u32(__p0, __p1) __extension__ ({ \
   22976   uint32x4_t __s0 = __p0; \
   22977   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   22978   uint32x4_t __ret; \
   22979   __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 50); \
   22980   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   22981   __ret; \
   22982 })
   22983 #endif
   22984 
   22985 #ifdef __LITTLE_ENDIAN__
   22986 #define vshlq_n_u64(__p0, __p1) __extension__ ({ \
   22987   uint64x2_t __s0 = __p0; \
   22988   uint64x2_t __ret; \
   22989   __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 51); \
   22990   __ret; \
   22991 })
   22992 #else
   22993 #define vshlq_n_u64(__p0, __p1) __extension__ ({ \
   22994   uint64x2_t __s0 = __p0; \
   22995   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   22996   uint64x2_t __ret; \
   22997   __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 51); \
   22998   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   22999   __ret; \
   23000 })
   23001 #endif
   23002 
   23003 #ifdef __LITTLE_ENDIAN__
   23004 #define vshlq_n_u16(__p0, __p1) __extension__ ({ \
   23005   uint16x8_t __s0 = __p0; \
   23006   uint16x8_t __ret; \
   23007   __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 49); \
   23008   __ret; \
   23009 })
   23010 #else
   23011 #define vshlq_n_u16(__p0, __p1) __extension__ ({ \
   23012   uint16x8_t __s0 = __p0; \
   23013   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23014   uint16x8_t __ret; \
   23015   __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 49); \
   23016   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23017   __ret; \
   23018 })
   23019 #endif
   23020 
   23021 #ifdef __LITTLE_ENDIAN__
   23022 #define vshlq_n_s8(__p0, __p1) __extension__ ({ \
   23023   int8x16_t __s0 = __p0; \
   23024   int8x16_t __ret; \
   23025   __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 32); \
   23026   __ret; \
   23027 })
   23028 #else
   23029 #define vshlq_n_s8(__p0, __p1) __extension__ ({ \
   23030   int8x16_t __s0 = __p0; \
   23031   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23032   int8x16_t __ret; \
   23033   __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 32); \
   23034   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23035   __ret; \
   23036 })
   23037 #endif
   23038 
   23039 #ifdef __LITTLE_ENDIAN__
   23040 #define vshlq_n_s32(__p0, __p1) __extension__ ({ \
   23041   int32x4_t __s0 = __p0; \
   23042   int32x4_t __ret; \
   23043   __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 34); \
   23044   __ret; \
   23045 })
   23046 #else
   23047 #define vshlq_n_s32(__p0, __p1) __extension__ ({ \
   23048   int32x4_t __s0 = __p0; \
   23049   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23050   int32x4_t __ret; \
   23051   __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 34); \
   23052   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23053   __ret; \
   23054 })
   23055 #endif
   23056 
   23057 #ifdef __LITTLE_ENDIAN__
   23058 #define vshlq_n_s64(__p0, __p1) __extension__ ({ \
   23059   int64x2_t __s0 = __p0; \
   23060   int64x2_t __ret; \
   23061   __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 35); \
   23062   __ret; \
   23063 })
   23064 #else
   23065 #define vshlq_n_s64(__p0, __p1) __extension__ ({ \
   23066   int64x2_t __s0 = __p0; \
   23067   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23068   int64x2_t __ret; \
   23069   __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 35); \
   23070   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23071   __ret; \
   23072 })
   23073 #endif
   23074 
   23075 #ifdef __LITTLE_ENDIAN__
   23076 #define vshlq_n_s16(__p0, __p1) __extension__ ({ \
   23077   int16x8_t __s0 = __p0; \
   23078   int16x8_t __ret; \
   23079   __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 33); \
   23080   __ret; \
   23081 })
   23082 #else
   23083 #define vshlq_n_s16(__p0, __p1) __extension__ ({ \
   23084   int16x8_t __s0 = __p0; \
   23085   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23086   int16x8_t __ret; \
   23087   __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 33); \
   23088   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23089   __ret; \
   23090 })
   23091 #endif
   23092 
   23093 #ifdef __LITTLE_ENDIAN__
   23094 #define vshl_n_u8(__p0, __p1) __extension__ ({ \
   23095   uint8x8_t __s0 = __p0; \
   23096   uint8x8_t __ret; \
   23097   __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 16); \
   23098   __ret; \
   23099 })
   23100 #else
   23101 #define vshl_n_u8(__p0, __p1) __extension__ ({ \
   23102   uint8x8_t __s0 = __p0; \
   23103   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23104   uint8x8_t __ret; \
   23105   __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 16); \
   23106   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23107   __ret; \
   23108 })
   23109 #endif
   23110 
   23111 #ifdef __LITTLE_ENDIAN__
   23112 #define vshl_n_u32(__p0, __p1) __extension__ ({ \
   23113   uint32x2_t __s0 = __p0; \
   23114   uint32x2_t __ret; \
   23115   __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 18); \
   23116   __ret; \
   23117 })
   23118 #else
   23119 #define vshl_n_u32(__p0, __p1) __extension__ ({ \
   23120   uint32x2_t __s0 = __p0; \
   23121   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23122   uint32x2_t __ret; \
   23123   __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 18); \
   23124   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23125   __ret; \
   23126 })
   23127 #endif
   23128 
   23129 #ifdef __LITTLE_ENDIAN__
   23130 #define vshl_n_u64(__p0, __p1) __extension__ ({ \
   23131   uint64x1_t __s0 = __p0; \
   23132   uint64x1_t __ret; \
   23133   __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
   23134   __ret; \
   23135 })
   23136 #else
   23137 #define vshl_n_u64(__p0, __p1) __extension__ ({ \
   23138   uint64x1_t __s0 = __p0; \
   23139   uint64x1_t __ret; \
   23140   __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
   23141   __ret; \
   23142 })
   23143 #endif
   23144 
   23145 #ifdef __LITTLE_ENDIAN__
   23146 #define vshl_n_u16(__p0, __p1) __extension__ ({ \
   23147   uint16x4_t __s0 = __p0; \
   23148   uint16x4_t __ret; \
   23149   __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 17); \
   23150   __ret; \
   23151 })
   23152 #else
   23153 #define vshl_n_u16(__p0, __p1) __extension__ ({ \
   23154   uint16x4_t __s0 = __p0; \
   23155   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23156   uint16x4_t __ret; \
   23157   __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 17); \
   23158   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23159   __ret; \
   23160 })
   23161 #endif
   23162 
   23163 #ifdef __LITTLE_ENDIAN__
   23164 #define vshl_n_s8(__p0, __p1) __extension__ ({ \
   23165   int8x8_t __s0 = __p0; \
   23166   int8x8_t __ret; \
   23167   __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 0); \
   23168   __ret; \
   23169 })
   23170 #else
   23171 #define vshl_n_s8(__p0, __p1) __extension__ ({ \
   23172   int8x8_t __s0 = __p0; \
   23173   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23174   int8x8_t __ret; \
   23175   __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 0); \
   23176   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23177   __ret; \
   23178 })
   23179 #endif
   23180 
   23181 #ifdef __LITTLE_ENDIAN__
   23182 #define vshl_n_s32(__p0, __p1) __extension__ ({ \
   23183   int32x2_t __s0 = __p0; \
   23184   int32x2_t __ret; \
   23185   __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 2); \
   23186   __ret; \
   23187 })
   23188 #else
   23189 #define vshl_n_s32(__p0, __p1) __extension__ ({ \
   23190   int32x2_t __s0 = __p0; \
   23191   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23192   int32x2_t __ret; \
   23193   __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 2); \
   23194   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23195   __ret; \
   23196 })
   23197 #endif
   23198 
   23199 #ifdef __LITTLE_ENDIAN__
   23200 #define vshl_n_s64(__p0, __p1) __extension__ ({ \
   23201   int64x1_t __s0 = __p0; \
   23202   int64x1_t __ret; \
   23203   __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
   23204   __ret; \
   23205 })
   23206 #else
   23207 #define vshl_n_s64(__p0, __p1) __extension__ ({ \
   23208   int64x1_t __s0 = __p0; \
   23209   int64x1_t __ret; \
   23210   __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
   23211   __ret; \
   23212 })
   23213 #endif
   23214 
   23215 #ifdef __LITTLE_ENDIAN__
   23216 #define vshl_n_s16(__p0, __p1) __extension__ ({ \
   23217   int16x4_t __s0 = __p0; \
   23218   int16x4_t __ret; \
   23219   __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 1); \
   23220   __ret; \
   23221 })
   23222 #else
   23223 #define vshl_n_s16(__p0, __p1) __extension__ ({ \
   23224   int16x4_t __s0 = __p0; \
   23225   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23226   int16x4_t __ret; \
   23227   __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 1); \
   23228   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23229   __ret; \
   23230 })
   23231 #endif
   23232 
   23233 #ifdef __LITTLE_ENDIAN__
   23234 #define vshll_n_u8(__p0, __p1) __extension__ ({ \
   23235   uint8x8_t __s0 = __p0; \
   23236   uint16x8_t __ret; \
   23237   __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
   23238   __ret; \
   23239 })
   23240 #else
   23241 #define vshll_n_u8(__p0, __p1) __extension__ ({ \
   23242   uint8x8_t __s0 = __p0; \
   23243   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23244   uint16x8_t __ret; \
   23245   __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 49); \
   23246   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23247   __ret; \
   23248 })
   23249 #define __noswap_vshll_n_u8(__p0, __p1) __extension__ ({ \
   23250   uint8x8_t __s0 = __p0; \
   23251   uint16x8_t __ret; \
   23252   __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
   23253   __ret; \
   23254 })
   23255 #endif
   23256 
   23257 #ifdef __LITTLE_ENDIAN__
   23258 #define vshll_n_u32(__p0, __p1) __extension__ ({ \
   23259   uint32x2_t __s0 = __p0; \
   23260   uint64x2_t __ret; \
   23261   __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
   23262   __ret; \
   23263 })
   23264 #else
   23265 #define vshll_n_u32(__p0, __p1) __extension__ ({ \
   23266   uint32x2_t __s0 = __p0; \
   23267   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23268   uint64x2_t __ret; \
   23269   __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 51); \
   23270   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23271   __ret; \
   23272 })
   23273 #define __noswap_vshll_n_u32(__p0, __p1) __extension__ ({ \
   23274   uint32x2_t __s0 = __p0; \
   23275   uint64x2_t __ret; \
   23276   __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
   23277   __ret; \
   23278 })
   23279 #endif
   23280 
   23281 #ifdef __LITTLE_ENDIAN__
   23282 #define vshll_n_u16(__p0, __p1) __extension__ ({ \
   23283   uint16x4_t __s0 = __p0; \
   23284   uint32x4_t __ret; \
   23285   __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
   23286   __ret; \
   23287 })
   23288 #else
   23289 #define vshll_n_u16(__p0, __p1) __extension__ ({ \
   23290   uint16x4_t __s0 = __p0; \
   23291   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23292   uint32x4_t __ret; \
   23293   __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 50); \
   23294   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23295   __ret; \
   23296 })
   23297 #define __noswap_vshll_n_u16(__p0, __p1) __extension__ ({ \
   23298   uint16x4_t __s0 = __p0; \
   23299   uint32x4_t __ret; \
   23300   __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
   23301   __ret; \
   23302 })
   23303 #endif
   23304 
   23305 #ifdef __LITTLE_ENDIAN__
   23306 #define vshll_n_s8(__p0, __p1) __extension__ ({ \
   23307   int8x8_t __s0 = __p0; \
   23308   int16x8_t __ret; \
   23309   __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
   23310   __ret; \
   23311 })
   23312 #else
   23313 #define vshll_n_s8(__p0, __p1) __extension__ ({ \
   23314   int8x8_t __s0 = __p0; \
   23315   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23316   int16x8_t __ret; \
   23317   __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 33); \
   23318   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23319   __ret; \
   23320 })
   23321 #define __noswap_vshll_n_s8(__p0, __p1) __extension__ ({ \
   23322   int8x8_t __s0 = __p0; \
   23323   int16x8_t __ret; \
   23324   __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
   23325   __ret; \
   23326 })
   23327 #endif
   23328 
   23329 #ifdef __LITTLE_ENDIAN__
   23330 #define vshll_n_s32(__p0, __p1) __extension__ ({ \
   23331   int32x2_t __s0 = __p0; \
   23332   int64x2_t __ret; \
   23333   __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
   23334   __ret; \
   23335 })
   23336 #else
   23337 #define vshll_n_s32(__p0, __p1) __extension__ ({ \
   23338   int32x2_t __s0 = __p0; \
   23339   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23340   int64x2_t __ret; \
   23341   __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 35); \
   23342   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23343   __ret; \
   23344 })
   23345 #define __noswap_vshll_n_s32(__p0, __p1) __extension__ ({ \
   23346   int32x2_t __s0 = __p0; \
   23347   int64x2_t __ret; \
   23348   __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
   23349   __ret; \
   23350 })
   23351 #endif
   23352 
   23353 #ifdef __LITTLE_ENDIAN__
   23354 #define vshll_n_s16(__p0, __p1) __extension__ ({ \
   23355   int16x4_t __s0 = __p0; \
   23356   int32x4_t __ret; \
   23357   __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
   23358   __ret; \
   23359 })
   23360 #else
   23361 #define vshll_n_s16(__p0, __p1) __extension__ ({ \
   23362   int16x4_t __s0 = __p0; \
   23363   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23364   int32x4_t __ret; \
   23365   __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 34); \
   23366   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23367   __ret; \
   23368 })
   23369 #define __noswap_vshll_n_s16(__p0, __p1) __extension__ ({ \
   23370   int16x4_t __s0 = __p0; \
   23371   int32x4_t __ret; \
   23372   __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
   23373   __ret; \
   23374 })
   23375 #endif
   23376 
   23377 #ifdef __LITTLE_ENDIAN__
   23378 #define vshrq_n_u8(__p0, __p1) __extension__ ({ \
   23379   uint8x16_t __s0 = __p0; \
   23380   uint8x16_t __ret; \
   23381   __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 48); \
   23382   __ret; \
   23383 })
   23384 #else
   23385 #define vshrq_n_u8(__p0, __p1) __extension__ ({ \
   23386   uint8x16_t __s0 = __p0; \
   23387   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23388   uint8x16_t __ret; \
   23389   __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 48); \
   23390   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23391   __ret; \
   23392 })
   23393 #endif
   23394 
   23395 #ifdef __LITTLE_ENDIAN__
   23396 #define vshrq_n_u32(__p0, __p1) __extension__ ({ \
   23397   uint32x4_t __s0 = __p0; \
   23398   uint32x4_t __ret; \
   23399   __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 50); \
   23400   __ret; \
   23401 })
   23402 #else
   23403 #define vshrq_n_u32(__p0, __p1) __extension__ ({ \
   23404   uint32x4_t __s0 = __p0; \
   23405   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23406   uint32x4_t __ret; \
   23407   __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 50); \
   23408   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23409   __ret; \
   23410 })
   23411 #endif
   23412 
   23413 #ifdef __LITTLE_ENDIAN__
   23414 #define vshrq_n_u64(__p0, __p1) __extension__ ({ \
   23415   uint64x2_t __s0 = __p0; \
   23416   uint64x2_t __ret; \
   23417   __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 51); \
   23418   __ret; \
   23419 })
   23420 #else
   23421 #define vshrq_n_u64(__p0, __p1) __extension__ ({ \
   23422   uint64x2_t __s0 = __p0; \
   23423   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23424   uint64x2_t __ret; \
   23425   __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 51); \
   23426   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23427   __ret; \
   23428 })
   23429 #endif
   23430 
   23431 #ifdef __LITTLE_ENDIAN__
   23432 #define vshrq_n_u16(__p0, __p1) __extension__ ({ \
   23433   uint16x8_t __s0 = __p0; \
   23434   uint16x8_t __ret; \
   23435   __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 49); \
   23436   __ret; \
   23437 })
   23438 #else
   23439 #define vshrq_n_u16(__p0, __p1) __extension__ ({ \
   23440   uint16x8_t __s0 = __p0; \
   23441   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23442   uint16x8_t __ret; \
   23443   __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 49); \
   23444   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23445   __ret; \
   23446 })
   23447 #endif
   23448 
   23449 #ifdef __LITTLE_ENDIAN__
   23450 #define vshrq_n_s8(__p0, __p1) __extension__ ({ \
   23451   int8x16_t __s0 = __p0; \
   23452   int8x16_t __ret; \
   23453   __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 32); \
   23454   __ret; \
   23455 })
   23456 #else
   23457 #define vshrq_n_s8(__p0, __p1) __extension__ ({ \
   23458   int8x16_t __s0 = __p0; \
   23459   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23460   int8x16_t __ret; \
   23461   __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 32); \
   23462   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23463   __ret; \
   23464 })
   23465 #endif
   23466 
   23467 #ifdef __LITTLE_ENDIAN__
   23468 #define vshrq_n_s32(__p0, __p1) __extension__ ({ \
   23469   int32x4_t __s0 = __p0; \
   23470   int32x4_t __ret; \
   23471   __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 34); \
   23472   __ret; \
   23473 })
   23474 #else
   23475 #define vshrq_n_s32(__p0, __p1) __extension__ ({ \
   23476   int32x4_t __s0 = __p0; \
   23477   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23478   int32x4_t __ret; \
   23479   __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 34); \
   23480   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23481   __ret; \
   23482 })
   23483 #endif
   23484 
   23485 #ifdef __LITTLE_ENDIAN__
   23486 #define vshrq_n_s64(__p0, __p1) __extension__ ({ \
   23487   int64x2_t __s0 = __p0; \
   23488   int64x2_t __ret; \
   23489   __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 35); \
   23490   __ret; \
   23491 })
   23492 #else
   23493 #define vshrq_n_s64(__p0, __p1) __extension__ ({ \
   23494   int64x2_t __s0 = __p0; \
   23495   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23496   int64x2_t __ret; \
   23497   __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 35); \
   23498   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23499   __ret; \
   23500 })
   23501 #endif
   23502 
   23503 #ifdef __LITTLE_ENDIAN__
   23504 #define vshrq_n_s16(__p0, __p1) __extension__ ({ \
   23505   int16x8_t __s0 = __p0; \
   23506   int16x8_t __ret; \
   23507   __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 33); \
   23508   __ret; \
   23509 })
   23510 #else
   23511 #define vshrq_n_s16(__p0, __p1) __extension__ ({ \
   23512   int16x8_t __s0 = __p0; \
   23513   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23514   int16x8_t __ret; \
   23515   __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 33); \
   23516   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23517   __ret; \
   23518 })
   23519 #endif
   23520 
   23521 #ifdef __LITTLE_ENDIAN__
   23522 #define vshr_n_u8(__p0, __p1) __extension__ ({ \
   23523   uint8x8_t __s0 = __p0; \
   23524   uint8x8_t __ret; \
   23525   __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 16); \
   23526   __ret; \
   23527 })
   23528 #else
   23529 #define vshr_n_u8(__p0, __p1) __extension__ ({ \
   23530   uint8x8_t __s0 = __p0; \
   23531   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23532   uint8x8_t __ret; \
   23533   __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 16); \
   23534   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23535   __ret; \
   23536 })
   23537 #endif
   23538 
   23539 #ifdef __LITTLE_ENDIAN__
   23540 #define vshr_n_u32(__p0, __p1) __extension__ ({ \
   23541   uint32x2_t __s0 = __p0; \
   23542   uint32x2_t __ret; \
   23543   __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 18); \
   23544   __ret; \
   23545 })
   23546 #else
   23547 #define vshr_n_u32(__p0, __p1) __extension__ ({ \
   23548   uint32x2_t __s0 = __p0; \
   23549   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23550   uint32x2_t __ret; \
   23551   __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 18); \
   23552   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23553   __ret; \
   23554 })
   23555 #endif
   23556 
   23557 #ifdef __LITTLE_ENDIAN__
   23558 #define vshr_n_u64(__p0, __p1) __extension__ ({ \
   23559   uint64x1_t __s0 = __p0; \
   23560   uint64x1_t __ret; \
   23561   __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
   23562   __ret; \
   23563 })
   23564 #else
   23565 #define vshr_n_u64(__p0, __p1) __extension__ ({ \
   23566   uint64x1_t __s0 = __p0; \
   23567   uint64x1_t __ret; \
   23568   __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
   23569   __ret; \
   23570 })
   23571 #endif
   23572 
   23573 #ifdef __LITTLE_ENDIAN__
   23574 #define vshr_n_u16(__p0, __p1) __extension__ ({ \
   23575   uint16x4_t __s0 = __p0; \
   23576   uint16x4_t __ret; \
   23577   __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 17); \
   23578   __ret; \
   23579 })
   23580 #else
   23581 #define vshr_n_u16(__p0, __p1) __extension__ ({ \
   23582   uint16x4_t __s0 = __p0; \
   23583   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23584   uint16x4_t __ret; \
   23585   __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 17); \
   23586   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23587   __ret; \
   23588 })
   23589 #endif
   23590 
   23591 #ifdef __LITTLE_ENDIAN__
   23592 #define vshr_n_s8(__p0, __p1) __extension__ ({ \
   23593   int8x8_t __s0 = __p0; \
   23594   int8x8_t __ret; \
   23595   __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 0); \
   23596   __ret; \
   23597 })
   23598 #else
   23599 #define vshr_n_s8(__p0, __p1) __extension__ ({ \
   23600   int8x8_t __s0 = __p0; \
   23601   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23602   int8x8_t __ret; \
   23603   __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 0); \
   23604   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23605   __ret; \
   23606 })
   23607 #endif
   23608 
   23609 #ifdef __LITTLE_ENDIAN__
   23610 #define vshr_n_s32(__p0, __p1) __extension__ ({ \
   23611   int32x2_t __s0 = __p0; \
   23612   int32x2_t __ret; \
   23613   __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 2); \
   23614   __ret; \
   23615 })
   23616 #else
   23617 #define vshr_n_s32(__p0, __p1) __extension__ ({ \
   23618   int32x2_t __s0 = __p0; \
   23619   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23620   int32x2_t __ret; \
   23621   __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 2); \
   23622   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23623   __ret; \
   23624 })
   23625 #endif
   23626 
   23627 #ifdef __LITTLE_ENDIAN__
   23628 #define vshr_n_s64(__p0, __p1) __extension__ ({ \
   23629   int64x1_t __s0 = __p0; \
   23630   int64x1_t __ret; \
   23631   __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
   23632   __ret; \
   23633 })
   23634 #else
   23635 #define vshr_n_s64(__p0, __p1) __extension__ ({ \
   23636   int64x1_t __s0 = __p0; \
   23637   int64x1_t __ret; \
   23638   __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
   23639   __ret; \
   23640 })
   23641 #endif
   23642 
   23643 #ifdef __LITTLE_ENDIAN__
   23644 #define vshr_n_s16(__p0, __p1) __extension__ ({ \
   23645   int16x4_t __s0 = __p0; \
   23646   int16x4_t __ret; \
   23647   __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 1); \
   23648   __ret; \
   23649 })
   23650 #else
   23651 #define vshr_n_s16(__p0, __p1) __extension__ ({ \
   23652   int16x4_t __s0 = __p0; \
   23653   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23654   int16x4_t __ret; \
   23655   __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 1); \
   23656   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23657   __ret; \
   23658 })
   23659 #endif
   23660 
   23661 #ifdef __LITTLE_ENDIAN__
   23662 #define vshrn_n_u32(__p0, __p1) __extension__ ({ \
   23663   uint32x4_t __s0 = __p0; \
   23664   uint16x4_t __ret; \
   23665   __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
   23666   __ret; \
   23667 })
   23668 #else
   23669 #define vshrn_n_u32(__p0, __p1) __extension__ ({ \
   23670   uint32x4_t __s0 = __p0; \
   23671   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23672   uint16x4_t __ret; \
   23673   __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 17); \
   23674   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23675   __ret; \
   23676 })
   23677 #define __noswap_vshrn_n_u32(__p0, __p1) __extension__ ({ \
   23678   uint32x4_t __s0 = __p0; \
   23679   uint16x4_t __ret; \
   23680   __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
   23681   __ret; \
   23682 })
   23683 #endif
   23684 
   23685 #ifdef __LITTLE_ENDIAN__
   23686 #define vshrn_n_u64(__p0, __p1) __extension__ ({ \
   23687   uint64x2_t __s0 = __p0; \
   23688   uint32x2_t __ret; \
   23689   __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
   23690   __ret; \
   23691 })
   23692 #else
   23693 #define vshrn_n_u64(__p0, __p1) __extension__ ({ \
   23694   uint64x2_t __s0 = __p0; \
   23695   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23696   uint32x2_t __ret; \
   23697   __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 18); \
   23698   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23699   __ret; \
   23700 })
   23701 #define __noswap_vshrn_n_u64(__p0, __p1) __extension__ ({ \
   23702   uint64x2_t __s0 = __p0; \
   23703   uint32x2_t __ret; \
   23704   __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
   23705   __ret; \
   23706 })
   23707 #endif
   23708 
   23709 #ifdef __LITTLE_ENDIAN__
   23710 #define vshrn_n_u16(__p0, __p1) __extension__ ({ \
   23711   uint16x8_t __s0 = __p0; \
   23712   uint8x8_t __ret; \
   23713   __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
   23714   __ret; \
   23715 })
   23716 #else
   23717 #define vshrn_n_u16(__p0, __p1) __extension__ ({ \
   23718   uint16x8_t __s0 = __p0; \
   23719   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23720   uint8x8_t __ret; \
   23721   __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 16); \
   23722   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23723   __ret; \
   23724 })
   23725 #define __noswap_vshrn_n_u16(__p0, __p1) __extension__ ({ \
   23726   uint16x8_t __s0 = __p0; \
   23727   uint8x8_t __ret; \
   23728   __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
   23729   __ret; \
   23730 })
   23731 #endif
   23732 
   23733 #ifdef __LITTLE_ENDIAN__
   23734 #define vshrn_n_s32(__p0, __p1) __extension__ ({ \
   23735   int32x4_t __s0 = __p0; \
   23736   int16x4_t __ret; \
   23737   __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
   23738   __ret; \
   23739 })
   23740 #else
   23741 #define vshrn_n_s32(__p0, __p1) __extension__ ({ \
   23742   int32x4_t __s0 = __p0; \
   23743   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23744   int16x4_t __ret; \
   23745   __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 1); \
   23746   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23747   __ret; \
   23748 })
   23749 #define __noswap_vshrn_n_s32(__p0, __p1) __extension__ ({ \
   23750   int32x4_t __s0 = __p0; \
   23751   int16x4_t __ret; \
   23752   __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
   23753   __ret; \
   23754 })
   23755 #endif
   23756 
   23757 #ifdef __LITTLE_ENDIAN__
   23758 #define vshrn_n_s64(__p0, __p1) __extension__ ({ \
   23759   int64x2_t __s0 = __p0; \
   23760   int32x2_t __ret; \
   23761   __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
   23762   __ret; \
   23763 })
   23764 #else
   23765 #define vshrn_n_s64(__p0, __p1) __extension__ ({ \
   23766   int64x2_t __s0 = __p0; \
   23767   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23768   int32x2_t __ret; \
   23769   __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 2); \
   23770   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23771   __ret; \
   23772 })
   23773 #define __noswap_vshrn_n_s64(__p0, __p1) __extension__ ({ \
   23774   int64x2_t __s0 = __p0; \
   23775   int32x2_t __ret; \
   23776   __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
   23777   __ret; \
   23778 })
   23779 #endif
   23780 
   23781 #ifdef __LITTLE_ENDIAN__
   23782 #define vshrn_n_s16(__p0, __p1) __extension__ ({ \
   23783   int16x8_t __s0 = __p0; \
   23784   int8x8_t __ret; \
   23785   __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
   23786   __ret; \
   23787 })
   23788 #else
   23789 #define vshrn_n_s16(__p0, __p1) __extension__ ({ \
   23790   int16x8_t __s0 = __p0; \
   23791   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23792   int8x8_t __ret; \
   23793   __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 0); \
   23794   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23795   __ret; \
   23796 })
   23797 #define __noswap_vshrn_n_s16(__p0, __p1) __extension__ ({ \
   23798   int16x8_t __s0 = __p0; \
   23799   int8x8_t __ret; \
   23800   __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
   23801   __ret; \
   23802 })
   23803 #endif
   23804 
   23805 #ifdef __LITTLE_ENDIAN__
   23806 #define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
   23807   poly8x8_t __s0 = __p0; \
   23808   poly8x8_t __s1 = __p1; \
   23809   poly8x8_t __ret; \
   23810   __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
   23811   __ret; \
   23812 })
   23813 #else
   23814 #define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
   23815   poly8x8_t __s0 = __p0; \
   23816   poly8x8_t __s1 = __p1; \
   23817   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23818   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   23819   poly8x8_t __ret; \
   23820   __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
   23821   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23822   __ret; \
   23823 })
   23824 #endif
   23825 
   23826 #ifdef __LITTLE_ENDIAN__
   23827 #define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
   23828   poly16x4_t __s0 = __p0; \
   23829   poly16x4_t __s1 = __p1; \
   23830   poly16x4_t __ret; \
   23831   __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
   23832   __ret; \
   23833 })
   23834 #else
   23835 #define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
   23836   poly16x4_t __s0 = __p0; \
   23837   poly16x4_t __s1 = __p1; \
   23838   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23839   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   23840   poly16x4_t __ret; \
   23841   __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
   23842   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23843   __ret; \
   23844 })
   23845 #endif
   23846 
   23847 #ifdef __LITTLE_ENDIAN__
   23848 #define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
   23849   poly8x16_t __s0 = __p0; \
   23850   poly8x16_t __s1 = __p1; \
   23851   poly8x16_t __ret; \
   23852   __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
   23853   __ret; \
   23854 })
   23855 #else
   23856 #define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
   23857   poly8x16_t __s0 = __p0; \
   23858   poly8x16_t __s1 = __p1; \
   23859   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23860   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23861   poly8x16_t __ret; \
   23862   __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
   23863   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23864   __ret; \
   23865 })
   23866 #endif
   23867 
   23868 #ifdef __LITTLE_ENDIAN__
   23869 #define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
   23870   poly16x8_t __s0 = __p0; \
   23871   poly16x8_t __s1 = __p1; \
   23872   poly16x8_t __ret; \
   23873   __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
   23874   __ret; \
   23875 })
   23876 #else
   23877 #define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
   23878   poly16x8_t __s0 = __p0; \
   23879   poly16x8_t __s1 = __p1; \
   23880   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23881   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   23882   poly16x8_t __ret; \
   23883   __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
   23884   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23885   __ret; \
   23886 })
   23887 #endif
   23888 
   23889 #ifdef __LITTLE_ENDIAN__
   23890 #define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
   23891   uint8x16_t __s0 = __p0; \
   23892   uint8x16_t __s1 = __p1; \
   23893   uint8x16_t __ret; \
   23894   __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
   23895   __ret; \
   23896 })
   23897 #else
   23898 #define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
   23899   uint8x16_t __s0 = __p0; \
   23900   uint8x16_t __s1 = __p1; \
   23901   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23902   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23903   uint8x16_t __ret; \
   23904   __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
   23905   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23906   __ret; \
   23907 })
   23908 #endif
   23909 
   23910 #ifdef __LITTLE_ENDIAN__
   23911 #define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
   23912   uint32x4_t __s0 = __p0; \
   23913   uint32x4_t __s1 = __p1; \
   23914   uint32x4_t __ret; \
   23915   __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
   23916   __ret; \
   23917 })
   23918 #else
   23919 #define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
   23920   uint32x4_t __s0 = __p0; \
   23921   uint32x4_t __s1 = __p1; \
   23922   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   23923   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   23924   uint32x4_t __ret; \
   23925   __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
   23926   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   23927   __ret; \
   23928 })
   23929 #endif
   23930 
   23931 #ifdef __LITTLE_ENDIAN__
   23932 #define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
   23933   uint64x2_t __s0 = __p0; \
   23934   uint64x2_t __s1 = __p1; \
   23935   uint64x2_t __ret; \
   23936   __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
   23937   __ret; \
   23938 })
   23939 #else
   23940 #define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
   23941   uint64x2_t __s0 = __p0; \
   23942   uint64x2_t __s1 = __p1; \
   23943   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   23944   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   23945   uint64x2_t __ret; \
   23946   __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
   23947   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   23948   __ret; \
   23949 })
   23950 #endif
   23951 
   23952 #ifdef __LITTLE_ENDIAN__
   23953 #define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
   23954   uint16x8_t __s0 = __p0; \
   23955   uint16x8_t __s1 = __p1; \
   23956   uint16x8_t __ret; \
   23957   __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
   23958   __ret; \
   23959 })
   23960 #else
   23961 #define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
   23962   uint16x8_t __s0 = __p0; \
   23963   uint16x8_t __s1 = __p1; \
   23964   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   23965   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   23966   uint16x8_t __ret; \
   23967   __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
   23968   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   23969   __ret; \
   23970 })
   23971 #endif
   23972 
   23973 #ifdef __LITTLE_ENDIAN__
   23974 #define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
   23975   int8x16_t __s0 = __p0; \
   23976   int8x16_t __s1 = __p1; \
   23977   int8x16_t __ret; \
   23978   __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
   23979   __ret; \
   23980 })
   23981 #else
   23982 #define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
   23983   int8x16_t __s0 = __p0; \
   23984   int8x16_t __s1 = __p1; \
   23985   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23986   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23987   int8x16_t __ret; \
   23988   __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
   23989   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   23990   __ret; \
   23991 })
   23992 #endif
   23993 
   23994 #ifdef __LITTLE_ENDIAN__
   23995 #define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
   23996   int32x4_t __s0 = __p0; \
   23997   int32x4_t __s1 = __p1; \
   23998   int32x4_t __ret; \
   23999   __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
   24000   __ret; \
   24001 })
   24002 #else
   24003 #define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24004   int32x4_t __s0 = __p0; \
   24005   int32x4_t __s1 = __p1; \
   24006   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24007   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24008   int32x4_t __ret; \
   24009   __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
   24010   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24011   __ret; \
   24012 })
   24013 #endif
   24014 
   24015 #ifdef __LITTLE_ENDIAN__
   24016 #define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24017   int64x2_t __s0 = __p0; \
   24018   int64x2_t __s1 = __p1; \
   24019   int64x2_t __ret; \
   24020   __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
   24021   __ret; \
   24022 })
   24023 #else
   24024 #define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24025   int64x2_t __s0 = __p0; \
   24026   int64x2_t __s1 = __p1; \
   24027   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24028   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24029   int64x2_t __ret; \
   24030   __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
   24031   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24032   __ret; \
   24033 })
   24034 #endif
   24035 
   24036 #ifdef __LITTLE_ENDIAN__
   24037 #define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24038   int16x8_t __s0 = __p0; \
   24039   int16x8_t __s1 = __p1; \
   24040   int16x8_t __ret; \
   24041   __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
   24042   __ret; \
   24043 })
   24044 #else
   24045 #define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24046   int16x8_t __s0 = __p0; \
   24047   int16x8_t __s1 = __p1; \
   24048   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24049   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24050   int16x8_t __ret; \
   24051   __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
   24052   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24053   __ret; \
   24054 })
   24055 #endif
   24056 
   24057 #ifdef __LITTLE_ENDIAN__
   24058 #define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24059   uint8x8_t __s0 = __p0; \
   24060   uint8x8_t __s1 = __p1; \
   24061   uint8x8_t __ret; \
   24062   __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
   24063   __ret; \
   24064 })
   24065 #else
   24066 #define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24067   uint8x8_t __s0 = __p0; \
   24068   uint8x8_t __s1 = __p1; \
   24069   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24070   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24071   uint8x8_t __ret; \
   24072   __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
   24073   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24074   __ret; \
   24075 })
   24076 #endif
   24077 
   24078 #ifdef __LITTLE_ENDIAN__
   24079 #define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24080   uint32x2_t __s0 = __p0; \
   24081   uint32x2_t __s1 = __p1; \
   24082   uint32x2_t __ret; \
   24083   __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
   24084   __ret; \
   24085 })
   24086 #else
   24087 #define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24088   uint32x2_t __s0 = __p0; \
   24089   uint32x2_t __s1 = __p1; \
   24090   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24091   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24092   uint32x2_t __ret; \
   24093   __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
   24094   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24095   __ret; \
   24096 })
   24097 #endif
   24098 
   24099 #ifdef __LITTLE_ENDIAN__
   24100 #define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24101   uint64x1_t __s0 = __p0; \
   24102   uint64x1_t __s1 = __p1; \
   24103   uint64x1_t __ret; \
   24104   __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   24105   __ret; \
   24106 })
   24107 #else
   24108 #define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24109   uint64x1_t __s0 = __p0; \
   24110   uint64x1_t __s1 = __p1; \
   24111   uint64x1_t __ret; \
   24112   __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   24113   __ret; \
   24114 })
   24115 #endif
   24116 
   24117 #ifdef __LITTLE_ENDIAN__
   24118 #define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24119   uint16x4_t __s0 = __p0; \
   24120   uint16x4_t __s1 = __p1; \
   24121   uint16x4_t __ret; \
   24122   __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
   24123   __ret; \
   24124 })
   24125 #else
   24126 #define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24127   uint16x4_t __s0 = __p0; \
   24128   uint16x4_t __s1 = __p1; \
   24129   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24130   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24131   uint16x4_t __ret; \
   24132   __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
   24133   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24134   __ret; \
   24135 })
   24136 #endif
   24137 
   24138 #ifdef __LITTLE_ENDIAN__
   24139 #define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24140   int8x8_t __s0 = __p0; \
   24141   int8x8_t __s1 = __p1; \
   24142   int8x8_t __ret; \
   24143   __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
   24144   __ret; \
   24145 })
   24146 #else
   24147 #define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24148   int8x8_t __s0 = __p0; \
   24149   int8x8_t __s1 = __p1; \
   24150   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24151   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24152   int8x8_t __ret; \
   24153   __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
   24154   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24155   __ret; \
   24156 })
   24157 #endif
   24158 
   24159 #ifdef __LITTLE_ENDIAN__
   24160 #define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24161   int32x2_t __s0 = __p0; \
   24162   int32x2_t __s1 = __p1; \
   24163   int32x2_t __ret; \
   24164   __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
   24165   __ret; \
   24166 })
   24167 #else
   24168 #define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24169   int32x2_t __s0 = __p0; \
   24170   int32x2_t __s1 = __p1; \
   24171   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24172   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24173   int32x2_t __ret; \
   24174   __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
   24175   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24176   __ret; \
   24177 })
   24178 #endif
   24179 
   24180 #ifdef __LITTLE_ENDIAN__
   24181 #define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24182   int64x1_t __s0 = __p0; \
   24183   int64x1_t __s1 = __p1; \
   24184   int64x1_t __ret; \
   24185   __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   24186   __ret; \
   24187 })
   24188 #else
   24189 #define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24190   int64x1_t __s0 = __p0; \
   24191   int64x1_t __s1 = __p1; \
   24192   int64x1_t __ret; \
   24193   __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   24194   __ret; \
   24195 })
   24196 #endif
   24197 
   24198 #ifdef __LITTLE_ENDIAN__
   24199 #define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24200   int16x4_t __s0 = __p0; \
   24201   int16x4_t __s1 = __p1; \
   24202   int16x4_t __ret; \
   24203   __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
   24204   __ret; \
   24205 })
   24206 #else
   24207 #define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24208   int16x4_t __s0 = __p0; \
   24209   int16x4_t __s1 = __p1; \
   24210   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24211   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24212   int16x4_t __ret; \
   24213   __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
   24214   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24215   __ret; \
   24216 })
   24217 #endif
   24218 
   24219 #ifdef __LITTLE_ENDIAN__
   24220 #define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24221   uint8x16_t __s0 = __p0; \
   24222   uint8x16_t __s1 = __p1; \
   24223   uint8x16_t __ret; \
   24224   __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
   24225   __ret; \
   24226 })
   24227 #else
   24228 #define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24229   uint8x16_t __s0 = __p0; \
   24230   uint8x16_t __s1 = __p1; \
   24231   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24232   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24233   uint8x16_t __ret; \
   24234   __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
   24235   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24236   __ret; \
   24237 })
   24238 #endif
   24239 
   24240 #ifdef __LITTLE_ENDIAN__
   24241 #define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24242   uint32x4_t __s0 = __p0; \
   24243   uint32x4_t __s1 = __p1; \
   24244   uint32x4_t __ret; \
   24245   __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
   24246   __ret; \
   24247 })
   24248 #else
   24249 #define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24250   uint32x4_t __s0 = __p0; \
   24251   uint32x4_t __s1 = __p1; \
   24252   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24253   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24254   uint32x4_t __ret; \
   24255   __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
   24256   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24257   __ret; \
   24258 })
   24259 #endif
   24260 
   24261 #ifdef __LITTLE_ENDIAN__
   24262 #define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24263   uint64x2_t __s0 = __p0; \
   24264   uint64x2_t __s1 = __p1; \
   24265   uint64x2_t __ret; \
   24266   __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
   24267   __ret; \
   24268 })
   24269 #else
   24270 #define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24271   uint64x2_t __s0 = __p0; \
   24272   uint64x2_t __s1 = __p1; \
   24273   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24274   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24275   uint64x2_t __ret; \
   24276   __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
   24277   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24278   __ret; \
   24279 })
   24280 #endif
   24281 
   24282 #ifdef __LITTLE_ENDIAN__
   24283 #define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24284   uint16x8_t __s0 = __p0; \
   24285   uint16x8_t __s1 = __p1; \
   24286   uint16x8_t __ret; \
   24287   __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
   24288   __ret; \
   24289 })
   24290 #else
   24291 #define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24292   uint16x8_t __s0 = __p0; \
   24293   uint16x8_t __s1 = __p1; \
   24294   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24295   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24296   uint16x8_t __ret; \
   24297   __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
   24298   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24299   __ret; \
   24300 })
   24301 #endif
   24302 
   24303 #ifdef __LITTLE_ENDIAN__
   24304 #define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24305   int8x16_t __s0 = __p0; \
   24306   int8x16_t __s1 = __p1; \
   24307   int8x16_t __ret; \
   24308   __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
   24309   __ret; \
   24310 })
   24311 #else
   24312 #define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24313   int8x16_t __s0 = __p0; \
   24314   int8x16_t __s1 = __p1; \
   24315   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24316   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24317   int8x16_t __ret; \
   24318   __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
   24319   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24320   __ret; \
   24321 })
   24322 #endif
   24323 
   24324 #ifdef __LITTLE_ENDIAN__
   24325 #define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24326   int32x4_t __s0 = __p0; \
   24327   int32x4_t __s1 = __p1; \
   24328   int32x4_t __ret; \
   24329   __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
   24330   __ret; \
   24331 })
   24332 #else
   24333 #define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24334   int32x4_t __s0 = __p0; \
   24335   int32x4_t __s1 = __p1; \
   24336   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24337   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24338   int32x4_t __ret; \
   24339   __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
   24340   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24341   __ret; \
   24342 })
   24343 #endif
   24344 
   24345 #ifdef __LITTLE_ENDIAN__
   24346 #define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24347   int64x2_t __s0 = __p0; \
   24348   int64x2_t __s1 = __p1; \
   24349   int64x2_t __ret; \
   24350   __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
   24351   __ret; \
   24352 })
   24353 #else
   24354 #define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24355   int64x2_t __s0 = __p0; \
   24356   int64x2_t __s1 = __p1; \
   24357   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24358   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24359   int64x2_t __ret; \
   24360   __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
   24361   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24362   __ret; \
   24363 })
   24364 #endif
   24365 
   24366 #ifdef __LITTLE_ENDIAN__
   24367 #define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24368   int16x8_t __s0 = __p0; \
   24369   int16x8_t __s1 = __p1; \
   24370   int16x8_t __ret; \
   24371   __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
   24372   __ret; \
   24373 })
   24374 #else
   24375 #define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24376   int16x8_t __s0 = __p0; \
   24377   int16x8_t __s1 = __p1; \
   24378   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24379   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24380   int16x8_t __ret; \
   24381   __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
   24382   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24383   __ret; \
   24384 })
   24385 #endif
   24386 
   24387 #ifdef __LITTLE_ENDIAN__
   24388 #define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24389   uint8x8_t __s0 = __p0; \
   24390   uint8x8_t __s1 = __p1; \
   24391   uint8x8_t __ret; \
   24392   __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
   24393   __ret; \
   24394 })
   24395 #else
   24396 #define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24397   uint8x8_t __s0 = __p0; \
   24398   uint8x8_t __s1 = __p1; \
   24399   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24400   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24401   uint8x8_t __ret; \
   24402   __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
   24403   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24404   __ret; \
   24405 })
   24406 #endif
   24407 
   24408 #ifdef __LITTLE_ENDIAN__
   24409 #define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24410   uint32x2_t __s0 = __p0; \
   24411   uint32x2_t __s1 = __p1; \
   24412   uint32x2_t __ret; \
   24413   __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
   24414   __ret; \
   24415 })
   24416 #else
   24417 #define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24418   uint32x2_t __s0 = __p0; \
   24419   uint32x2_t __s1 = __p1; \
   24420   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24421   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24422   uint32x2_t __ret; \
   24423   __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
   24424   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24425   __ret; \
   24426 })
   24427 #endif
   24428 
   24429 #ifdef __LITTLE_ENDIAN__
   24430 #define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24431   uint64x1_t __s0 = __p0; \
   24432   uint64x1_t __s1 = __p1; \
   24433   uint64x1_t __ret; \
   24434   __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   24435   __ret; \
   24436 })
   24437 #else
   24438 #define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24439   uint64x1_t __s0 = __p0; \
   24440   uint64x1_t __s1 = __p1; \
   24441   uint64x1_t __ret; \
   24442   __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   24443   __ret; \
   24444 })
   24445 #endif
   24446 
   24447 #ifdef __LITTLE_ENDIAN__
   24448 #define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24449   uint16x4_t __s0 = __p0; \
   24450   uint16x4_t __s1 = __p1; \
   24451   uint16x4_t __ret; \
   24452   __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
   24453   __ret; \
   24454 })
   24455 #else
   24456 #define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24457   uint16x4_t __s0 = __p0; \
   24458   uint16x4_t __s1 = __p1; \
   24459   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24460   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24461   uint16x4_t __ret; \
   24462   __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
   24463   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24464   __ret; \
   24465 })
   24466 #endif
   24467 
   24468 #ifdef __LITTLE_ENDIAN__
   24469 #define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24470   int8x8_t __s0 = __p0; \
   24471   int8x8_t __s1 = __p1; \
   24472   int8x8_t __ret; \
   24473   __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
   24474   __ret; \
   24475 })
   24476 #else
   24477 #define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24478   int8x8_t __s0 = __p0; \
   24479   int8x8_t __s1 = __p1; \
   24480   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24481   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24482   int8x8_t __ret; \
   24483   __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
   24484   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24485   __ret; \
   24486 })
   24487 #endif
   24488 
   24489 #ifdef __LITTLE_ENDIAN__
   24490 #define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24491   int32x2_t __s0 = __p0; \
   24492   int32x2_t __s1 = __p1; \
   24493   int32x2_t __ret; \
   24494   __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
   24495   __ret; \
   24496 })
   24497 #else
   24498 #define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24499   int32x2_t __s0 = __p0; \
   24500   int32x2_t __s1 = __p1; \
   24501   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24502   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24503   int32x2_t __ret; \
   24504   __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
   24505   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24506   __ret; \
   24507 })
   24508 #endif
   24509 
   24510 #ifdef __LITTLE_ENDIAN__
   24511 #define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24512   int64x1_t __s0 = __p0; \
   24513   int64x1_t __s1 = __p1; \
   24514   int64x1_t __ret; \
   24515   __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   24516   __ret; \
   24517 })
   24518 #else
   24519 #define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24520   int64x1_t __s0 = __p0; \
   24521   int64x1_t __s1 = __p1; \
   24522   int64x1_t __ret; \
   24523   __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   24524   __ret; \
   24525 })
   24526 #endif
   24527 
   24528 #ifdef __LITTLE_ENDIAN__
   24529 #define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24530   int16x4_t __s0 = __p0; \
   24531   int16x4_t __s1 = __p1; \
   24532   int16x4_t __ret; \
   24533   __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
   24534   __ret; \
   24535 })
   24536 #else
   24537 #define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24538   int16x4_t __s0 = __p0; \
   24539   int16x4_t __s1 = __p1; \
   24540   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24541   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24542   int16x4_t __ret; \
   24543   __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
   24544   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24545   __ret; \
   24546 })
   24547 #endif
   24548 
   24549 #ifdef __LITTLE_ENDIAN__
   24550 #define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
   24551   poly8x8_t __s0 = __p0; \
   24552   poly8x8_t __s1 = __p1; \
   24553   poly8x8_t __ret; \
   24554   __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
   24555   __ret; \
   24556 })
   24557 #else
   24558 #define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
   24559   poly8x8_t __s0 = __p0; \
   24560   poly8x8_t __s1 = __p1; \
   24561   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24562   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24563   poly8x8_t __ret; \
   24564   __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
   24565   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24566   __ret; \
   24567 })
   24568 #endif
   24569 
   24570 #ifdef __LITTLE_ENDIAN__
   24571 #define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
   24572   poly16x4_t __s0 = __p0; \
   24573   poly16x4_t __s1 = __p1; \
   24574   poly16x4_t __ret; \
   24575   __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
   24576   __ret; \
   24577 })
   24578 #else
   24579 #define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
   24580   poly16x4_t __s0 = __p0; \
   24581   poly16x4_t __s1 = __p1; \
   24582   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24583   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24584   poly16x4_t __ret; \
   24585   __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
   24586   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24587   __ret; \
   24588 })
   24589 #endif
   24590 
   24591 #ifdef __LITTLE_ENDIAN__
   24592 #define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
   24593   poly8x16_t __s0 = __p0; \
   24594   poly8x16_t __s1 = __p1; \
   24595   poly8x16_t __ret; \
   24596   __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
   24597   __ret; \
   24598 })
   24599 #else
   24600 #define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
   24601   poly8x16_t __s0 = __p0; \
   24602   poly8x16_t __s1 = __p1; \
   24603   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24604   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24605   poly8x16_t __ret; \
   24606   __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
   24607   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24608   __ret; \
   24609 })
   24610 #endif
   24611 
   24612 #ifdef __LITTLE_ENDIAN__
   24613 #define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
   24614   poly16x8_t __s0 = __p0; \
   24615   poly16x8_t __s1 = __p1; \
   24616   poly16x8_t __ret; \
   24617   __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
   24618   __ret; \
   24619 })
   24620 #else
   24621 #define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
   24622   poly16x8_t __s0 = __p0; \
   24623   poly16x8_t __s1 = __p1; \
   24624   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24625   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24626   poly16x8_t __ret; \
   24627   __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
   24628   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24629   __ret; \
   24630 })
   24631 #endif
   24632 
   24633 #ifdef __LITTLE_ENDIAN__
   24634 #define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24635   uint8x16_t __s0 = __p0; \
   24636   uint8x16_t __s1 = __p1; \
   24637   uint8x16_t __ret; \
   24638   __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
   24639   __ret; \
   24640 })
   24641 #else
   24642 #define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24643   uint8x16_t __s0 = __p0; \
   24644   uint8x16_t __s1 = __p1; \
   24645   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24646   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24647   uint8x16_t __ret; \
   24648   __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
   24649   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24650   __ret; \
   24651 })
   24652 #endif
   24653 
   24654 #ifdef __LITTLE_ENDIAN__
   24655 #define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24656   uint32x4_t __s0 = __p0; \
   24657   uint32x4_t __s1 = __p1; \
   24658   uint32x4_t __ret; \
   24659   __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
   24660   __ret; \
   24661 })
   24662 #else
   24663 #define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24664   uint32x4_t __s0 = __p0; \
   24665   uint32x4_t __s1 = __p1; \
   24666   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24667   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24668   uint32x4_t __ret; \
   24669   __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
   24670   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24671   __ret; \
   24672 })
   24673 #endif
   24674 
   24675 #ifdef __LITTLE_ENDIAN__
   24676 #define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24677   uint64x2_t __s0 = __p0; \
   24678   uint64x2_t __s1 = __p1; \
   24679   uint64x2_t __ret; \
   24680   __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
   24681   __ret; \
   24682 })
   24683 #else
   24684 #define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24685   uint64x2_t __s0 = __p0; \
   24686   uint64x2_t __s1 = __p1; \
   24687   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24688   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24689   uint64x2_t __ret; \
   24690   __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
   24691   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24692   __ret; \
   24693 })
   24694 #endif
   24695 
   24696 #ifdef __LITTLE_ENDIAN__
   24697 #define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24698   uint16x8_t __s0 = __p0; \
   24699   uint16x8_t __s1 = __p1; \
   24700   uint16x8_t __ret; \
   24701   __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
   24702   __ret; \
   24703 })
   24704 #else
   24705 #define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24706   uint16x8_t __s0 = __p0; \
   24707   uint16x8_t __s1 = __p1; \
   24708   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24709   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24710   uint16x8_t __ret; \
   24711   __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
   24712   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24713   __ret; \
   24714 })
   24715 #endif
   24716 
   24717 #ifdef __LITTLE_ENDIAN__
   24718 #define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24719   int8x16_t __s0 = __p0; \
   24720   int8x16_t __s1 = __p1; \
   24721   int8x16_t __ret; \
   24722   __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
   24723   __ret; \
   24724 })
   24725 #else
   24726 #define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24727   int8x16_t __s0 = __p0; \
   24728   int8x16_t __s1 = __p1; \
   24729   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24730   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24731   int8x16_t __ret; \
   24732   __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
   24733   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24734   __ret; \
   24735 })
   24736 #endif
   24737 
   24738 #ifdef __LITTLE_ENDIAN__
   24739 #define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24740   int32x4_t __s0 = __p0; \
   24741   int32x4_t __s1 = __p1; \
   24742   int32x4_t __ret; \
   24743   __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
   24744   __ret; \
   24745 })
   24746 #else
   24747 #define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24748   int32x4_t __s0 = __p0; \
   24749   int32x4_t __s1 = __p1; \
   24750   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24751   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24752   int32x4_t __ret; \
   24753   __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
   24754   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24755   __ret; \
   24756 })
   24757 #endif
   24758 
   24759 #ifdef __LITTLE_ENDIAN__
   24760 #define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24761   int64x2_t __s0 = __p0; \
   24762   int64x2_t __s1 = __p1; \
   24763   int64x2_t __ret; \
   24764   __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
   24765   __ret; \
   24766 })
   24767 #else
   24768 #define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24769   int64x2_t __s0 = __p0; \
   24770   int64x2_t __s1 = __p1; \
   24771   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24772   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24773   int64x2_t __ret; \
   24774   __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
   24775   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24776   __ret; \
   24777 })
   24778 #endif
   24779 
   24780 #ifdef __LITTLE_ENDIAN__
   24781 #define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24782   int16x8_t __s0 = __p0; \
   24783   int16x8_t __s1 = __p1; \
   24784   int16x8_t __ret; \
   24785   __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
   24786   __ret; \
   24787 })
   24788 #else
   24789 #define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24790   int16x8_t __s0 = __p0; \
   24791   int16x8_t __s1 = __p1; \
   24792   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24793   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24794   int16x8_t __ret; \
   24795   __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
   24796   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24797   __ret; \
   24798 })
   24799 #endif
   24800 
   24801 #ifdef __LITTLE_ENDIAN__
   24802 #define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24803   uint8x8_t __s0 = __p0; \
   24804   uint8x8_t __s1 = __p1; \
   24805   uint8x8_t __ret; \
   24806   __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
   24807   __ret; \
   24808 })
   24809 #else
   24810 #define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
   24811   uint8x8_t __s0 = __p0; \
   24812   uint8x8_t __s1 = __p1; \
   24813   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24814   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24815   uint8x8_t __ret; \
   24816   __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
   24817   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24818   __ret; \
   24819 })
   24820 #endif
   24821 
   24822 #ifdef __LITTLE_ENDIAN__
   24823 #define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24824   uint32x2_t __s0 = __p0; \
   24825   uint32x2_t __s1 = __p1; \
   24826   uint32x2_t __ret; \
   24827   __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
   24828   __ret; \
   24829 })
   24830 #else
   24831 #define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
   24832   uint32x2_t __s0 = __p0; \
   24833   uint32x2_t __s1 = __p1; \
   24834   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24835   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24836   uint32x2_t __ret; \
   24837   __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
   24838   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24839   __ret; \
   24840 })
   24841 #endif
   24842 
   24843 #ifdef __LITTLE_ENDIAN__
   24844 #define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24845   uint64x1_t __s0 = __p0; \
   24846   uint64x1_t __s1 = __p1; \
   24847   uint64x1_t __ret; \
   24848   __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   24849   __ret; \
   24850 })
   24851 #else
   24852 #define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
   24853   uint64x1_t __s0 = __p0; \
   24854   uint64x1_t __s1 = __p1; \
   24855   uint64x1_t __ret; \
   24856   __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
   24857   __ret; \
   24858 })
   24859 #endif
   24860 
   24861 #ifdef __LITTLE_ENDIAN__
   24862 #define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24863   uint16x4_t __s0 = __p0; \
   24864   uint16x4_t __s1 = __p1; \
   24865   uint16x4_t __ret; \
   24866   __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
   24867   __ret; \
   24868 })
   24869 #else
   24870 #define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
   24871   uint16x4_t __s0 = __p0; \
   24872   uint16x4_t __s1 = __p1; \
   24873   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24874   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24875   uint16x4_t __ret; \
   24876   __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
   24877   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24878   __ret; \
   24879 })
   24880 #endif
   24881 
   24882 #ifdef __LITTLE_ENDIAN__
   24883 #define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24884   int8x8_t __s0 = __p0; \
   24885   int8x8_t __s1 = __p1; \
   24886   int8x8_t __ret; \
   24887   __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
   24888   __ret; \
   24889 })
   24890 #else
   24891 #define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
   24892   int8x8_t __s0 = __p0; \
   24893   int8x8_t __s1 = __p1; \
   24894   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   24895   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24896   int8x8_t __ret; \
   24897   __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
   24898   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   24899   __ret; \
   24900 })
   24901 #endif
   24902 
   24903 #ifdef __LITTLE_ENDIAN__
   24904 #define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24905   int32x2_t __s0 = __p0; \
   24906   int32x2_t __s1 = __p1; \
   24907   int32x2_t __ret; \
   24908   __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
   24909   __ret; \
   24910 })
   24911 #else
   24912 #define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
   24913   int32x2_t __s0 = __p0; \
   24914   int32x2_t __s1 = __p1; \
   24915   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   24916   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   24917   int32x2_t __ret; \
   24918   __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
   24919   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   24920   __ret; \
   24921 })
   24922 #endif
   24923 
   24924 #ifdef __LITTLE_ENDIAN__
   24925 #define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24926   int64x1_t __s0 = __p0; \
   24927   int64x1_t __s1 = __p1; \
   24928   int64x1_t __ret; \
   24929   __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   24930   __ret; \
   24931 })
   24932 #else
   24933 #define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
   24934   int64x1_t __s0 = __p0; \
   24935   int64x1_t __s1 = __p1; \
   24936   int64x1_t __ret; \
   24937   __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
   24938   __ret; \
   24939 })
   24940 #endif
   24941 
   24942 #ifdef __LITTLE_ENDIAN__
   24943 #define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24944   int16x4_t __s0 = __p0; \
   24945   int16x4_t __s1 = __p1; \
   24946   int16x4_t __ret; \
   24947   __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
   24948   __ret; \
   24949 })
   24950 #else
   24951 #define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
   24952   int16x4_t __s0 = __p0; \
   24953   int16x4_t __s1 = __p1; \
   24954   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   24955   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24956   int16x4_t __ret; \
   24957   __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
   24958   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   24959   __ret; \
   24960 })
   24961 #endif
   24962 
   24963 #ifdef __LITTLE_ENDIAN__
   24964 #define vst1_p8(__p0, __p1) __extension__ ({ \
   24965   poly8x8_t __s1 = __p1; \
   24966   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 4); \
   24967 })
   24968 #else
   24969 #define vst1_p8(__p0, __p1) __extension__ ({ \
   24970   poly8x8_t __s1 = __p1; \
   24971   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   24972   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 4); \
   24973 })
   24974 #endif
   24975 
   24976 #ifdef __LITTLE_ENDIAN__
   24977 #define vst1_p16(__p0, __p1) __extension__ ({ \
   24978   poly16x4_t __s1 = __p1; \
   24979   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 5); \
   24980 })
   24981 #else
   24982 #define vst1_p16(__p0, __p1) __extension__ ({ \
   24983   poly16x4_t __s1 = __p1; \
   24984   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   24985   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 5); \
   24986 })
   24987 #endif
   24988 
   24989 #ifdef __LITTLE_ENDIAN__
   24990 #define vst1q_p8(__p0, __p1) __extension__ ({ \
   24991   poly8x16_t __s1 = __p1; \
   24992   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 36); \
   24993 })
   24994 #else
   24995 #define vst1q_p8(__p0, __p1) __extension__ ({ \
   24996   poly8x16_t __s1 = __p1; \
   24997   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   24998   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 36); \
   24999 })
   25000 #endif
   25001 
   25002 #ifdef __LITTLE_ENDIAN__
   25003 #define vst1q_p16(__p0, __p1) __extension__ ({ \
   25004   poly16x8_t __s1 = __p1; \
   25005   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 37); \
   25006 })
   25007 #else
   25008 #define vst1q_p16(__p0, __p1) __extension__ ({ \
   25009   poly16x8_t __s1 = __p1; \
   25010   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25011   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 37); \
   25012 })
   25013 #endif
   25014 
   25015 #ifdef __LITTLE_ENDIAN__
   25016 #define vst1q_u8(__p0, __p1) __extension__ ({ \
   25017   uint8x16_t __s1 = __p1; \
   25018   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 48); \
   25019 })
   25020 #else
   25021 #define vst1q_u8(__p0, __p1) __extension__ ({ \
   25022   uint8x16_t __s1 = __p1; \
   25023   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25024   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 48); \
   25025 })
   25026 #endif
   25027 
   25028 #ifdef __LITTLE_ENDIAN__
   25029 #define vst1q_u32(__p0, __p1) __extension__ ({ \
   25030   uint32x4_t __s1 = __p1; \
   25031   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 50); \
   25032 })
   25033 #else
   25034 #define vst1q_u32(__p0, __p1) __extension__ ({ \
   25035   uint32x4_t __s1 = __p1; \
   25036   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25037   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 50); \
   25038 })
   25039 #endif
   25040 
   25041 #ifdef __LITTLE_ENDIAN__
   25042 #define vst1q_u64(__p0, __p1) __extension__ ({ \
   25043   uint64x2_t __s1 = __p1; \
   25044   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 51); \
   25045 })
   25046 #else
   25047 #define vst1q_u64(__p0, __p1) __extension__ ({ \
   25048   uint64x2_t __s1 = __p1; \
   25049   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25050   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 51); \
   25051 })
   25052 #endif
   25053 
   25054 #ifdef __LITTLE_ENDIAN__
   25055 #define vst1q_u16(__p0, __p1) __extension__ ({ \
   25056   uint16x8_t __s1 = __p1; \
   25057   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 49); \
   25058 })
   25059 #else
   25060 #define vst1q_u16(__p0, __p1) __extension__ ({ \
   25061   uint16x8_t __s1 = __p1; \
   25062   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25063   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 49); \
   25064 })
   25065 #endif
   25066 
   25067 #ifdef __LITTLE_ENDIAN__
   25068 #define vst1q_s8(__p0, __p1) __extension__ ({ \
   25069   int8x16_t __s1 = __p1; \
   25070   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 32); \
   25071 })
   25072 #else
   25073 #define vst1q_s8(__p0, __p1) __extension__ ({ \
   25074   int8x16_t __s1 = __p1; \
   25075   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25076   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 32); \
   25077 })
   25078 #endif
   25079 
   25080 #ifdef __LITTLE_ENDIAN__
   25081 #define vst1q_f32(__p0, __p1) __extension__ ({ \
   25082   float32x4_t __s1 = __p1; \
   25083   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 41); \
   25084 })
   25085 #else
   25086 #define vst1q_f32(__p0, __p1) __extension__ ({ \
   25087   float32x4_t __s1 = __p1; \
   25088   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25089   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 41); \
   25090 })
   25091 #endif
   25092 
   25093 #ifdef __LITTLE_ENDIAN__
   25094 #define vst1q_f16(__p0, __p1) __extension__ ({ \
   25095   float16x8_t __s1 = __p1; \
   25096   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 40); \
   25097 })
   25098 #else
   25099 #define vst1q_f16(__p0, __p1) __extension__ ({ \
   25100   float16x8_t __s1 = __p1; \
   25101   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25102   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 40); \
   25103 })
   25104 #endif
   25105 
   25106 #ifdef __LITTLE_ENDIAN__
   25107 #define vst1q_s32(__p0, __p1) __extension__ ({ \
   25108   int32x4_t __s1 = __p1; \
   25109   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 34); \
   25110 })
   25111 #else
   25112 #define vst1q_s32(__p0, __p1) __extension__ ({ \
   25113   int32x4_t __s1 = __p1; \
   25114   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25115   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 34); \
   25116 })
   25117 #endif
   25118 
   25119 #ifdef __LITTLE_ENDIAN__
   25120 #define vst1q_s64(__p0, __p1) __extension__ ({ \
   25121   int64x2_t __s1 = __p1; \
   25122   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 35); \
   25123 })
   25124 #else
   25125 #define vst1q_s64(__p0, __p1) __extension__ ({ \
   25126   int64x2_t __s1 = __p1; \
   25127   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25128   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 35); \
   25129 })
   25130 #endif
   25131 
   25132 #ifdef __LITTLE_ENDIAN__
   25133 #define vst1q_s16(__p0, __p1) __extension__ ({ \
   25134   int16x8_t __s1 = __p1; \
   25135   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 33); \
   25136 })
   25137 #else
   25138 #define vst1q_s16(__p0, __p1) __extension__ ({ \
   25139   int16x8_t __s1 = __p1; \
   25140   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25141   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 33); \
   25142 })
   25143 #endif
   25144 
   25145 #ifdef __LITTLE_ENDIAN__
   25146 #define vst1_u8(__p0, __p1) __extension__ ({ \
   25147   uint8x8_t __s1 = __p1; \
   25148   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 16); \
   25149 })
   25150 #else
   25151 #define vst1_u8(__p0, __p1) __extension__ ({ \
   25152   uint8x8_t __s1 = __p1; \
   25153   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25154   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 16); \
   25155 })
   25156 #endif
   25157 
   25158 #ifdef __LITTLE_ENDIAN__
   25159 #define vst1_u32(__p0, __p1) __extension__ ({ \
   25160   uint32x2_t __s1 = __p1; \
   25161   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 18); \
   25162 })
   25163 #else
   25164 #define vst1_u32(__p0, __p1) __extension__ ({ \
   25165   uint32x2_t __s1 = __p1; \
   25166   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25167   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 18); \
   25168 })
   25169 #endif
   25170 
   25171 #ifdef __LITTLE_ENDIAN__
   25172 #define vst1_u64(__p0, __p1) __extension__ ({ \
   25173   uint64x1_t __s1 = __p1; \
   25174   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
   25175 })
   25176 #else
   25177 #define vst1_u64(__p0, __p1) __extension__ ({ \
   25178   uint64x1_t __s1 = __p1; \
   25179   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
   25180 })
   25181 #endif
   25182 
   25183 #ifdef __LITTLE_ENDIAN__
   25184 #define vst1_u16(__p0, __p1) __extension__ ({ \
   25185   uint16x4_t __s1 = __p1; \
   25186   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 17); \
   25187 })
   25188 #else
   25189 #define vst1_u16(__p0, __p1) __extension__ ({ \
   25190   uint16x4_t __s1 = __p1; \
   25191   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25192   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 17); \
   25193 })
   25194 #endif
   25195 
   25196 #ifdef __LITTLE_ENDIAN__
   25197 #define vst1_s8(__p0, __p1) __extension__ ({ \
   25198   int8x8_t __s1 = __p1; \
   25199   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 0); \
   25200 })
   25201 #else
   25202 #define vst1_s8(__p0, __p1) __extension__ ({ \
   25203   int8x8_t __s1 = __p1; \
   25204   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25205   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 0); \
   25206 })
   25207 #endif
   25208 
   25209 #ifdef __LITTLE_ENDIAN__
   25210 #define vst1_f32(__p0, __p1) __extension__ ({ \
   25211   float32x2_t __s1 = __p1; \
   25212   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 9); \
   25213 })
   25214 #else
   25215 #define vst1_f32(__p0, __p1) __extension__ ({ \
   25216   float32x2_t __s1 = __p1; \
   25217   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25218   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 9); \
   25219 })
   25220 #endif
   25221 
   25222 #ifdef __LITTLE_ENDIAN__
   25223 #define vst1_f16(__p0, __p1) __extension__ ({ \
   25224   float16x4_t __s1 = __p1; \
   25225   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 8); \
   25226 })
   25227 #else
   25228 #define vst1_f16(__p0, __p1) __extension__ ({ \
   25229   float16x4_t __s1 = __p1; \
   25230   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25231   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 8); \
   25232 })
   25233 #endif
   25234 
   25235 #ifdef __LITTLE_ENDIAN__
   25236 #define vst1_s32(__p0, __p1) __extension__ ({ \
   25237   int32x2_t __s1 = __p1; \
   25238   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 2); \
   25239 })
   25240 #else
   25241 #define vst1_s32(__p0, __p1) __extension__ ({ \
   25242   int32x2_t __s1 = __p1; \
   25243   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25244   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 2); \
   25245 })
   25246 #endif
   25247 
   25248 #ifdef __LITTLE_ENDIAN__
   25249 #define vst1_s64(__p0, __p1) __extension__ ({ \
   25250   int64x1_t __s1 = __p1; \
   25251   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
   25252 })
   25253 #else
   25254 #define vst1_s64(__p0, __p1) __extension__ ({ \
   25255   int64x1_t __s1 = __p1; \
   25256   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
   25257 })
   25258 #endif
   25259 
   25260 #ifdef __LITTLE_ENDIAN__
   25261 #define vst1_s16(__p0, __p1) __extension__ ({ \
   25262   int16x4_t __s1 = __p1; \
   25263   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 1); \
   25264 })
   25265 #else
   25266 #define vst1_s16(__p0, __p1) __extension__ ({ \
   25267   int16x4_t __s1 = __p1; \
   25268   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25269   __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 1); \
   25270 })
   25271 #endif
   25272 
   25273 #ifdef __LITTLE_ENDIAN__
   25274 #define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   25275   poly8x8_t __s1 = __p1; \
   25276   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
   25277 })
   25278 #else
   25279 #define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   25280   poly8x8_t __s1 = __p1; \
   25281   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25282   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
   25283 })
   25284 #endif
   25285 
   25286 #ifdef __LITTLE_ENDIAN__
   25287 #define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   25288   poly16x4_t __s1 = __p1; \
   25289   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
   25290 })
   25291 #else
   25292 #define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   25293   poly16x4_t __s1 = __p1; \
   25294   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25295   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
   25296 })
   25297 #endif
   25298 
   25299 #ifdef __LITTLE_ENDIAN__
   25300 #define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   25301   poly8x16_t __s1 = __p1; \
   25302   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
   25303 })
   25304 #else
   25305 #define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   25306   poly8x16_t __s1 = __p1; \
   25307   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25308   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
   25309 })
   25310 #endif
   25311 
   25312 #ifdef __LITTLE_ENDIAN__
   25313 #define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   25314   poly16x8_t __s1 = __p1; \
   25315   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
   25316 })
   25317 #else
   25318 #define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   25319   poly16x8_t __s1 = __p1; \
   25320   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25321   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
   25322 })
   25323 #endif
   25324 
   25325 #ifdef __LITTLE_ENDIAN__
   25326 #define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   25327   uint8x16_t __s1 = __p1; \
   25328   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
   25329 })
   25330 #else
   25331 #define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   25332   uint8x16_t __s1 = __p1; \
   25333   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25334   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
   25335 })
   25336 #endif
   25337 
   25338 #ifdef __LITTLE_ENDIAN__
   25339 #define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   25340   uint32x4_t __s1 = __p1; \
   25341   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
   25342 })
   25343 #else
   25344 #define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   25345   uint32x4_t __s1 = __p1; \
   25346   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25347   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
   25348 })
   25349 #endif
   25350 
   25351 #ifdef __LITTLE_ENDIAN__
   25352 #define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   25353   uint64x2_t __s1 = __p1; \
   25354   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
   25355 })
   25356 #else
   25357 #define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   25358   uint64x2_t __s1 = __p1; \
   25359   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25360   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
   25361 })
   25362 #endif
   25363 
   25364 #ifdef __LITTLE_ENDIAN__
   25365 #define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   25366   uint16x8_t __s1 = __p1; \
   25367   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
   25368 })
   25369 #else
   25370 #define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   25371   uint16x8_t __s1 = __p1; \
   25372   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25373   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
   25374 })
   25375 #endif
   25376 
   25377 #ifdef __LITTLE_ENDIAN__
   25378 #define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   25379   int8x16_t __s1 = __p1; \
   25380   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
   25381 })
   25382 #else
   25383 #define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   25384   int8x16_t __s1 = __p1; \
   25385   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25386   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
   25387 })
   25388 #endif
   25389 
   25390 #ifdef __LITTLE_ENDIAN__
   25391 #define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   25392   float32x4_t __s1 = __p1; \
   25393   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
   25394 })
   25395 #else
   25396 #define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   25397   float32x4_t __s1 = __p1; \
   25398   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25399   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
   25400 })
   25401 #endif
   25402 
   25403 #ifdef __LITTLE_ENDIAN__
   25404 #define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   25405   float16x8_t __s1 = __p1; \
   25406   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
   25407 })
   25408 #else
   25409 #define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   25410   float16x8_t __s1 = __p1; \
   25411   float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25412   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
   25413 })
   25414 #endif
   25415 
   25416 #ifdef __LITTLE_ENDIAN__
   25417 #define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   25418   int32x4_t __s1 = __p1; \
   25419   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
   25420 })
   25421 #else
   25422 #define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   25423   int32x4_t __s1 = __p1; \
   25424   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25425   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
   25426 })
   25427 #endif
   25428 
   25429 #ifdef __LITTLE_ENDIAN__
   25430 #define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   25431   int64x2_t __s1 = __p1; \
   25432   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
   25433 })
   25434 #else
   25435 #define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   25436   int64x2_t __s1 = __p1; \
   25437   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25438   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
   25439 })
   25440 #endif
   25441 
   25442 #ifdef __LITTLE_ENDIAN__
   25443 #define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   25444   int16x8_t __s1 = __p1; \
   25445   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
   25446 })
   25447 #else
   25448 #define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   25449   int16x8_t __s1 = __p1; \
   25450   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25451   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
   25452 })
   25453 #endif
   25454 
   25455 #ifdef __LITTLE_ENDIAN__
   25456 #define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   25457   uint8x8_t __s1 = __p1; \
   25458   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
   25459 })
   25460 #else
   25461 #define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   25462   uint8x8_t __s1 = __p1; \
   25463   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25464   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
   25465 })
   25466 #endif
   25467 
   25468 #ifdef __LITTLE_ENDIAN__
   25469 #define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   25470   uint32x2_t __s1 = __p1; \
   25471   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
   25472 })
   25473 #else
   25474 #define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   25475   uint32x2_t __s1 = __p1; \
   25476   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25477   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
   25478 })
   25479 #endif
   25480 
   25481 #ifdef __LITTLE_ENDIAN__
   25482 #define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   25483   uint64x1_t __s1 = __p1; \
   25484   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
   25485 })
   25486 #else
   25487 #define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   25488   uint64x1_t __s1 = __p1; \
   25489   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
   25490 })
   25491 #endif
   25492 
   25493 #ifdef __LITTLE_ENDIAN__
   25494 #define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   25495   uint16x4_t __s1 = __p1; \
   25496   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
   25497 })
   25498 #else
   25499 #define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   25500   uint16x4_t __s1 = __p1; \
   25501   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25502   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
   25503 })
   25504 #endif
   25505 
   25506 #ifdef __LITTLE_ENDIAN__
   25507 #define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   25508   int8x8_t __s1 = __p1; \
   25509   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
   25510 })
   25511 #else
   25512 #define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   25513   int8x8_t __s1 = __p1; \
   25514   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   25515   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
   25516 })
   25517 #endif
   25518 
   25519 #ifdef __LITTLE_ENDIAN__
   25520 #define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   25521   float32x2_t __s1 = __p1; \
   25522   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
   25523 })
   25524 #else
   25525 #define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   25526   float32x2_t __s1 = __p1; \
   25527   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25528   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
   25529 })
   25530 #endif
   25531 
   25532 #ifdef __LITTLE_ENDIAN__
   25533 #define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   25534   float16x4_t __s1 = __p1; \
   25535   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
   25536 })
   25537 #else
   25538 #define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   25539   float16x4_t __s1 = __p1; \
   25540   float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25541   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
   25542 })
   25543 #endif
   25544 
   25545 #ifdef __LITTLE_ENDIAN__
   25546 #define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   25547   int32x2_t __s1 = __p1; \
   25548   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
   25549 })
   25550 #else
   25551 #define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   25552   int32x2_t __s1 = __p1; \
   25553   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   25554   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
   25555 })
   25556 #endif
   25557 
   25558 #ifdef __LITTLE_ENDIAN__
   25559 #define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   25560   int64x1_t __s1 = __p1; \
   25561   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
   25562 })
   25563 #else
   25564 #define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   25565   int64x1_t __s1 = __p1; \
   25566   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
   25567 })
   25568 #endif
   25569 
   25570 #ifdef __LITTLE_ENDIAN__
   25571 #define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   25572   int16x4_t __s1 = __p1; \
   25573   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
   25574 })
   25575 #else
   25576 #define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   25577   int16x4_t __s1 = __p1; \
   25578   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   25579   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
   25580 })
   25581 #endif
   25582 
   25583 #ifdef __LITTLE_ENDIAN__
   25584 #define vst2_p8(__p0, __p1) __extension__ ({ \
   25585   poly8x8x2_t __s1 = __p1; \
   25586   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
   25587 })
   25588 #else
   25589 #define vst2_p8(__p0, __p1) __extension__ ({ \
   25590   poly8x8x2_t __s1 = __p1; \
   25591   poly8x8x2_t __rev1; \
   25592   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25593   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25594   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
   25595 })
   25596 #endif
   25597 
   25598 #ifdef __LITTLE_ENDIAN__
   25599 #define vst2_p16(__p0, __p1) __extension__ ({ \
   25600   poly16x4x2_t __s1 = __p1; \
   25601   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
   25602 })
   25603 #else
   25604 #define vst2_p16(__p0, __p1) __extension__ ({ \
   25605   poly16x4x2_t __s1 = __p1; \
   25606   poly16x4x2_t __rev1; \
   25607   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25608   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25609   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
   25610 })
   25611 #endif
   25612 
   25613 #ifdef __LITTLE_ENDIAN__
   25614 #define vst2q_p8(__p0, __p1) __extension__ ({ \
   25615   poly8x16x2_t __s1 = __p1; \
   25616   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
   25617 })
   25618 #else
   25619 #define vst2q_p8(__p0, __p1) __extension__ ({ \
   25620   poly8x16x2_t __s1 = __p1; \
   25621   poly8x16x2_t __rev1; \
   25622   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25623   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25624   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
   25625 })
   25626 #endif
   25627 
   25628 #ifdef __LITTLE_ENDIAN__
   25629 #define vst2q_p16(__p0, __p1) __extension__ ({ \
   25630   poly16x8x2_t __s1 = __p1; \
   25631   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
   25632 })
   25633 #else
   25634 #define vst2q_p16(__p0, __p1) __extension__ ({ \
   25635   poly16x8x2_t __s1 = __p1; \
   25636   poly16x8x2_t __rev1; \
   25637   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25638   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25639   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
   25640 })
   25641 #endif
   25642 
   25643 #ifdef __LITTLE_ENDIAN__
   25644 #define vst2q_u8(__p0, __p1) __extension__ ({ \
   25645   uint8x16x2_t __s1 = __p1; \
   25646   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
   25647 })
   25648 #else
   25649 #define vst2q_u8(__p0, __p1) __extension__ ({ \
   25650   uint8x16x2_t __s1 = __p1; \
   25651   uint8x16x2_t __rev1; \
   25652   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25653   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25654   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
   25655 })
   25656 #endif
   25657 
   25658 #ifdef __LITTLE_ENDIAN__
   25659 #define vst2q_u32(__p0, __p1) __extension__ ({ \
   25660   uint32x4x2_t __s1 = __p1; \
   25661   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
   25662 })
   25663 #else
   25664 #define vst2q_u32(__p0, __p1) __extension__ ({ \
   25665   uint32x4x2_t __s1 = __p1; \
   25666   uint32x4x2_t __rev1; \
   25667   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25668   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25669   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
   25670 })
   25671 #endif
   25672 
   25673 #ifdef __LITTLE_ENDIAN__
   25674 #define vst2q_u16(__p0, __p1) __extension__ ({ \
   25675   uint16x8x2_t __s1 = __p1; \
   25676   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
   25677 })
   25678 #else
   25679 #define vst2q_u16(__p0, __p1) __extension__ ({ \
   25680   uint16x8x2_t __s1 = __p1; \
   25681   uint16x8x2_t __rev1; \
   25682   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25683   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25684   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
   25685 })
   25686 #endif
   25687 
   25688 #ifdef __LITTLE_ENDIAN__
   25689 #define vst2q_s8(__p0, __p1) __extension__ ({ \
   25690   int8x16x2_t __s1 = __p1; \
   25691   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
   25692 })
   25693 #else
   25694 #define vst2q_s8(__p0, __p1) __extension__ ({ \
   25695   int8x16x2_t __s1 = __p1; \
   25696   int8x16x2_t __rev1; \
   25697   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25698   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   25699   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
   25700 })
   25701 #endif
   25702 
   25703 #ifdef __LITTLE_ENDIAN__
   25704 #define vst2q_f32(__p0, __p1) __extension__ ({ \
   25705   float32x4x2_t __s1 = __p1; \
   25706   __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 41); \
   25707 })
   25708 #else
   25709 #define vst2q_f32(__p0, __p1) __extension__ ({ \
   25710   float32x4x2_t __s1 = __p1; \
   25711   float32x4x2_t __rev1; \
   25712   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25713   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25714   __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
   25715 })
   25716 #endif
   25717 
   25718 #ifdef __LITTLE_ENDIAN__
   25719 #define vst2q_f16(__p0, __p1) __extension__ ({ \
   25720   float16x8x2_t __s1 = __p1; \
   25721   __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 40); \
   25722 })
   25723 #else
   25724 #define vst2q_f16(__p0, __p1) __extension__ ({ \
   25725   float16x8x2_t __s1 = __p1; \
   25726   float16x8x2_t __rev1; \
   25727   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25728   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25729   __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
   25730 })
   25731 #endif
   25732 
   25733 #ifdef __LITTLE_ENDIAN__
   25734 #define vst2q_s32(__p0, __p1) __extension__ ({ \
   25735   int32x4x2_t __s1 = __p1; \
   25736   __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 34); \
   25737 })
   25738 #else
   25739 #define vst2q_s32(__p0, __p1) __extension__ ({ \
   25740   int32x4x2_t __s1 = __p1; \
   25741   int32x4x2_t __rev1; \
   25742   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25743   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25744   __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
   25745 })
   25746 #endif
   25747 
   25748 #ifdef __LITTLE_ENDIAN__
   25749 #define vst2q_s16(__p0, __p1) __extension__ ({ \
   25750   int16x8x2_t __s1 = __p1; \
   25751   __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 33); \
   25752 })
   25753 #else
   25754 #define vst2q_s16(__p0, __p1) __extension__ ({ \
   25755   int16x8x2_t __s1 = __p1; \
   25756   int16x8x2_t __rev1; \
   25757   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25758   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25759   __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
   25760 })
   25761 #endif
   25762 
   25763 #ifdef __LITTLE_ENDIAN__
   25764 #define vst2_u8(__p0, __p1) __extension__ ({ \
   25765   uint8x8x2_t __s1 = __p1; \
   25766   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
   25767 })
   25768 #else
   25769 #define vst2_u8(__p0, __p1) __extension__ ({ \
   25770   uint8x8x2_t __s1 = __p1; \
   25771   uint8x8x2_t __rev1; \
   25772   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25773   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25774   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
   25775 })
   25776 #endif
   25777 
   25778 #ifdef __LITTLE_ENDIAN__
   25779 #define vst2_u32(__p0, __p1) __extension__ ({ \
   25780   uint32x2x2_t __s1 = __p1; \
   25781   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
   25782 })
   25783 #else
   25784 #define vst2_u32(__p0, __p1) __extension__ ({ \
   25785   uint32x2x2_t __s1 = __p1; \
   25786   uint32x2x2_t __rev1; \
   25787   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   25788   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   25789   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
   25790 })
   25791 #endif
   25792 
   25793 #ifdef __LITTLE_ENDIAN__
   25794 #define vst2_u64(__p0, __p1) __extension__ ({ \
   25795   uint64x1x2_t __s1 = __p1; \
   25796   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
   25797 })
   25798 #else
   25799 #define vst2_u64(__p0, __p1) __extension__ ({ \
   25800   uint64x1x2_t __s1 = __p1; \
   25801   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
   25802 })
   25803 #endif
   25804 
   25805 #ifdef __LITTLE_ENDIAN__
   25806 #define vst2_u16(__p0, __p1) __extension__ ({ \
   25807   uint16x4x2_t __s1 = __p1; \
   25808   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
   25809 })
   25810 #else
   25811 #define vst2_u16(__p0, __p1) __extension__ ({ \
   25812   uint16x4x2_t __s1 = __p1; \
   25813   uint16x4x2_t __rev1; \
   25814   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25815   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25816   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
   25817 })
   25818 #endif
   25819 
   25820 #ifdef __LITTLE_ENDIAN__
   25821 #define vst2_s8(__p0, __p1) __extension__ ({ \
   25822   int8x8x2_t __s1 = __p1; \
   25823   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
   25824 })
   25825 #else
   25826 #define vst2_s8(__p0, __p1) __extension__ ({ \
   25827   int8x8x2_t __s1 = __p1; \
   25828   int8x8x2_t __rev1; \
   25829   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25830   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25831   __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
   25832 })
   25833 #endif
   25834 
   25835 #ifdef __LITTLE_ENDIAN__
   25836 #define vst2_f32(__p0, __p1) __extension__ ({ \
   25837   float32x2x2_t __s1 = __p1; \
   25838   __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 9); \
   25839 })
   25840 #else
   25841 #define vst2_f32(__p0, __p1) __extension__ ({ \
   25842   float32x2x2_t __s1 = __p1; \
   25843   float32x2x2_t __rev1; \
   25844   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   25845   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   25846   __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
   25847 })
   25848 #endif
   25849 
   25850 #ifdef __LITTLE_ENDIAN__
   25851 #define vst2_f16(__p0, __p1) __extension__ ({ \
   25852   float16x4x2_t __s1 = __p1; \
   25853   __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 8); \
   25854 })
   25855 #else
   25856 #define vst2_f16(__p0, __p1) __extension__ ({ \
   25857   float16x4x2_t __s1 = __p1; \
   25858   float16x4x2_t __rev1; \
   25859   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25860   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25861   __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
   25862 })
   25863 #endif
   25864 
   25865 #ifdef __LITTLE_ENDIAN__
   25866 #define vst2_s32(__p0, __p1) __extension__ ({ \
   25867   int32x2x2_t __s1 = __p1; \
   25868   __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 2); \
   25869 })
   25870 #else
   25871 #define vst2_s32(__p0, __p1) __extension__ ({ \
   25872   int32x2x2_t __s1 = __p1; \
   25873   int32x2x2_t __rev1; \
   25874   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   25875   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   25876   __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
   25877 })
   25878 #endif
   25879 
   25880 #ifdef __LITTLE_ENDIAN__
   25881 #define vst2_s64(__p0, __p1) __extension__ ({ \
   25882   int64x1x2_t __s1 = __p1; \
   25883   __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
   25884 })
   25885 #else
   25886 #define vst2_s64(__p0, __p1) __extension__ ({ \
   25887   int64x1x2_t __s1 = __p1; \
   25888   __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
   25889 })
   25890 #endif
   25891 
   25892 #ifdef __LITTLE_ENDIAN__
   25893 #define vst2_s16(__p0, __p1) __extension__ ({ \
   25894   int16x4x2_t __s1 = __p1; \
   25895   __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 1); \
   25896 })
   25897 #else
   25898 #define vst2_s16(__p0, __p1) __extension__ ({ \
   25899   int16x4x2_t __s1 = __p1; \
   25900   int16x4x2_t __rev1; \
   25901   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25902   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25903   __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
   25904 })
   25905 #endif
   25906 
   25907 #ifdef __LITTLE_ENDIAN__
   25908 #define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   25909   poly8x8x2_t __s1 = __p1; \
   25910   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
   25911 })
   25912 #else
   25913 #define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   25914   poly8x8x2_t __s1 = __p1; \
   25915   poly8x8x2_t __rev1; \
   25916   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25917   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25918   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
   25919 })
   25920 #endif
   25921 
   25922 #ifdef __LITTLE_ENDIAN__
   25923 #define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   25924   poly16x4x2_t __s1 = __p1; \
   25925   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
   25926 })
   25927 #else
   25928 #define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   25929   poly16x4x2_t __s1 = __p1; \
   25930   poly16x4x2_t __rev1; \
   25931   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25932   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25933   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
   25934 })
   25935 #endif
   25936 
   25937 #ifdef __LITTLE_ENDIAN__
   25938 #define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   25939   poly16x8x2_t __s1 = __p1; \
   25940   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
   25941 })
   25942 #else
   25943 #define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   25944   poly16x8x2_t __s1 = __p1; \
   25945   poly16x8x2_t __rev1; \
   25946   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25947   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25948   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
   25949 })
   25950 #endif
   25951 
   25952 #ifdef __LITTLE_ENDIAN__
   25953 #define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   25954   uint32x4x2_t __s1 = __p1; \
   25955   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
   25956 })
   25957 #else
   25958 #define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   25959   uint32x4x2_t __s1 = __p1; \
   25960   uint32x4x2_t __rev1; \
   25961   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25962   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25963   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
   25964 })
   25965 #endif
   25966 
   25967 #ifdef __LITTLE_ENDIAN__
   25968 #define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   25969   uint16x8x2_t __s1 = __p1; \
   25970   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
   25971 })
   25972 #else
   25973 #define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   25974   uint16x8x2_t __s1 = __p1; \
   25975   uint16x8x2_t __rev1; \
   25976   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   25977   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   25978   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
   25979 })
   25980 #endif
   25981 
   25982 #ifdef __LITTLE_ENDIAN__
   25983 #define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   25984   float32x4x2_t __s1 = __p1; \
   25985   __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 41); \
   25986 })
   25987 #else
   25988 #define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   25989   float32x4x2_t __s1 = __p1; \
   25990   float32x4x2_t __rev1; \
   25991   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   25992   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   25993   __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
   25994 })
   25995 #endif
   25996 
   25997 #ifdef __LITTLE_ENDIAN__
   25998 #define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   25999   float16x8x2_t __s1 = __p1; \
   26000   __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 40); \
   26001 })
   26002 #else
   26003 #define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   26004   float16x8x2_t __s1 = __p1; \
   26005   float16x8x2_t __rev1; \
   26006   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26007   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26008   __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
   26009 })
   26010 #endif
   26011 
   26012 #ifdef __LITTLE_ENDIAN__
   26013 #define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   26014   int32x4x2_t __s1 = __p1; \
   26015   __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 34); \
   26016 })
   26017 #else
   26018 #define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   26019   int32x4x2_t __s1 = __p1; \
   26020   int32x4x2_t __rev1; \
   26021   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26022   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26023   __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
   26024 })
   26025 #endif
   26026 
   26027 #ifdef __LITTLE_ENDIAN__
   26028 #define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   26029   int16x8x2_t __s1 = __p1; \
   26030   __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 33); \
   26031 })
   26032 #else
   26033 #define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   26034   int16x8x2_t __s1 = __p1; \
   26035   int16x8x2_t __rev1; \
   26036   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26037   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26038   __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
   26039 })
   26040 #endif
   26041 
   26042 #ifdef __LITTLE_ENDIAN__
   26043 #define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   26044   uint8x8x2_t __s1 = __p1; \
   26045   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
   26046 })
   26047 #else
   26048 #define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   26049   uint8x8x2_t __s1 = __p1; \
   26050   uint8x8x2_t __rev1; \
   26051   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26052   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26053   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
   26054 })
   26055 #endif
   26056 
   26057 #ifdef __LITTLE_ENDIAN__
   26058 #define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   26059   uint32x2x2_t __s1 = __p1; \
   26060   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
   26061 })
   26062 #else
   26063 #define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   26064   uint32x2x2_t __s1 = __p1; \
   26065   uint32x2x2_t __rev1; \
   26066   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   26067   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   26068   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
   26069 })
   26070 #endif
   26071 
   26072 #ifdef __LITTLE_ENDIAN__
   26073 #define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   26074   uint16x4x2_t __s1 = __p1; \
   26075   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
   26076 })
   26077 #else
   26078 #define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   26079   uint16x4x2_t __s1 = __p1; \
   26080   uint16x4x2_t __rev1; \
   26081   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26082   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26083   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
   26084 })
   26085 #endif
   26086 
   26087 #ifdef __LITTLE_ENDIAN__
   26088 #define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   26089   int8x8x2_t __s1 = __p1; \
   26090   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
   26091 })
   26092 #else
   26093 #define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   26094   int8x8x2_t __s1 = __p1; \
   26095   int8x8x2_t __rev1; \
   26096   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26097   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26098   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
   26099 })
   26100 #endif
   26101 
   26102 #ifdef __LITTLE_ENDIAN__
   26103 #define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   26104   float32x2x2_t __s1 = __p1; \
   26105   __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 9); \
   26106 })
   26107 #else
   26108 #define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   26109   float32x2x2_t __s1 = __p1; \
   26110   float32x2x2_t __rev1; \
   26111   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   26112   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   26113   __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
   26114 })
   26115 #endif
   26116 
   26117 #ifdef __LITTLE_ENDIAN__
   26118 #define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   26119   float16x4x2_t __s1 = __p1; \
   26120   __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 8); \
   26121 })
   26122 #else
   26123 #define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   26124   float16x4x2_t __s1 = __p1; \
   26125   float16x4x2_t __rev1; \
   26126   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26127   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26128   __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
   26129 })
   26130 #endif
   26131 
   26132 #ifdef __LITTLE_ENDIAN__
   26133 #define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   26134   int32x2x2_t __s1 = __p1; \
   26135   __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 2); \
   26136 })
   26137 #else
   26138 #define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   26139   int32x2x2_t __s1 = __p1; \
   26140   int32x2x2_t __rev1; \
   26141   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   26142   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   26143   __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
   26144 })
   26145 #endif
   26146 
   26147 #ifdef __LITTLE_ENDIAN__
   26148 #define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   26149   int16x4x2_t __s1 = __p1; \
   26150   __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 1); \
   26151 })
   26152 #else
   26153 #define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   26154   int16x4x2_t __s1 = __p1; \
   26155   int16x4x2_t __rev1; \
   26156   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26157   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26158   __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
   26159 })
   26160 #endif
   26161 
   26162 #ifdef __LITTLE_ENDIAN__
   26163 #define vst3_p8(__p0, __p1) __extension__ ({ \
   26164   poly8x8x3_t __s1 = __p1; \
   26165   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
   26166 })
   26167 #else
   26168 #define vst3_p8(__p0, __p1) __extension__ ({ \
   26169   poly8x8x3_t __s1 = __p1; \
   26170   poly8x8x3_t __rev1; \
   26171   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26172   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26173   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26174   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
   26175 })
   26176 #endif
   26177 
   26178 #ifdef __LITTLE_ENDIAN__
   26179 #define vst3_p16(__p0, __p1) __extension__ ({ \
   26180   poly16x4x3_t __s1 = __p1; \
   26181   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
   26182 })
   26183 #else
   26184 #define vst3_p16(__p0, __p1) __extension__ ({ \
   26185   poly16x4x3_t __s1 = __p1; \
   26186   poly16x4x3_t __rev1; \
   26187   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26188   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26189   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26190   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
   26191 })
   26192 #endif
   26193 
   26194 #ifdef __LITTLE_ENDIAN__
   26195 #define vst3q_p8(__p0, __p1) __extension__ ({ \
   26196   poly8x16x3_t __s1 = __p1; \
   26197   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
   26198 })
   26199 #else
   26200 #define vst3q_p8(__p0, __p1) __extension__ ({ \
   26201   poly8x16x3_t __s1 = __p1; \
   26202   poly8x16x3_t __rev1; \
   26203   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26204   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26205   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26206   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
   26207 })
   26208 #endif
   26209 
   26210 #ifdef __LITTLE_ENDIAN__
   26211 #define vst3q_p16(__p0, __p1) __extension__ ({ \
   26212   poly16x8x3_t __s1 = __p1; \
   26213   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
   26214 })
   26215 #else
   26216 #define vst3q_p16(__p0, __p1) __extension__ ({ \
   26217   poly16x8x3_t __s1 = __p1; \
   26218   poly16x8x3_t __rev1; \
   26219   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26220   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26221   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26222   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
   26223 })
   26224 #endif
   26225 
   26226 #ifdef __LITTLE_ENDIAN__
   26227 #define vst3q_u8(__p0, __p1) __extension__ ({ \
   26228   uint8x16x3_t __s1 = __p1; \
   26229   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
   26230 })
   26231 #else
   26232 #define vst3q_u8(__p0, __p1) __extension__ ({ \
   26233   uint8x16x3_t __s1 = __p1; \
   26234   uint8x16x3_t __rev1; \
   26235   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26236   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26237   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26238   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
   26239 })
   26240 #endif
   26241 
   26242 #ifdef __LITTLE_ENDIAN__
   26243 #define vst3q_u32(__p0, __p1) __extension__ ({ \
   26244   uint32x4x3_t __s1 = __p1; \
   26245   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
   26246 })
   26247 #else
   26248 #define vst3q_u32(__p0, __p1) __extension__ ({ \
   26249   uint32x4x3_t __s1 = __p1; \
   26250   uint32x4x3_t __rev1; \
   26251   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26252   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26253   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26254   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
   26255 })
   26256 #endif
   26257 
   26258 #ifdef __LITTLE_ENDIAN__
   26259 #define vst3q_u16(__p0, __p1) __extension__ ({ \
   26260   uint16x8x3_t __s1 = __p1; \
   26261   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
   26262 })
   26263 #else
   26264 #define vst3q_u16(__p0, __p1) __extension__ ({ \
   26265   uint16x8x3_t __s1 = __p1; \
   26266   uint16x8x3_t __rev1; \
   26267   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26268   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26269   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26270   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
   26271 })
   26272 #endif
   26273 
   26274 #ifdef __LITTLE_ENDIAN__
   26275 #define vst3q_s8(__p0, __p1) __extension__ ({ \
   26276   int8x16x3_t __s1 = __p1; \
   26277   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
   26278 })
   26279 #else
   26280 #define vst3q_s8(__p0, __p1) __extension__ ({ \
   26281   int8x16x3_t __s1 = __p1; \
   26282   int8x16x3_t __rev1; \
   26283   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26284   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26285   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26286   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
   26287 })
   26288 #endif
   26289 
   26290 #ifdef __LITTLE_ENDIAN__
   26291 #define vst3q_f32(__p0, __p1) __extension__ ({ \
   26292   float32x4x3_t __s1 = __p1; \
   26293   __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
   26294 })
   26295 #else
   26296 #define vst3q_f32(__p0, __p1) __extension__ ({ \
   26297   float32x4x3_t __s1 = __p1; \
   26298   float32x4x3_t __rev1; \
   26299   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26300   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26301   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26302   __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
   26303 })
   26304 #endif
   26305 
   26306 #ifdef __LITTLE_ENDIAN__
   26307 #define vst3q_f16(__p0, __p1) __extension__ ({ \
   26308   float16x8x3_t __s1 = __p1; \
   26309   __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
   26310 })
   26311 #else
   26312 #define vst3q_f16(__p0, __p1) __extension__ ({ \
   26313   float16x8x3_t __s1 = __p1; \
   26314   float16x8x3_t __rev1; \
   26315   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26316   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26317   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26318   __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
   26319 })
   26320 #endif
   26321 
   26322 #ifdef __LITTLE_ENDIAN__
   26323 #define vst3q_s32(__p0, __p1) __extension__ ({ \
   26324   int32x4x3_t __s1 = __p1; \
   26325   __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
   26326 })
   26327 #else
   26328 #define vst3q_s32(__p0, __p1) __extension__ ({ \
   26329   int32x4x3_t __s1 = __p1; \
   26330   int32x4x3_t __rev1; \
   26331   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26332   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26333   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26334   __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
   26335 })
   26336 #endif
   26337 
   26338 #ifdef __LITTLE_ENDIAN__
   26339 #define vst3q_s16(__p0, __p1) __extension__ ({ \
   26340   int16x8x3_t __s1 = __p1; \
   26341   __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
   26342 })
   26343 #else
   26344 #define vst3q_s16(__p0, __p1) __extension__ ({ \
   26345   int16x8x3_t __s1 = __p1; \
   26346   int16x8x3_t __rev1; \
   26347   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26348   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26349   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26350   __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
   26351 })
   26352 #endif
   26353 
   26354 #ifdef __LITTLE_ENDIAN__
   26355 #define vst3_u8(__p0, __p1) __extension__ ({ \
   26356   uint8x8x3_t __s1 = __p1; \
   26357   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
   26358 })
   26359 #else
   26360 #define vst3_u8(__p0, __p1) __extension__ ({ \
   26361   uint8x8x3_t __s1 = __p1; \
   26362   uint8x8x3_t __rev1; \
   26363   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26364   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26365   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26366   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
   26367 })
   26368 #endif
   26369 
   26370 #ifdef __LITTLE_ENDIAN__
   26371 #define vst3_u32(__p0, __p1) __extension__ ({ \
   26372   uint32x2x3_t __s1 = __p1; \
   26373   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
   26374 })
   26375 #else
   26376 #define vst3_u32(__p0, __p1) __extension__ ({ \
   26377   uint32x2x3_t __s1 = __p1; \
   26378   uint32x2x3_t __rev1; \
   26379   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   26380   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   26381   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   26382   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
   26383 })
   26384 #endif
   26385 
   26386 #ifdef __LITTLE_ENDIAN__
   26387 #define vst3_u64(__p0, __p1) __extension__ ({ \
   26388   uint64x1x3_t __s1 = __p1; \
   26389   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
   26390 })
   26391 #else
   26392 #define vst3_u64(__p0, __p1) __extension__ ({ \
   26393   uint64x1x3_t __s1 = __p1; \
   26394   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
   26395 })
   26396 #endif
   26397 
   26398 #ifdef __LITTLE_ENDIAN__
   26399 #define vst3_u16(__p0, __p1) __extension__ ({ \
   26400   uint16x4x3_t __s1 = __p1; \
   26401   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
   26402 })
   26403 #else
   26404 #define vst3_u16(__p0, __p1) __extension__ ({ \
   26405   uint16x4x3_t __s1 = __p1; \
   26406   uint16x4x3_t __rev1; \
   26407   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26408   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26409   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26410   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
   26411 })
   26412 #endif
   26413 
   26414 #ifdef __LITTLE_ENDIAN__
   26415 #define vst3_s8(__p0, __p1) __extension__ ({ \
   26416   int8x8x3_t __s1 = __p1; \
   26417   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
   26418 })
   26419 #else
   26420 #define vst3_s8(__p0, __p1) __extension__ ({ \
   26421   int8x8x3_t __s1 = __p1; \
   26422   int8x8x3_t __rev1; \
   26423   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26424   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26425   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26426   __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
   26427 })
   26428 #endif
   26429 
   26430 #ifdef __LITTLE_ENDIAN__
   26431 #define vst3_f32(__p0, __p1) __extension__ ({ \
   26432   float32x2x3_t __s1 = __p1; \
   26433   __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
   26434 })
   26435 #else
   26436 #define vst3_f32(__p0, __p1) __extension__ ({ \
   26437   float32x2x3_t __s1 = __p1; \
   26438   float32x2x3_t __rev1; \
   26439   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   26440   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   26441   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   26442   __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
   26443 })
   26444 #endif
   26445 
   26446 #ifdef __LITTLE_ENDIAN__
   26447 #define vst3_f16(__p0, __p1) __extension__ ({ \
   26448   float16x4x3_t __s1 = __p1; \
   26449   __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
   26450 })
   26451 #else
   26452 #define vst3_f16(__p0, __p1) __extension__ ({ \
   26453   float16x4x3_t __s1 = __p1; \
   26454   float16x4x3_t __rev1; \
   26455   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26456   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26457   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26458   __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
   26459 })
   26460 #endif
   26461 
   26462 #ifdef __LITTLE_ENDIAN__
   26463 #define vst3_s32(__p0, __p1) __extension__ ({ \
   26464   int32x2x3_t __s1 = __p1; \
   26465   __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
   26466 })
   26467 #else
   26468 #define vst3_s32(__p0, __p1) __extension__ ({ \
   26469   int32x2x3_t __s1 = __p1; \
   26470   int32x2x3_t __rev1; \
   26471   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   26472   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   26473   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   26474   __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
   26475 })
   26476 #endif
   26477 
   26478 #ifdef __LITTLE_ENDIAN__
   26479 #define vst3_s64(__p0, __p1) __extension__ ({ \
   26480   int64x1x3_t __s1 = __p1; \
   26481   __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
   26482 })
   26483 #else
   26484 #define vst3_s64(__p0, __p1) __extension__ ({ \
   26485   int64x1x3_t __s1 = __p1; \
   26486   __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
   26487 })
   26488 #endif
   26489 
   26490 #ifdef __LITTLE_ENDIAN__
   26491 #define vst3_s16(__p0, __p1) __extension__ ({ \
   26492   int16x4x3_t __s1 = __p1; \
   26493   __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
   26494 })
   26495 #else
   26496 #define vst3_s16(__p0, __p1) __extension__ ({ \
   26497   int16x4x3_t __s1 = __p1; \
   26498   int16x4x3_t __rev1; \
   26499   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26500   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26501   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26502   __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
   26503 })
   26504 #endif
   26505 
   26506 #ifdef __LITTLE_ENDIAN__
   26507 #define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   26508   poly8x8x3_t __s1 = __p1; \
   26509   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
   26510 })
   26511 #else
   26512 #define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   26513   poly8x8x3_t __s1 = __p1; \
   26514   poly8x8x3_t __rev1; \
   26515   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26516   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26517   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26518   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
   26519 })
   26520 #endif
   26521 
   26522 #ifdef __LITTLE_ENDIAN__
   26523 #define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   26524   poly16x4x3_t __s1 = __p1; \
   26525   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
   26526 })
   26527 #else
   26528 #define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   26529   poly16x4x3_t __s1 = __p1; \
   26530   poly16x4x3_t __rev1; \
   26531   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26532   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26533   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26534   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
   26535 })
   26536 #endif
   26537 
   26538 #ifdef __LITTLE_ENDIAN__
   26539 #define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   26540   poly16x8x3_t __s1 = __p1; \
   26541   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
   26542 })
   26543 #else
   26544 #define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   26545   poly16x8x3_t __s1 = __p1; \
   26546   poly16x8x3_t __rev1; \
   26547   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26548   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26549   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26550   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
   26551 })
   26552 #endif
   26553 
   26554 #ifdef __LITTLE_ENDIAN__
   26555 #define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   26556   uint32x4x3_t __s1 = __p1; \
   26557   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
   26558 })
   26559 #else
   26560 #define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   26561   uint32x4x3_t __s1 = __p1; \
   26562   uint32x4x3_t __rev1; \
   26563   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26564   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26565   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26566   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
   26567 })
   26568 #endif
   26569 
   26570 #ifdef __LITTLE_ENDIAN__
   26571 #define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   26572   uint16x8x3_t __s1 = __p1; \
   26573   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
   26574 })
   26575 #else
   26576 #define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   26577   uint16x8x3_t __s1 = __p1; \
   26578   uint16x8x3_t __rev1; \
   26579   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26580   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26581   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26582   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
   26583 })
   26584 #endif
   26585 
   26586 #ifdef __LITTLE_ENDIAN__
   26587 #define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   26588   float32x4x3_t __s1 = __p1; \
   26589   __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
   26590 })
   26591 #else
   26592 #define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   26593   float32x4x3_t __s1 = __p1; \
   26594   float32x4x3_t __rev1; \
   26595   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26596   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26597   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26598   __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
   26599 })
   26600 #endif
   26601 
   26602 #ifdef __LITTLE_ENDIAN__
   26603 #define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   26604   float16x8x3_t __s1 = __p1; \
   26605   __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
   26606 })
   26607 #else
   26608 #define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   26609   float16x8x3_t __s1 = __p1; \
   26610   float16x8x3_t __rev1; \
   26611   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26612   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26613   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26614   __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
   26615 })
   26616 #endif
   26617 
   26618 #ifdef __LITTLE_ENDIAN__
   26619 #define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   26620   int32x4x3_t __s1 = __p1; \
   26621   __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
   26622 })
   26623 #else
   26624 #define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   26625   int32x4x3_t __s1 = __p1; \
   26626   int32x4x3_t __rev1; \
   26627   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26628   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26629   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26630   __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
   26631 })
   26632 #endif
   26633 
   26634 #ifdef __LITTLE_ENDIAN__
   26635 #define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   26636   int16x8x3_t __s1 = __p1; \
   26637   __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
   26638 })
   26639 #else
   26640 #define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   26641   int16x8x3_t __s1 = __p1; \
   26642   int16x8x3_t __rev1; \
   26643   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26644   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26645   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26646   __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
   26647 })
   26648 #endif
   26649 
   26650 #ifdef __LITTLE_ENDIAN__
   26651 #define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   26652   uint8x8x3_t __s1 = __p1; \
   26653   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
   26654 })
   26655 #else
   26656 #define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   26657   uint8x8x3_t __s1 = __p1; \
   26658   uint8x8x3_t __rev1; \
   26659   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26660   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26661   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26662   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
   26663 })
   26664 #endif
   26665 
   26666 #ifdef __LITTLE_ENDIAN__
   26667 #define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   26668   uint32x2x3_t __s1 = __p1; \
   26669   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
   26670 })
   26671 #else
   26672 #define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   26673   uint32x2x3_t __s1 = __p1; \
   26674   uint32x2x3_t __rev1; \
   26675   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   26676   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   26677   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   26678   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
   26679 })
   26680 #endif
   26681 
   26682 #ifdef __LITTLE_ENDIAN__
   26683 #define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   26684   uint16x4x3_t __s1 = __p1; \
   26685   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
   26686 })
   26687 #else
   26688 #define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   26689   uint16x4x3_t __s1 = __p1; \
   26690   uint16x4x3_t __rev1; \
   26691   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26692   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26693   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26694   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
   26695 })
   26696 #endif
   26697 
   26698 #ifdef __LITTLE_ENDIAN__
   26699 #define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   26700   int8x8x3_t __s1 = __p1; \
   26701   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
   26702 })
   26703 #else
   26704 #define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   26705   int8x8x3_t __s1 = __p1; \
   26706   int8x8x3_t __rev1; \
   26707   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26708   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26709   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26710   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
   26711 })
   26712 #endif
   26713 
   26714 #ifdef __LITTLE_ENDIAN__
   26715 #define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   26716   float32x2x3_t __s1 = __p1; \
   26717   __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
   26718 })
   26719 #else
   26720 #define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   26721   float32x2x3_t __s1 = __p1; \
   26722   float32x2x3_t __rev1; \
   26723   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   26724   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   26725   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   26726   __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
   26727 })
   26728 #endif
   26729 
   26730 #ifdef __LITTLE_ENDIAN__
   26731 #define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   26732   float16x4x3_t __s1 = __p1; \
   26733   __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
   26734 })
   26735 #else
   26736 #define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   26737   float16x4x3_t __s1 = __p1; \
   26738   float16x4x3_t __rev1; \
   26739   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26740   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26741   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26742   __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
   26743 })
   26744 #endif
   26745 
   26746 #ifdef __LITTLE_ENDIAN__
   26747 #define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   26748   int32x2x3_t __s1 = __p1; \
   26749   __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
   26750 })
   26751 #else
   26752 #define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   26753   int32x2x3_t __s1 = __p1; \
   26754   int32x2x3_t __rev1; \
   26755   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   26756   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   26757   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   26758   __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
   26759 })
   26760 #endif
   26761 
   26762 #ifdef __LITTLE_ENDIAN__
   26763 #define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   26764   int16x4x3_t __s1 = __p1; \
   26765   __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
   26766 })
   26767 #else
   26768 #define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   26769   int16x4x3_t __s1 = __p1; \
   26770   int16x4x3_t __rev1; \
   26771   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26772   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26773   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26774   __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
   26775 })
   26776 #endif
   26777 
   26778 #ifdef __LITTLE_ENDIAN__
   26779 #define vst4_p8(__p0, __p1) __extension__ ({ \
   26780   poly8x8x4_t __s1 = __p1; \
   26781   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
   26782 })
   26783 #else
   26784 #define vst4_p8(__p0, __p1) __extension__ ({ \
   26785   poly8x8x4_t __s1 = __p1; \
   26786   poly8x8x4_t __rev1; \
   26787   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26788   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26789   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26790   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   26791   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
   26792 })
   26793 #endif
   26794 
   26795 #ifdef __LITTLE_ENDIAN__
   26796 #define vst4_p16(__p0, __p1) __extension__ ({ \
   26797   poly16x4x4_t __s1 = __p1; \
   26798   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
   26799 })
   26800 #else
   26801 #define vst4_p16(__p0, __p1) __extension__ ({ \
   26802   poly16x4x4_t __s1 = __p1; \
   26803   poly16x4x4_t __rev1; \
   26804   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26805   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26806   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26807   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   26808   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
   26809 })
   26810 #endif
   26811 
   26812 #ifdef __LITTLE_ENDIAN__
   26813 #define vst4q_p8(__p0, __p1) __extension__ ({ \
   26814   poly8x16x4_t __s1 = __p1; \
   26815   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
   26816 })
   26817 #else
   26818 #define vst4q_p8(__p0, __p1) __extension__ ({ \
   26819   poly8x16x4_t __s1 = __p1; \
   26820   poly8x16x4_t __rev1; \
   26821   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26822   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26823   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26824   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26825   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
   26826 })
   26827 #endif
   26828 
   26829 #ifdef __LITTLE_ENDIAN__
   26830 #define vst4q_p16(__p0, __p1) __extension__ ({ \
   26831   poly16x8x4_t __s1 = __p1; \
   26832   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
   26833 })
   26834 #else
   26835 #define vst4q_p16(__p0, __p1) __extension__ ({ \
   26836   poly16x8x4_t __s1 = __p1; \
   26837   poly16x8x4_t __rev1; \
   26838   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26839   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26840   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26841   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   26842   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
   26843 })
   26844 #endif
   26845 
   26846 #ifdef __LITTLE_ENDIAN__
   26847 #define vst4q_u8(__p0, __p1) __extension__ ({ \
   26848   uint8x16x4_t __s1 = __p1; \
   26849   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
   26850 })
   26851 #else
   26852 #define vst4q_u8(__p0, __p1) __extension__ ({ \
   26853   uint8x16x4_t __s1 = __p1; \
   26854   uint8x16x4_t __rev1; \
   26855   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26856   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26857   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26858   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26859   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
   26860 })
   26861 #endif
   26862 
   26863 #ifdef __LITTLE_ENDIAN__
   26864 #define vst4q_u32(__p0, __p1) __extension__ ({ \
   26865   uint32x4x4_t __s1 = __p1; \
   26866   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
   26867 })
   26868 #else
   26869 #define vst4q_u32(__p0, __p1) __extension__ ({ \
   26870   uint32x4x4_t __s1 = __p1; \
   26871   uint32x4x4_t __rev1; \
   26872   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26873   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26874   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26875   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   26876   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
   26877 })
   26878 #endif
   26879 
   26880 #ifdef __LITTLE_ENDIAN__
   26881 #define vst4q_u16(__p0, __p1) __extension__ ({ \
   26882   uint16x8x4_t __s1 = __p1; \
   26883   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
   26884 })
   26885 #else
   26886 #define vst4q_u16(__p0, __p1) __extension__ ({ \
   26887   uint16x8x4_t __s1 = __p1; \
   26888   uint16x8x4_t __rev1; \
   26889   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26890   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26891   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26892   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   26893   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
   26894 })
   26895 #endif
   26896 
   26897 #ifdef __LITTLE_ENDIAN__
   26898 #define vst4q_s8(__p0, __p1) __extension__ ({ \
   26899   int8x16x4_t __s1 = __p1; \
   26900   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
   26901 })
   26902 #else
   26903 #define vst4q_s8(__p0, __p1) __extension__ ({ \
   26904   int8x16x4_t __s1 = __p1; \
   26905   int8x16x4_t __rev1; \
   26906   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26907   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26908   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26909   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   26910   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
   26911 })
   26912 #endif
   26913 
   26914 #ifdef __LITTLE_ENDIAN__
   26915 #define vst4q_f32(__p0, __p1) __extension__ ({ \
   26916   float32x4x4_t __s1 = __p1; \
   26917   __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
   26918 })
   26919 #else
   26920 #define vst4q_f32(__p0, __p1) __extension__ ({ \
   26921   float32x4x4_t __s1 = __p1; \
   26922   float32x4x4_t __rev1; \
   26923   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26924   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26925   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26926   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   26927   __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
   26928 })
   26929 #endif
   26930 
   26931 #ifdef __LITTLE_ENDIAN__
   26932 #define vst4q_f16(__p0, __p1) __extension__ ({ \
   26933   float16x8x4_t __s1 = __p1; \
   26934   __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
   26935 })
   26936 #else
   26937 #define vst4q_f16(__p0, __p1) __extension__ ({ \
   26938   float16x8x4_t __s1 = __p1; \
   26939   float16x8x4_t __rev1; \
   26940   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26941   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26942   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26943   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   26944   __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
   26945 })
   26946 #endif
   26947 
   26948 #ifdef __LITTLE_ENDIAN__
   26949 #define vst4q_s32(__p0, __p1) __extension__ ({ \
   26950   int32x4x4_t __s1 = __p1; \
   26951   __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
   26952 })
   26953 #else
   26954 #define vst4q_s32(__p0, __p1) __extension__ ({ \
   26955   int32x4x4_t __s1 = __p1; \
   26956   int32x4x4_t __rev1; \
   26957   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   26958   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   26959   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   26960   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   26961   __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
   26962 })
   26963 #endif
   26964 
   26965 #ifdef __LITTLE_ENDIAN__
   26966 #define vst4q_s16(__p0, __p1) __extension__ ({ \
   26967   int16x8x4_t __s1 = __p1; \
   26968   __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
   26969 })
   26970 #else
   26971 #define vst4q_s16(__p0, __p1) __extension__ ({ \
   26972   int16x8x4_t __s1 = __p1; \
   26973   int16x8x4_t __rev1; \
   26974   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26975   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26976   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26977   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   26978   __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
   26979 })
   26980 #endif
   26981 
   26982 #ifdef __LITTLE_ENDIAN__
   26983 #define vst4_u8(__p0, __p1) __extension__ ({ \
   26984   uint8x8x4_t __s1 = __p1; \
   26985   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
   26986 })
   26987 #else
   26988 #define vst4_u8(__p0, __p1) __extension__ ({ \
   26989   uint8x8x4_t __s1 = __p1; \
   26990   uint8x8x4_t __rev1; \
   26991   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   26992   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   26993   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   26994   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   26995   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
   26996 })
   26997 #endif
   26998 
   26999 #ifdef __LITTLE_ENDIAN__
   27000 #define vst4_u32(__p0, __p1) __extension__ ({ \
   27001   uint32x2x4_t __s1 = __p1; \
   27002   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
   27003 })
   27004 #else
   27005 #define vst4_u32(__p0, __p1) __extension__ ({ \
   27006   uint32x2x4_t __s1 = __p1; \
   27007   uint32x2x4_t __rev1; \
   27008   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   27009   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   27010   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   27011   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   27012   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
   27013 })
   27014 #endif
   27015 
   27016 #ifdef __LITTLE_ENDIAN__
   27017 #define vst4_u64(__p0, __p1) __extension__ ({ \
   27018   uint64x1x4_t __s1 = __p1; \
   27019   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
   27020 })
   27021 #else
   27022 #define vst4_u64(__p0, __p1) __extension__ ({ \
   27023   uint64x1x4_t __s1 = __p1; \
   27024   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
   27025 })
   27026 #endif
   27027 
   27028 #ifdef __LITTLE_ENDIAN__
   27029 #define vst4_u16(__p0, __p1) __extension__ ({ \
   27030   uint16x4x4_t __s1 = __p1; \
   27031   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
   27032 })
   27033 #else
   27034 #define vst4_u16(__p0, __p1) __extension__ ({ \
   27035   uint16x4x4_t __s1 = __p1; \
   27036   uint16x4x4_t __rev1; \
   27037   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27038   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27039   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27040   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27041   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
   27042 })
   27043 #endif
   27044 
   27045 #ifdef __LITTLE_ENDIAN__
   27046 #define vst4_s8(__p0, __p1) __extension__ ({ \
   27047   int8x8x4_t __s1 = __p1; \
   27048   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
   27049 })
   27050 #else
   27051 #define vst4_s8(__p0, __p1) __extension__ ({ \
   27052   int8x8x4_t __s1 = __p1; \
   27053   int8x8x4_t __rev1; \
   27054   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   27055   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   27056   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   27057   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   27058   __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
   27059 })
   27060 #endif
   27061 
   27062 #ifdef __LITTLE_ENDIAN__
   27063 #define vst4_f32(__p0, __p1) __extension__ ({ \
   27064   float32x2x4_t __s1 = __p1; \
   27065   __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
   27066 })
   27067 #else
   27068 #define vst4_f32(__p0, __p1) __extension__ ({ \
   27069   float32x2x4_t __s1 = __p1; \
   27070   float32x2x4_t __rev1; \
   27071   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   27072   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   27073   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   27074   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   27075   __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
   27076 })
   27077 #endif
   27078 
   27079 #ifdef __LITTLE_ENDIAN__
   27080 #define vst4_f16(__p0, __p1) __extension__ ({ \
   27081   float16x4x4_t __s1 = __p1; \
   27082   __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
   27083 })
   27084 #else
   27085 #define vst4_f16(__p0, __p1) __extension__ ({ \
   27086   float16x4x4_t __s1 = __p1; \
   27087   float16x4x4_t __rev1; \
   27088   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27089   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27090   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27091   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27092   __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
   27093 })
   27094 #endif
   27095 
   27096 #ifdef __LITTLE_ENDIAN__
   27097 #define vst4_s32(__p0, __p1) __extension__ ({ \
   27098   int32x2x4_t __s1 = __p1; \
   27099   __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
   27100 })
   27101 #else
   27102 #define vst4_s32(__p0, __p1) __extension__ ({ \
   27103   int32x2x4_t __s1 = __p1; \
   27104   int32x2x4_t __rev1; \
   27105   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   27106   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   27107   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   27108   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   27109   __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
   27110 })
   27111 #endif
   27112 
   27113 #ifdef __LITTLE_ENDIAN__
   27114 #define vst4_s64(__p0, __p1) __extension__ ({ \
   27115   int64x1x4_t __s1 = __p1; \
   27116   __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
   27117 })
   27118 #else
   27119 #define vst4_s64(__p0, __p1) __extension__ ({ \
   27120   int64x1x4_t __s1 = __p1; \
   27121   __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
   27122 })
   27123 #endif
   27124 
   27125 #ifdef __LITTLE_ENDIAN__
   27126 #define vst4_s16(__p0, __p1) __extension__ ({ \
   27127   int16x4x4_t __s1 = __p1; \
   27128   __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
   27129 })
   27130 #else
   27131 #define vst4_s16(__p0, __p1) __extension__ ({ \
   27132   int16x4x4_t __s1 = __p1; \
   27133   int16x4x4_t __rev1; \
   27134   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27135   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27136   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27137   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27138   __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
   27139 })
   27140 #endif
   27141 
   27142 #ifdef __LITTLE_ENDIAN__
   27143 #define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   27144   poly8x8x4_t __s1 = __p1; \
   27145   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
   27146 })
   27147 #else
   27148 #define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   27149   poly8x8x4_t __s1 = __p1; \
   27150   poly8x8x4_t __rev1; \
   27151   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   27152   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   27153   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   27154   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   27155   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
   27156 })
   27157 #endif
   27158 
   27159 #ifdef __LITTLE_ENDIAN__
   27160 #define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   27161   poly16x4x4_t __s1 = __p1; \
   27162   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
   27163 })
   27164 #else
   27165 #define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   27166   poly16x4x4_t __s1 = __p1; \
   27167   poly16x4x4_t __rev1; \
   27168   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27169   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27170   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27171   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27172   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
   27173 })
   27174 #endif
   27175 
   27176 #ifdef __LITTLE_ENDIAN__
   27177 #define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   27178   poly16x8x4_t __s1 = __p1; \
   27179   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
   27180 })
   27181 #else
   27182 #define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
   27183   poly16x8x4_t __s1 = __p1; \
   27184   poly16x8x4_t __rev1; \
   27185   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   27186   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   27187   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   27188   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   27189   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
   27190 })
   27191 #endif
   27192 
   27193 #ifdef __LITTLE_ENDIAN__
   27194 #define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   27195   uint32x4x4_t __s1 = __p1; \
   27196   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
   27197 })
   27198 #else
   27199 #define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   27200   uint32x4x4_t __s1 = __p1; \
   27201   uint32x4x4_t __rev1; \
   27202   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27203   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27204   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27205   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27206   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
   27207 })
   27208 #endif
   27209 
   27210 #ifdef __LITTLE_ENDIAN__
   27211 #define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   27212   uint16x8x4_t __s1 = __p1; \
   27213   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
   27214 })
   27215 #else
   27216 #define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   27217   uint16x8x4_t __s1 = __p1; \
   27218   uint16x8x4_t __rev1; \
   27219   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   27220   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   27221   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   27222   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   27223   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
   27224 })
   27225 #endif
   27226 
   27227 #ifdef __LITTLE_ENDIAN__
   27228 #define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   27229   float32x4x4_t __s1 = __p1; \
   27230   __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
   27231 })
   27232 #else
   27233 #define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   27234   float32x4x4_t __s1 = __p1; \
   27235   float32x4x4_t __rev1; \
   27236   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27237   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27238   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27239   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27240   __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
   27241 })
   27242 #endif
   27243 
   27244 #ifdef __LITTLE_ENDIAN__
   27245 #define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   27246   float16x8x4_t __s1 = __p1; \
   27247   __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
   27248 })
   27249 #else
   27250 #define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   27251   float16x8x4_t __s1 = __p1; \
   27252   float16x8x4_t __rev1; \
   27253   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   27254   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   27255   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   27256   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   27257   __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
   27258 })
   27259 #endif
   27260 
   27261 #ifdef __LITTLE_ENDIAN__
   27262 #define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   27263   int32x4x4_t __s1 = __p1; \
   27264   __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
   27265 })
   27266 #else
   27267 #define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   27268   int32x4x4_t __s1 = __p1; \
   27269   int32x4x4_t __rev1; \
   27270   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27271   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27272   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27273   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27274   __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
   27275 })
   27276 #endif
   27277 
   27278 #ifdef __LITTLE_ENDIAN__
   27279 #define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   27280   int16x8x4_t __s1 = __p1; \
   27281   __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
   27282 })
   27283 #else
   27284 #define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   27285   int16x8x4_t __s1 = __p1; \
   27286   int16x8x4_t __rev1; \
   27287   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   27288   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   27289   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   27290   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   27291   __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
   27292 })
   27293 #endif
   27294 
   27295 #ifdef __LITTLE_ENDIAN__
   27296 #define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   27297   uint8x8x4_t __s1 = __p1; \
   27298   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
   27299 })
   27300 #else
   27301 #define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   27302   uint8x8x4_t __s1 = __p1; \
   27303   uint8x8x4_t __rev1; \
   27304   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   27305   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   27306   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   27307   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   27308   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
   27309 })
   27310 #endif
   27311 
   27312 #ifdef __LITTLE_ENDIAN__
   27313 #define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   27314   uint32x2x4_t __s1 = __p1; \
   27315   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
   27316 })
   27317 #else
   27318 #define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   27319   uint32x2x4_t __s1 = __p1; \
   27320   uint32x2x4_t __rev1; \
   27321   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   27322   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   27323   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   27324   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   27325   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
   27326 })
   27327 #endif
   27328 
   27329 #ifdef __LITTLE_ENDIAN__
   27330 #define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   27331   uint16x4x4_t __s1 = __p1; \
   27332   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
   27333 })
   27334 #else
   27335 #define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   27336   uint16x4x4_t __s1 = __p1; \
   27337   uint16x4x4_t __rev1; \
   27338   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27339   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27340   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27341   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27342   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
   27343 })
   27344 #endif
   27345 
   27346 #ifdef __LITTLE_ENDIAN__
   27347 #define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   27348   int8x8x4_t __s1 = __p1; \
   27349   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
   27350 })
   27351 #else
   27352 #define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   27353   int8x8x4_t __s1 = __p1; \
   27354   int8x8x4_t __rev1; \
   27355   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   27356   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   27357   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   27358   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   27359   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
   27360 })
   27361 #endif
   27362 
   27363 #ifdef __LITTLE_ENDIAN__
   27364 #define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   27365   float32x2x4_t __s1 = __p1; \
   27366   __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
   27367 })
   27368 #else
   27369 #define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   27370   float32x2x4_t __s1 = __p1; \
   27371   float32x2x4_t __rev1; \
   27372   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   27373   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   27374   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   27375   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   27376   __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
   27377 })
   27378 #endif
   27379 
   27380 #ifdef __LITTLE_ENDIAN__
   27381 #define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   27382   float16x4x4_t __s1 = __p1; \
   27383   __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
   27384 })
   27385 #else
   27386 #define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
   27387   float16x4x4_t __s1 = __p1; \
   27388   float16x4x4_t __rev1; \
   27389   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27390   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27391   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27392   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27393   __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
   27394 })
   27395 #endif
   27396 
   27397 #ifdef __LITTLE_ENDIAN__
   27398 #define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   27399   int32x2x4_t __s1 = __p1; \
   27400   __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
   27401 })
   27402 #else
   27403 #define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   27404   int32x2x4_t __s1 = __p1; \
   27405   int32x2x4_t __rev1; \
   27406   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   27407   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   27408   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   27409   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   27410   __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
   27411 })
   27412 #endif
   27413 
   27414 #ifdef __LITTLE_ENDIAN__
   27415 #define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   27416   int16x4x4_t __s1 = __p1; \
   27417   __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
   27418 })
   27419 #else
   27420 #define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   27421   int16x4x4_t __s1 = __p1; \
   27422   int16x4x4_t __rev1; \
   27423   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   27424   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   27425   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   27426   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   27427   __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
   27428 })
   27429 #endif
   27430 
   27431 #ifdef __LITTLE_ENDIAN__
   27432 __ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   27433   uint8x16_t __ret;
   27434   __ret = __p0 - __p1;
   27435   return __ret;
   27436 }
   27437 #else
   27438 __ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   27439   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   27440   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   27441   uint8x16_t __ret;
   27442   __ret = __rev0 - __rev1;
   27443   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   27444   return __ret;
   27445 }
   27446 #endif
   27447 
   27448 #ifdef __LITTLE_ENDIAN__
   27449 __ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   27450   uint32x4_t __ret;
   27451   __ret = __p0 - __p1;
   27452   return __ret;
   27453 }
   27454 #else
   27455 __ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   27456   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   27457   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   27458   uint32x4_t __ret;
   27459   __ret = __rev0 - __rev1;
   27460   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   27461   return __ret;
   27462 }
   27463 #endif
   27464 
   27465 #ifdef __LITTLE_ENDIAN__
   27466 __ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   27467   uint64x2_t __ret;
   27468   __ret = __p0 - __p1;
   27469   return __ret;
   27470 }
   27471 #else
   27472 __ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   27473   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27474   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27475   uint64x2_t __ret;
   27476   __ret = __rev0 - __rev1;
   27477   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27478   return __ret;
   27479 }
   27480 #endif
   27481 
   27482 #ifdef __LITTLE_ENDIAN__
   27483 __ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   27484   uint16x8_t __ret;
   27485   __ret = __p0 - __p1;
   27486   return __ret;
   27487 }
   27488 #else
   27489 __ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   27490   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   27491   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   27492   uint16x8_t __ret;
   27493   __ret = __rev0 - __rev1;
   27494   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   27495   return __ret;
   27496 }
   27497 #endif
   27498 
   27499 #ifdef __LITTLE_ENDIAN__
   27500 __ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
   27501   int8x16_t __ret;
   27502   __ret = __p0 - __p1;
   27503   return __ret;
   27504 }
   27505 #else
   27506 __ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
   27507   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   27508   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   27509   int8x16_t __ret;
   27510   __ret = __rev0 - __rev1;
   27511   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   27512   return __ret;
   27513 }
   27514 #endif
   27515 
   27516 #ifdef __LITTLE_ENDIAN__
   27517 __ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
   27518   float32x4_t __ret;
   27519   __ret = __p0 - __p1;
   27520   return __ret;
   27521 }
   27522 #else
   27523 __ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
   27524   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   27525   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   27526   float32x4_t __ret;
   27527   __ret = __rev0 - __rev1;
   27528   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   27529   return __ret;
   27530 }
   27531 #endif
   27532 
   27533 #ifdef __LITTLE_ENDIAN__
   27534 __ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
   27535   int32x4_t __ret;
   27536   __ret = __p0 - __p1;
   27537   return __ret;
   27538 }
   27539 #else
   27540 __ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
   27541   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   27542   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   27543   int32x4_t __ret;
   27544   __ret = __rev0 - __rev1;
   27545   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   27546   return __ret;
   27547 }
   27548 #endif
   27549 
   27550 #ifdef __LITTLE_ENDIAN__
   27551 __ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
   27552   int64x2_t __ret;
   27553   __ret = __p0 - __p1;
   27554   return __ret;
   27555 }
   27556 #else
   27557 __ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
   27558   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27559   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27560   int64x2_t __ret;
   27561   __ret = __rev0 - __rev1;
   27562   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27563   return __ret;
   27564 }
   27565 #endif
   27566 
   27567 #ifdef __LITTLE_ENDIAN__
   27568 __ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
   27569   int16x8_t __ret;
   27570   __ret = __p0 - __p1;
   27571   return __ret;
   27572 }
   27573 #else
   27574 __ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
   27575   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   27576   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   27577   int16x8_t __ret;
   27578   __ret = __rev0 - __rev1;
   27579   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   27580   return __ret;
   27581 }
   27582 #endif
   27583 
   27584 #ifdef __LITTLE_ENDIAN__
   27585 __ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
   27586   uint8x8_t __ret;
   27587   __ret = __p0 - __p1;
   27588   return __ret;
   27589 }
   27590 #else
   27591 __ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
   27592   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   27593   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   27594   uint8x8_t __ret;
   27595   __ret = __rev0 - __rev1;
   27596   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   27597   return __ret;
   27598 }
   27599 #endif
   27600 
   27601 #ifdef __LITTLE_ENDIAN__
   27602 __ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
   27603   uint32x2_t __ret;
   27604   __ret = __p0 - __p1;
   27605   return __ret;
   27606 }
   27607 #else
   27608 __ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
   27609   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27610   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27611   uint32x2_t __ret;
   27612   __ret = __rev0 - __rev1;
   27613   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27614   return __ret;
   27615 }
   27616 #endif
   27617 
   27618 #ifdef __LITTLE_ENDIAN__
   27619 __ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
   27620   uint64x1_t __ret;
   27621   __ret = __p0 - __p1;
   27622   return __ret;
   27623 }
   27624 #else
   27625 __ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
   27626   uint64x1_t __ret;
   27627   __ret = __p0 - __p1;
   27628   return __ret;
   27629 }
   27630 #endif
   27631 
   27632 #ifdef __LITTLE_ENDIAN__
   27633 __ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
   27634   uint16x4_t __ret;
   27635   __ret = __p0 - __p1;
   27636   return __ret;
   27637 }
   27638 #else
   27639 __ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
   27640   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   27641   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   27642   uint16x4_t __ret;
   27643   __ret = __rev0 - __rev1;
   27644   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   27645   return __ret;
   27646 }
   27647 #endif
   27648 
   27649 #ifdef __LITTLE_ENDIAN__
   27650 __ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
   27651   int8x8_t __ret;
   27652   __ret = __p0 - __p1;
   27653   return __ret;
   27654 }
   27655 #else
   27656 __ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
   27657   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   27658   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   27659   int8x8_t __ret;
   27660   __ret = __rev0 - __rev1;
   27661   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   27662   return __ret;
   27663 }
   27664 #endif
   27665 
   27666 #ifdef __LITTLE_ENDIAN__
   27667 __ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
   27668   float32x2_t __ret;
   27669   __ret = __p0 - __p1;
   27670   return __ret;
   27671 }
   27672 #else
   27673 __ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
   27674   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27675   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27676   float32x2_t __ret;
   27677   __ret = __rev0 - __rev1;
   27678   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27679   return __ret;
   27680 }
   27681 #endif
   27682 
   27683 #ifdef __LITTLE_ENDIAN__
   27684 __ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
   27685   int32x2_t __ret;
   27686   __ret = __p0 - __p1;
   27687   return __ret;
   27688 }
   27689 #else
   27690 __ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
   27691   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27692   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27693   int32x2_t __ret;
   27694   __ret = __rev0 - __rev1;
   27695   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27696   return __ret;
   27697 }
   27698 #endif
   27699 
   27700 #ifdef __LITTLE_ENDIAN__
   27701 __ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
   27702   int64x1_t __ret;
   27703   __ret = __p0 - __p1;
   27704   return __ret;
   27705 }
   27706 #else
   27707 __ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
   27708   int64x1_t __ret;
   27709   __ret = __p0 - __p1;
   27710   return __ret;
   27711 }
   27712 #endif
   27713 
   27714 #ifdef __LITTLE_ENDIAN__
   27715 __ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
   27716   int16x4_t __ret;
   27717   __ret = __p0 - __p1;
   27718   return __ret;
   27719 }
   27720 #else
   27721 __ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
   27722   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   27723   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   27724   int16x4_t __ret;
   27725   __ret = __rev0 - __rev1;
   27726   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   27727   return __ret;
   27728 }
   27729 #endif
   27730 
   27731 #ifdef __LITTLE_ENDIAN__
   27732 __ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   27733   uint16x4_t __ret;
   27734   __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
   27735   return __ret;
   27736 }
   27737 #else
   27738 __ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   27739   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   27740   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   27741   uint16x4_t __ret;
   27742   __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
   27743   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   27744   return __ret;
   27745 }
   27746 __ai uint16x4_t __noswap_vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
   27747   uint16x4_t __ret;
   27748   __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
   27749   return __ret;
   27750 }
   27751 #endif
   27752 
   27753 #ifdef __LITTLE_ENDIAN__
   27754 __ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   27755   uint32x2_t __ret;
   27756   __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
   27757   return __ret;
   27758 }
   27759 #else
   27760 __ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   27761   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27762   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27763   uint32x2_t __ret;
   27764   __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
   27765   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27766   return __ret;
   27767 }
   27768 __ai uint32x2_t __noswap_vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
   27769   uint32x2_t __ret;
   27770   __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
   27771   return __ret;
   27772 }
   27773 #endif
   27774 
   27775 #ifdef __LITTLE_ENDIAN__
   27776 __ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   27777   uint8x8_t __ret;
   27778   __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
   27779   return __ret;
   27780 }
   27781 #else
   27782 __ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   27783   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   27784   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   27785   uint8x8_t __ret;
   27786   __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
   27787   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   27788   return __ret;
   27789 }
   27790 __ai uint8x8_t __noswap_vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
   27791   uint8x8_t __ret;
   27792   __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
   27793   return __ret;
   27794 }
   27795 #endif
   27796 
   27797 #ifdef __LITTLE_ENDIAN__
   27798 __ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
   27799   int16x4_t __ret;
   27800   __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
   27801   return __ret;
   27802 }
   27803 #else
   27804 __ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
   27805   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   27806   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   27807   int16x4_t __ret;
   27808   __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
   27809   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   27810   return __ret;
   27811 }
   27812 __ai int16x4_t __noswap_vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
   27813   int16x4_t __ret;
   27814   __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
   27815   return __ret;
   27816 }
   27817 #endif
   27818 
   27819 #ifdef __LITTLE_ENDIAN__
   27820 __ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
   27821   int32x2_t __ret;
   27822   __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
   27823   return __ret;
   27824 }
   27825 #else
   27826 __ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
   27827   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27828   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27829   int32x2_t __ret;
   27830   __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
   27831   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27832   return __ret;
   27833 }
   27834 __ai int32x2_t __noswap_vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
   27835   int32x2_t __ret;
   27836   __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
   27837   return __ret;
   27838 }
   27839 #endif
   27840 
   27841 #ifdef __LITTLE_ENDIAN__
   27842 __ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
   27843   int8x8_t __ret;
   27844   __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
   27845   return __ret;
   27846 }
   27847 #else
   27848 __ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
   27849   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   27850   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   27851   int8x8_t __ret;
   27852   __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
   27853   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   27854   return __ret;
   27855 }
   27856 __ai int8x8_t __noswap_vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
   27857   int8x8_t __ret;
   27858   __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
   27859   return __ret;
   27860 }
   27861 #endif
   27862 
   27863 #ifdef __LITTLE_ENDIAN__
   27864 __ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
   27865   uint16x8_t __ret;
   27866   __ret = vmovl_u8(__p0) - vmovl_u8(__p1);
   27867   return __ret;
   27868 }
   27869 #else
   27870 __ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
   27871   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   27872   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   27873   uint16x8_t __ret;
   27874   __ret = __noswap_vmovl_u8(__rev0) - __noswap_vmovl_u8(__rev1);
   27875   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   27876   return __ret;
   27877 }
   27878 #endif
   27879 
   27880 #ifdef __LITTLE_ENDIAN__
   27881 __ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
   27882   uint64x2_t __ret;
   27883   __ret = vmovl_u32(__p0) - vmovl_u32(__p1);
   27884   return __ret;
   27885 }
   27886 #else
   27887 __ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
   27888   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27889   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27890   uint64x2_t __ret;
   27891   __ret = __noswap_vmovl_u32(__rev0) - __noswap_vmovl_u32(__rev1);
   27892   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27893   return __ret;
   27894 }
   27895 #endif
   27896 
   27897 #ifdef __LITTLE_ENDIAN__
   27898 __ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
   27899   uint32x4_t __ret;
   27900   __ret = vmovl_u16(__p0) - vmovl_u16(__p1);
   27901   return __ret;
   27902 }
   27903 #else
   27904 __ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
   27905   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   27906   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   27907   uint32x4_t __ret;
   27908   __ret = __noswap_vmovl_u16(__rev0) - __noswap_vmovl_u16(__rev1);
   27909   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   27910   return __ret;
   27911 }
   27912 #endif
   27913 
   27914 #ifdef __LITTLE_ENDIAN__
   27915 __ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
   27916   int16x8_t __ret;
   27917   __ret = vmovl_s8(__p0) - vmovl_s8(__p1);
   27918   return __ret;
   27919 }
   27920 #else
   27921 __ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
   27922   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   27923   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   27924   int16x8_t __ret;
   27925   __ret = __noswap_vmovl_s8(__rev0) - __noswap_vmovl_s8(__rev1);
   27926   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   27927   return __ret;
   27928 }
   27929 #endif
   27930 
   27931 #ifdef __LITTLE_ENDIAN__
   27932 __ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
   27933   int64x2_t __ret;
   27934   __ret = vmovl_s32(__p0) - vmovl_s32(__p1);
   27935   return __ret;
   27936 }
   27937 #else
   27938 __ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
   27939   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27940   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27941   int64x2_t __ret;
   27942   __ret = __noswap_vmovl_s32(__rev0) - __noswap_vmovl_s32(__rev1);
   27943   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27944   return __ret;
   27945 }
   27946 #endif
   27947 
   27948 #ifdef __LITTLE_ENDIAN__
   27949 __ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
   27950   int32x4_t __ret;
   27951   __ret = vmovl_s16(__p0) - vmovl_s16(__p1);
   27952   return __ret;
   27953 }
   27954 #else
   27955 __ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
   27956   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   27957   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   27958   int32x4_t __ret;
   27959   __ret = __noswap_vmovl_s16(__rev0) - __noswap_vmovl_s16(__rev1);
   27960   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   27961   return __ret;
   27962 }
   27963 #endif
   27964 
   27965 #ifdef __LITTLE_ENDIAN__
   27966 __ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
   27967   uint16x8_t __ret;
   27968   __ret = __p0 - vmovl_u8(__p1);
   27969   return __ret;
   27970 }
   27971 #else
   27972 __ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
   27973   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   27974   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   27975   uint16x8_t __ret;
   27976   __ret = __rev0 - __noswap_vmovl_u8(__rev1);
   27977   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   27978   return __ret;
   27979 }
   27980 #endif
   27981 
   27982 #ifdef __LITTLE_ENDIAN__
   27983 __ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
   27984   uint64x2_t __ret;
   27985   __ret = __p0 - vmovl_u32(__p1);
   27986   return __ret;
   27987 }
   27988 #else
   27989 __ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
   27990   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   27991   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   27992   uint64x2_t __ret;
   27993   __ret = __rev0 - __noswap_vmovl_u32(__rev1);
   27994   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   27995   return __ret;
   27996 }
   27997 #endif
   27998 
   27999 #ifdef __LITTLE_ENDIAN__
   28000 __ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
   28001   uint32x4_t __ret;
   28002   __ret = __p0 - vmovl_u16(__p1);
   28003   return __ret;
   28004 }
   28005 #else
   28006 __ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
   28007   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28008   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28009   uint32x4_t __ret;
   28010   __ret = __rev0 - __noswap_vmovl_u16(__rev1);
   28011   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   28012   return __ret;
   28013 }
   28014 #endif
   28015 
   28016 #ifdef __LITTLE_ENDIAN__
   28017 __ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
   28018   int16x8_t __ret;
   28019   __ret = __p0 - vmovl_s8(__p1);
   28020   return __ret;
   28021 }
   28022 #else
   28023 __ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
   28024   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28025   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28026   int16x8_t __ret;
   28027   __ret = __rev0 - __noswap_vmovl_s8(__rev1);
   28028   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28029   return __ret;
   28030 }
   28031 #endif
   28032 
   28033 #ifdef __LITTLE_ENDIAN__
   28034 __ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
   28035   int64x2_t __ret;
   28036   __ret = __p0 - vmovl_s32(__p1);
   28037   return __ret;
   28038 }
   28039 #else
   28040 __ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
   28041   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   28042   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   28043   int64x2_t __ret;
   28044   __ret = __rev0 - __noswap_vmovl_s32(__rev1);
   28045   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   28046   return __ret;
   28047 }
   28048 #endif
   28049 
   28050 #ifdef __LITTLE_ENDIAN__
   28051 __ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
   28052   int32x4_t __ret;
   28053   __ret = __p0 - vmovl_s16(__p1);
   28054   return __ret;
   28055 }
   28056 #else
   28057 __ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
   28058   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28059   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28060   int32x4_t __ret;
   28061   __ret = __rev0 - __noswap_vmovl_s16(__rev1);
   28062   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   28063   return __ret;
   28064 }
   28065 #endif
   28066 
   28067 #ifdef __LITTLE_ENDIAN__
   28068 __ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
   28069   poly8x8_t __ret;
   28070   __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
   28071   return __ret;
   28072 }
   28073 #else
   28074 __ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
   28075   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28076   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28077   poly8x8_t __ret;
   28078   __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
   28079   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28080   return __ret;
   28081 }
   28082 #endif
   28083 
   28084 #ifdef __LITTLE_ENDIAN__
   28085 __ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
   28086   uint8x8_t __ret;
   28087   __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   28088   return __ret;
   28089 }
   28090 #else
   28091 __ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
   28092   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28093   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28094   uint8x8_t __ret;
   28095   __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   28096   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28097   return __ret;
   28098 }
   28099 #endif
   28100 
   28101 #ifdef __LITTLE_ENDIAN__
   28102 __ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
   28103   int8x8_t __ret;
   28104   __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   28105   return __ret;
   28106 }
   28107 #else
   28108 __ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
   28109   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28110   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28111   int8x8_t __ret;
   28112   __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   28113   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28114   return __ret;
   28115 }
   28116 #endif
   28117 
   28118 #ifdef __LITTLE_ENDIAN__
   28119 __ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
   28120   poly8x8_t __ret;
   28121   __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 4);
   28122   return __ret;
   28123 }
   28124 #else
   28125 __ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
   28126   poly8x8x2_t __rev0;
   28127   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28128   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28129   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28130   poly8x8_t __ret;
   28131   __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 4);
   28132   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28133   return __ret;
   28134 }
   28135 #endif
   28136 
   28137 #ifdef __LITTLE_ENDIAN__
   28138 __ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
   28139   uint8x8_t __ret;
   28140   __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 16);
   28141   return __ret;
   28142 }
   28143 #else
   28144 __ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
   28145   uint8x8x2_t __rev0;
   28146   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28147   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28148   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28149   uint8x8_t __ret;
   28150   __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 16);
   28151   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28152   return __ret;
   28153 }
   28154 #endif
   28155 
   28156 #ifdef __LITTLE_ENDIAN__
   28157 __ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
   28158   int8x8_t __ret;
   28159   __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 0);
   28160   return __ret;
   28161 }
   28162 #else
   28163 __ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
   28164   int8x8x2_t __rev0;
   28165   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28166   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28167   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28168   int8x8_t __ret;
   28169   __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 0);
   28170   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28171   return __ret;
   28172 }
   28173 #endif
   28174 
   28175 #ifdef __LITTLE_ENDIAN__
   28176 __ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
   28177   poly8x8_t __ret;
   28178   __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 4);
   28179   return __ret;
   28180 }
   28181 #else
   28182 __ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
   28183   poly8x8x3_t __rev0;
   28184   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28185   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28186   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28187   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28188   poly8x8_t __ret;
   28189   __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 4);
   28190   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28191   return __ret;
   28192 }
   28193 #endif
   28194 
   28195 #ifdef __LITTLE_ENDIAN__
   28196 __ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
   28197   uint8x8_t __ret;
   28198   __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 16);
   28199   return __ret;
   28200 }
   28201 #else
   28202 __ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
   28203   uint8x8x3_t __rev0;
   28204   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28205   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28206   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28207   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28208   uint8x8_t __ret;
   28209   __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 16);
   28210   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28211   return __ret;
   28212 }
   28213 #endif
   28214 
   28215 #ifdef __LITTLE_ENDIAN__
   28216 __ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
   28217   int8x8_t __ret;
   28218   __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 0);
   28219   return __ret;
   28220 }
   28221 #else
   28222 __ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
   28223   int8x8x3_t __rev0;
   28224   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28225   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28226   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28227   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28228   int8x8_t __ret;
   28229   __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 0);
   28230   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28231   return __ret;
   28232 }
   28233 #endif
   28234 
   28235 #ifdef __LITTLE_ENDIAN__
   28236 __ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
   28237   poly8x8_t __ret;
   28238   __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 4);
   28239   return __ret;
   28240 }
   28241 #else
   28242 __ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
   28243   poly8x8x4_t __rev0;
   28244   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28245   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28246   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28247   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
   28248   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28249   poly8x8_t __ret;
   28250   __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 4);
   28251   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28252   return __ret;
   28253 }
   28254 #endif
   28255 
   28256 #ifdef __LITTLE_ENDIAN__
   28257 __ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
   28258   uint8x8_t __ret;
   28259   __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 16);
   28260   return __ret;
   28261 }
   28262 #else
   28263 __ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
   28264   uint8x8x4_t __rev0;
   28265   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28266   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28267   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28268   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
   28269   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28270   uint8x8_t __ret;
   28271   __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 16);
   28272   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28273   return __ret;
   28274 }
   28275 #endif
   28276 
   28277 #ifdef __LITTLE_ENDIAN__
   28278 __ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
   28279   int8x8_t __ret;
   28280   __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 0);
   28281   return __ret;
   28282 }
   28283 #else
   28284 __ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
   28285   int8x8x4_t __rev0;
   28286   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28287   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28288   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28289   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
   28290   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28291   int8x8_t __ret;
   28292   __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 0);
   28293   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28294   return __ret;
   28295 }
   28296 #endif
   28297 
   28298 #ifdef __LITTLE_ENDIAN__
   28299 __ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
   28300   poly8x8_t __ret;
   28301   __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
   28302   return __ret;
   28303 }
   28304 #else
   28305 __ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
   28306   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28307   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28308   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28309   poly8x8_t __ret;
   28310   __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
   28311   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28312   return __ret;
   28313 }
   28314 #endif
   28315 
   28316 #ifdef __LITTLE_ENDIAN__
   28317 __ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   28318   uint8x8_t __ret;
   28319   __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
   28320   return __ret;
   28321 }
   28322 #else
   28323 __ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   28324   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28325   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28326   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28327   uint8x8_t __ret;
   28328   __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
   28329   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28330   return __ret;
   28331 }
   28332 #endif
   28333 
   28334 #ifdef __LITTLE_ENDIAN__
   28335 __ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   28336   int8x8_t __ret;
   28337   __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
   28338   return __ret;
   28339 }
   28340 #else
   28341 __ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   28342   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28343   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28344   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28345   int8x8_t __ret;
   28346   __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
   28347   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28348   return __ret;
   28349 }
   28350 #endif
   28351 
   28352 #ifdef __LITTLE_ENDIAN__
   28353 __ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
   28354   poly8x8_t __ret;
   28355   __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 4);
   28356   return __ret;
   28357 }
   28358 #else
   28359 __ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
   28360   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28361   poly8x8x2_t __rev1;
   28362   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28363   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28364   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28365   poly8x8_t __ret;
   28366   __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 4);
   28367   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28368   return __ret;
   28369 }
   28370 #endif
   28371 
   28372 #ifdef __LITTLE_ENDIAN__
   28373 __ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
   28374   uint8x8_t __ret;
   28375   __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 16);
   28376   return __ret;
   28377 }
   28378 #else
   28379 __ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
   28380   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28381   uint8x8x2_t __rev1;
   28382   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28383   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28384   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28385   uint8x8_t __ret;
   28386   __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 16);
   28387   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28388   return __ret;
   28389 }
   28390 #endif
   28391 
   28392 #ifdef __LITTLE_ENDIAN__
   28393 __ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
   28394   int8x8_t __ret;
   28395   __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 0);
   28396   return __ret;
   28397 }
   28398 #else
   28399 __ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
   28400   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28401   int8x8x2_t __rev1;
   28402   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28403   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28404   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28405   int8x8_t __ret;
   28406   __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 0);
   28407   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28408   return __ret;
   28409 }
   28410 #endif
   28411 
   28412 #ifdef __LITTLE_ENDIAN__
   28413 __ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
   28414   poly8x8_t __ret;
   28415   __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 4);
   28416   return __ret;
   28417 }
   28418 #else
   28419 __ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
   28420   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28421   poly8x8x3_t __rev1;
   28422   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28423   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28424   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28425   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28426   poly8x8_t __ret;
   28427   __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 4);
   28428   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28429   return __ret;
   28430 }
   28431 #endif
   28432 
   28433 #ifdef __LITTLE_ENDIAN__
   28434 __ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
   28435   uint8x8_t __ret;
   28436   __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 16);
   28437   return __ret;
   28438 }
   28439 #else
   28440 __ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
   28441   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28442   uint8x8x3_t __rev1;
   28443   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28444   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28445   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28446   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28447   uint8x8_t __ret;
   28448   __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 16);
   28449   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28450   return __ret;
   28451 }
   28452 #endif
   28453 
   28454 #ifdef __LITTLE_ENDIAN__
   28455 __ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
   28456   int8x8_t __ret;
   28457   __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 0);
   28458   return __ret;
   28459 }
   28460 #else
   28461 __ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
   28462   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28463   int8x8x3_t __rev1;
   28464   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28465   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28466   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28467   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28468   int8x8_t __ret;
   28469   __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 0);
   28470   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28471   return __ret;
   28472 }
   28473 #endif
   28474 
   28475 #ifdef __LITTLE_ENDIAN__
   28476 __ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
   28477   poly8x8_t __ret;
   28478   __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 4);
   28479   return __ret;
   28480 }
   28481 #else
   28482 __ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
   28483   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28484   poly8x8x4_t __rev1;
   28485   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28486   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28487   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28488   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
   28489   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28490   poly8x8_t __ret;
   28491   __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 4);
   28492   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28493   return __ret;
   28494 }
   28495 #endif
   28496 
   28497 #ifdef __LITTLE_ENDIAN__
   28498 __ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
   28499   uint8x8_t __ret;
   28500   __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 16);
   28501   return __ret;
   28502 }
   28503 #else
   28504 __ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
   28505   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28506   uint8x8x4_t __rev1;
   28507   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28508   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28509   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28510   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
   28511   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28512   uint8x8_t __ret;
   28513   __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 16);
   28514   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28515   return __ret;
   28516 }
   28517 #endif
   28518 
   28519 #ifdef __LITTLE_ENDIAN__
   28520 __ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
   28521   int8x8_t __ret;
   28522   __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 0);
   28523   return __ret;
   28524 }
   28525 #else
   28526 __ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
   28527   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28528   int8x8x4_t __rev1;
   28529   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28530   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28531   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
   28532   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
   28533   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   28534   int8x8_t __ret;
   28535   __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 0);
   28536   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28537   return __ret;
   28538 }
   28539 #endif
   28540 
   28541 #ifdef __LITTLE_ENDIAN__
   28542 __ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
   28543   poly8x8x2_t __ret;
   28544   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
   28545   return __ret;
   28546 }
   28547 #else
   28548 __ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
   28549   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28550   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28551   poly8x8x2_t __ret;
   28552   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
   28553 
   28554   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28555   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28556   return __ret;
   28557 }
   28558 #endif
   28559 
   28560 #ifdef __LITTLE_ENDIAN__
   28561 __ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
   28562   poly16x4x2_t __ret;
   28563   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
   28564   return __ret;
   28565 }
   28566 #else
   28567 __ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
   28568   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28569   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28570   poly16x4x2_t __ret;
   28571   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
   28572 
   28573   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   28574   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   28575   return __ret;
   28576 }
   28577 #endif
   28578 
   28579 #ifdef __LITTLE_ENDIAN__
   28580 __ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   28581   poly8x16x2_t __ret;
   28582   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
   28583   return __ret;
   28584 }
   28585 #else
   28586 __ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   28587   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28588   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28589   poly8x16x2_t __ret;
   28590   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
   28591 
   28592   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28593   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28594   return __ret;
   28595 }
   28596 #endif
   28597 
   28598 #ifdef __LITTLE_ENDIAN__
   28599 __ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
   28600   poly16x8x2_t __ret;
   28601   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
   28602   return __ret;
   28603 }
   28604 #else
   28605 __ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
   28606   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28607   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28608   poly16x8x2_t __ret;
   28609   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
   28610 
   28611   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28612   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28613   return __ret;
   28614 }
   28615 #endif
   28616 
   28617 #ifdef __LITTLE_ENDIAN__
   28618 __ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   28619   uint8x16x2_t __ret;
   28620   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
   28621   return __ret;
   28622 }
   28623 #else
   28624 __ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   28625   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28626   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28627   uint8x16x2_t __ret;
   28628   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   28629 
   28630   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28631   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28632   return __ret;
   28633 }
   28634 #endif
   28635 
   28636 #ifdef __LITTLE_ENDIAN__
   28637 __ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   28638   uint32x4x2_t __ret;
   28639   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
   28640   return __ret;
   28641 }
   28642 #else
   28643 __ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   28644   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28645   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28646   uint32x4x2_t __ret;
   28647   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   28648 
   28649   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   28650   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   28651   return __ret;
   28652 }
   28653 #endif
   28654 
   28655 #ifdef __LITTLE_ENDIAN__
   28656 __ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   28657   uint16x8x2_t __ret;
   28658   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
   28659   return __ret;
   28660 }
   28661 #else
   28662 __ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   28663   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28664   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28665   uint16x8x2_t __ret;
   28666   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   28667 
   28668   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28669   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28670   return __ret;
   28671 }
   28672 #endif
   28673 
   28674 #ifdef __LITTLE_ENDIAN__
   28675 __ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
   28676   int8x16x2_t __ret;
   28677   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
   28678   return __ret;
   28679 }
   28680 #else
   28681 __ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
   28682   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28683   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28684   int8x16x2_t __ret;
   28685   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   28686 
   28687   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28688   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28689   return __ret;
   28690 }
   28691 #endif
   28692 
   28693 #ifdef __LITTLE_ENDIAN__
   28694 __ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
   28695   float32x4x2_t __ret;
   28696   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
   28697   return __ret;
   28698 }
   28699 #else
   28700 __ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
   28701   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28702   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28703   float32x4x2_t __ret;
   28704   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   28705 
   28706   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   28707   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   28708   return __ret;
   28709 }
   28710 #endif
   28711 
   28712 #ifdef __LITTLE_ENDIAN__
   28713 __ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
   28714   int32x4x2_t __ret;
   28715   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
   28716   return __ret;
   28717 }
   28718 #else
   28719 __ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
   28720   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28721   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28722   int32x4x2_t __ret;
   28723   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   28724 
   28725   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   28726   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   28727   return __ret;
   28728 }
   28729 #endif
   28730 
   28731 #ifdef __LITTLE_ENDIAN__
   28732 __ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
   28733   int16x8x2_t __ret;
   28734   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
   28735   return __ret;
   28736 }
   28737 #else
   28738 __ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
   28739   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28740   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28741   int16x8x2_t __ret;
   28742   __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   28743 
   28744   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28745   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28746   return __ret;
   28747 }
   28748 #endif
   28749 
   28750 #ifdef __LITTLE_ENDIAN__
   28751 __ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
   28752   uint8x8x2_t __ret;
   28753   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
   28754   return __ret;
   28755 }
   28756 #else
   28757 __ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
   28758   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28759   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28760   uint8x8x2_t __ret;
   28761   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   28762 
   28763   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28764   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28765   return __ret;
   28766 }
   28767 #endif
   28768 
   28769 #ifdef __LITTLE_ENDIAN__
   28770 __ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
   28771   uint32x2x2_t __ret;
   28772   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
   28773   return __ret;
   28774 }
   28775 #else
   28776 __ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
   28777   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   28778   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   28779   uint32x2x2_t __ret;
   28780   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   28781 
   28782   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
   28783   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
   28784   return __ret;
   28785 }
   28786 #endif
   28787 
   28788 #ifdef __LITTLE_ENDIAN__
   28789 __ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
   28790   uint16x4x2_t __ret;
   28791   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
   28792   return __ret;
   28793 }
   28794 #else
   28795 __ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
   28796   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28797   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28798   uint16x4x2_t __ret;
   28799   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   28800 
   28801   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   28802   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   28803   return __ret;
   28804 }
   28805 #endif
   28806 
   28807 #ifdef __LITTLE_ENDIAN__
   28808 __ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
   28809   int8x8x2_t __ret;
   28810   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
   28811   return __ret;
   28812 }
   28813 #else
   28814 __ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
   28815   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28816   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28817   int8x8x2_t __ret;
   28818   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   28819 
   28820   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   28821   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   28822   return __ret;
   28823 }
   28824 #endif
   28825 
   28826 #ifdef __LITTLE_ENDIAN__
   28827 __ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
   28828   float32x2x2_t __ret;
   28829   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
   28830   return __ret;
   28831 }
   28832 #else
   28833 __ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
   28834   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   28835   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   28836   float32x2x2_t __ret;
   28837   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   28838 
   28839   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
   28840   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
   28841   return __ret;
   28842 }
   28843 #endif
   28844 
   28845 #ifdef __LITTLE_ENDIAN__
   28846 __ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
   28847   int32x2x2_t __ret;
   28848   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
   28849   return __ret;
   28850 }
   28851 #else
   28852 __ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
   28853   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   28854   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   28855   int32x2x2_t __ret;
   28856   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   28857 
   28858   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
   28859   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
   28860   return __ret;
   28861 }
   28862 #endif
   28863 
   28864 #ifdef __LITTLE_ENDIAN__
   28865 __ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
   28866   int16x4x2_t __ret;
   28867   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
   28868   return __ret;
   28869 }
   28870 #else
   28871 __ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
   28872   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28873   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28874   int16x4x2_t __ret;
   28875   __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   28876 
   28877   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   28878   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   28879   return __ret;
   28880 }
   28881 #endif
   28882 
   28883 #ifdef __LITTLE_ENDIAN__
   28884 __ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
   28885   uint8x8_t __ret;
   28886   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   28887   return __ret;
   28888 }
   28889 #else
   28890 __ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
   28891   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28892   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28893   uint8x8_t __ret;
   28894   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   28895   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28896   return __ret;
   28897 }
   28898 #endif
   28899 
   28900 #ifdef __LITTLE_ENDIAN__
   28901 __ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
   28902   uint16x4_t __ret;
   28903   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   28904   return __ret;
   28905 }
   28906 #else
   28907 __ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
   28908   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28909   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28910   uint16x4_t __ret;
   28911   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   28912   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   28913   return __ret;
   28914 }
   28915 #endif
   28916 
   28917 #ifdef __LITTLE_ENDIAN__
   28918 __ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   28919   uint8x16_t __ret;
   28920   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   28921   return __ret;
   28922 }
   28923 #else
   28924 __ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   28925   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28926   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28927   uint8x16_t __ret;
   28928   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   28929   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28930   return __ret;
   28931 }
   28932 #endif
   28933 
   28934 #ifdef __LITTLE_ENDIAN__
   28935 __ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
   28936   uint16x8_t __ret;
   28937   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   28938   return __ret;
   28939 }
   28940 #else
   28941 __ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
   28942   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28943   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28944   uint16x8_t __ret;
   28945   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   28946   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28947   return __ret;
   28948 }
   28949 #endif
   28950 
   28951 #ifdef __LITTLE_ENDIAN__
   28952 __ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   28953   uint8x16_t __ret;
   28954   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   28955   return __ret;
   28956 }
   28957 #else
   28958 __ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   28959   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28960   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28961   uint8x16_t __ret;
   28962   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   28963   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   28964   return __ret;
   28965 }
   28966 #endif
   28967 
   28968 #ifdef __LITTLE_ENDIAN__
   28969 __ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   28970   uint32x4_t __ret;
   28971   __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   28972   return __ret;
   28973 }
   28974 #else
   28975 __ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   28976   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   28977   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   28978   uint32x4_t __ret;
   28979   __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   28980   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   28981   return __ret;
   28982 }
   28983 #endif
   28984 
   28985 #ifdef __LITTLE_ENDIAN__
   28986 __ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   28987   uint16x8_t __ret;
   28988   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   28989   return __ret;
   28990 }
   28991 #else
   28992 __ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   28993   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   28994   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   28995   uint16x8_t __ret;
   28996   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   28997   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   28998   return __ret;
   28999 }
   29000 #endif
   29001 
   29002 #ifdef __LITTLE_ENDIAN__
   29003 __ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
   29004   uint8x16_t __ret;
   29005   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   29006   return __ret;
   29007 }
   29008 #else
   29009 __ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
   29010   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29011   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29012   uint8x16_t __ret;
   29013   __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   29014   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29015   return __ret;
   29016 }
   29017 #endif
   29018 
   29019 #ifdef __LITTLE_ENDIAN__
   29020 __ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
   29021   uint32x4_t __ret;
   29022   __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   29023   return __ret;
   29024 }
   29025 #else
   29026 __ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
   29027   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29028   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29029   uint32x4_t __ret;
   29030   __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   29031   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   29032   return __ret;
   29033 }
   29034 #endif
   29035 
   29036 #ifdef __LITTLE_ENDIAN__
   29037 __ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
   29038   uint16x8_t __ret;
   29039   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   29040   return __ret;
   29041 }
   29042 #else
   29043 __ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
   29044   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29045   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29046   uint16x8_t __ret;
   29047   __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   29048   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   29049   return __ret;
   29050 }
   29051 #endif
   29052 
   29053 #ifdef __LITTLE_ENDIAN__
   29054 __ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
   29055   uint8x8_t __ret;
   29056   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   29057   return __ret;
   29058 }
   29059 #else
   29060 __ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
   29061   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29062   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29063   uint8x8_t __ret;
   29064   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   29065   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   29066   return __ret;
   29067 }
   29068 #endif
   29069 
   29070 #ifdef __LITTLE_ENDIAN__
   29071 __ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
   29072   uint32x2_t __ret;
   29073   __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   29074   return __ret;
   29075 }
   29076 #else
   29077 __ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
   29078   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   29079   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   29080   uint32x2_t __ret;
   29081   __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   29082   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   29083   return __ret;
   29084 }
   29085 #endif
   29086 
   29087 #ifdef __LITTLE_ENDIAN__
   29088 __ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
   29089   uint16x4_t __ret;
   29090   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   29091   return __ret;
   29092 }
   29093 #else
   29094 __ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
   29095   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29096   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29097   uint16x4_t __ret;
   29098   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   29099   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   29100   return __ret;
   29101 }
   29102 #endif
   29103 
   29104 #ifdef __LITTLE_ENDIAN__
   29105 __ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
   29106   uint8x8_t __ret;
   29107   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   29108   return __ret;
   29109 }
   29110 #else
   29111 __ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
   29112   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29113   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29114   uint8x8_t __ret;
   29115   __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   29116   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   29117   return __ret;
   29118 }
   29119 #endif
   29120 
   29121 #ifdef __LITTLE_ENDIAN__
   29122 __ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
   29123   uint32x2_t __ret;
   29124   __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   29125   return __ret;
   29126 }
   29127 #else
   29128 __ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
   29129   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   29130   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   29131   uint32x2_t __ret;
   29132   __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   29133   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   29134   return __ret;
   29135 }
   29136 #endif
   29137 
   29138 #ifdef __LITTLE_ENDIAN__
   29139 __ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
   29140   uint16x4_t __ret;
   29141   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   29142   return __ret;
   29143 }
   29144 #else
   29145 __ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
   29146   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29147   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29148   uint16x4_t __ret;
   29149   __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   29150   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   29151   return __ret;
   29152 }
   29153 #endif
   29154 
   29155 #ifdef __LITTLE_ENDIAN__
   29156 __ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
   29157   poly8x8x2_t __ret;
   29158   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
   29159   return __ret;
   29160 }
   29161 #else
   29162 __ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
   29163   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29164   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29165   poly8x8x2_t __ret;
   29166   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
   29167 
   29168   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29169   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29170   return __ret;
   29171 }
   29172 #endif
   29173 
   29174 #ifdef __LITTLE_ENDIAN__
   29175 __ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
   29176   poly16x4x2_t __ret;
   29177   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
   29178   return __ret;
   29179 }
   29180 #else
   29181 __ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
   29182   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29183   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29184   poly16x4x2_t __ret;
   29185   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
   29186 
   29187   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29188   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29189   return __ret;
   29190 }
   29191 #endif
   29192 
   29193 #ifdef __LITTLE_ENDIAN__
   29194 __ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   29195   poly8x16x2_t __ret;
   29196   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
   29197   return __ret;
   29198 }
   29199 #else
   29200 __ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   29201   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29202   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29203   poly8x16x2_t __ret;
   29204   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
   29205 
   29206   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29207   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29208   return __ret;
   29209 }
   29210 #endif
   29211 
   29212 #ifdef __LITTLE_ENDIAN__
   29213 __ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
   29214   poly16x8x2_t __ret;
   29215   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
   29216   return __ret;
   29217 }
   29218 #else
   29219 __ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
   29220   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29221   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29222   poly16x8x2_t __ret;
   29223   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
   29224 
   29225   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29226   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29227   return __ret;
   29228 }
   29229 #endif
   29230 
   29231 #ifdef __LITTLE_ENDIAN__
   29232 __ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   29233   uint8x16x2_t __ret;
   29234   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
   29235   return __ret;
   29236 }
   29237 #else
   29238 __ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   29239   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29240   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29241   uint8x16x2_t __ret;
   29242   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   29243 
   29244   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29245   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29246   return __ret;
   29247 }
   29248 #endif
   29249 
   29250 #ifdef __LITTLE_ENDIAN__
   29251 __ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   29252   uint32x4x2_t __ret;
   29253   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
   29254   return __ret;
   29255 }
   29256 #else
   29257 __ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   29258   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29259   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29260   uint32x4x2_t __ret;
   29261   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   29262 
   29263   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29264   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29265   return __ret;
   29266 }
   29267 #endif
   29268 
   29269 #ifdef __LITTLE_ENDIAN__
   29270 __ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   29271   uint16x8x2_t __ret;
   29272   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
   29273   return __ret;
   29274 }
   29275 #else
   29276 __ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   29277   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29278   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29279   uint16x8x2_t __ret;
   29280   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   29281 
   29282   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29283   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29284   return __ret;
   29285 }
   29286 #endif
   29287 
   29288 #ifdef __LITTLE_ENDIAN__
   29289 __ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
   29290   int8x16x2_t __ret;
   29291   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
   29292   return __ret;
   29293 }
   29294 #else
   29295 __ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
   29296   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29297   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29298   int8x16x2_t __ret;
   29299   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   29300 
   29301   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29302   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29303   return __ret;
   29304 }
   29305 #endif
   29306 
   29307 #ifdef __LITTLE_ENDIAN__
   29308 __ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
   29309   float32x4x2_t __ret;
   29310   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
   29311   return __ret;
   29312 }
   29313 #else
   29314 __ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
   29315   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29316   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29317   float32x4x2_t __ret;
   29318   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   29319 
   29320   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29321   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29322   return __ret;
   29323 }
   29324 #endif
   29325 
   29326 #ifdef __LITTLE_ENDIAN__
   29327 __ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
   29328   int32x4x2_t __ret;
   29329   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
   29330   return __ret;
   29331 }
   29332 #else
   29333 __ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
   29334   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29335   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29336   int32x4x2_t __ret;
   29337   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   29338 
   29339   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29340   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29341   return __ret;
   29342 }
   29343 #endif
   29344 
   29345 #ifdef __LITTLE_ENDIAN__
   29346 __ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
   29347   int16x8x2_t __ret;
   29348   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
   29349   return __ret;
   29350 }
   29351 #else
   29352 __ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
   29353   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29354   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29355   int16x8x2_t __ret;
   29356   __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   29357 
   29358   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29359   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29360   return __ret;
   29361 }
   29362 #endif
   29363 
   29364 #ifdef __LITTLE_ENDIAN__
   29365 __ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
   29366   uint8x8x2_t __ret;
   29367   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
   29368   return __ret;
   29369 }
   29370 #else
   29371 __ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
   29372   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29373   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29374   uint8x8x2_t __ret;
   29375   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   29376 
   29377   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29378   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29379   return __ret;
   29380 }
   29381 #endif
   29382 
   29383 #ifdef __LITTLE_ENDIAN__
   29384 __ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
   29385   uint32x2x2_t __ret;
   29386   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
   29387   return __ret;
   29388 }
   29389 #else
   29390 __ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
   29391   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   29392   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   29393   uint32x2x2_t __ret;
   29394   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   29395 
   29396   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
   29397   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
   29398   return __ret;
   29399 }
   29400 #endif
   29401 
   29402 #ifdef __LITTLE_ENDIAN__
   29403 __ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
   29404   uint16x4x2_t __ret;
   29405   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
   29406   return __ret;
   29407 }
   29408 #else
   29409 __ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
   29410   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29411   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29412   uint16x4x2_t __ret;
   29413   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   29414 
   29415   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29416   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29417   return __ret;
   29418 }
   29419 #endif
   29420 
   29421 #ifdef __LITTLE_ENDIAN__
   29422 __ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
   29423   int8x8x2_t __ret;
   29424   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
   29425   return __ret;
   29426 }
   29427 #else
   29428 __ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
   29429   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29430   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29431   int8x8x2_t __ret;
   29432   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   29433 
   29434   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29435   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29436   return __ret;
   29437 }
   29438 #endif
   29439 
   29440 #ifdef __LITTLE_ENDIAN__
   29441 __ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
   29442   float32x2x2_t __ret;
   29443   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
   29444   return __ret;
   29445 }
   29446 #else
   29447 __ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
   29448   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   29449   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   29450   float32x2x2_t __ret;
   29451   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   29452 
   29453   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
   29454   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
   29455   return __ret;
   29456 }
   29457 #endif
   29458 
   29459 #ifdef __LITTLE_ENDIAN__
   29460 __ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
   29461   int32x2x2_t __ret;
   29462   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
   29463   return __ret;
   29464 }
   29465 #else
   29466 __ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
   29467   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   29468   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   29469   int32x2x2_t __ret;
   29470   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   29471 
   29472   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
   29473   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
   29474   return __ret;
   29475 }
   29476 #endif
   29477 
   29478 #ifdef __LITTLE_ENDIAN__
   29479 __ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
   29480   int16x4x2_t __ret;
   29481   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
   29482   return __ret;
   29483 }
   29484 #else
   29485 __ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
   29486   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29487   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29488   int16x4x2_t __ret;
   29489   __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   29490 
   29491   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29492   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29493   return __ret;
   29494 }
   29495 #endif
   29496 
   29497 #ifdef __LITTLE_ENDIAN__
   29498 __ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
   29499   poly8x8x2_t __ret;
   29500   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
   29501   return __ret;
   29502 }
   29503 #else
   29504 __ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
   29505   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29506   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29507   poly8x8x2_t __ret;
   29508   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
   29509 
   29510   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29511   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29512   return __ret;
   29513 }
   29514 #endif
   29515 
   29516 #ifdef __LITTLE_ENDIAN__
   29517 __ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
   29518   poly16x4x2_t __ret;
   29519   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
   29520   return __ret;
   29521 }
   29522 #else
   29523 __ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
   29524   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29525   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29526   poly16x4x2_t __ret;
   29527   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
   29528 
   29529   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29530   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29531   return __ret;
   29532 }
   29533 #endif
   29534 
   29535 #ifdef __LITTLE_ENDIAN__
   29536 __ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   29537   poly8x16x2_t __ret;
   29538   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
   29539   return __ret;
   29540 }
   29541 #else
   29542 __ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
   29543   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29544   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29545   poly8x16x2_t __ret;
   29546   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
   29547 
   29548   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29549   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29550   return __ret;
   29551 }
   29552 #endif
   29553 
   29554 #ifdef __LITTLE_ENDIAN__
   29555 __ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
   29556   poly16x8x2_t __ret;
   29557   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
   29558   return __ret;
   29559 }
   29560 #else
   29561 __ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
   29562   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29563   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29564   poly16x8x2_t __ret;
   29565   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
   29566 
   29567   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29568   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29569   return __ret;
   29570 }
   29571 #endif
   29572 
   29573 #ifdef __LITTLE_ENDIAN__
   29574 __ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   29575   uint8x16x2_t __ret;
   29576   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
   29577   return __ret;
   29578 }
   29579 #else
   29580 __ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   29581   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29582   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29583   uint8x16x2_t __ret;
   29584   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   29585 
   29586   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29587   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29588   return __ret;
   29589 }
   29590 #endif
   29591 
   29592 #ifdef __LITTLE_ENDIAN__
   29593 __ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   29594   uint32x4x2_t __ret;
   29595   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
   29596   return __ret;
   29597 }
   29598 #else
   29599 __ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   29600   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29601   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29602   uint32x4x2_t __ret;
   29603   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   29604 
   29605   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29606   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29607   return __ret;
   29608 }
   29609 #endif
   29610 
   29611 #ifdef __LITTLE_ENDIAN__
   29612 __ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   29613   uint16x8x2_t __ret;
   29614   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
   29615   return __ret;
   29616 }
   29617 #else
   29618 __ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   29619   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29620   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29621   uint16x8x2_t __ret;
   29622   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   29623 
   29624   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29625   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29626   return __ret;
   29627 }
   29628 #endif
   29629 
   29630 #ifdef __LITTLE_ENDIAN__
   29631 __ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
   29632   int8x16x2_t __ret;
   29633   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
   29634   return __ret;
   29635 }
   29636 #else
   29637 __ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
   29638   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29639   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29640   int8x16x2_t __ret;
   29641   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   29642 
   29643   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29644   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   29645   return __ret;
   29646 }
   29647 #endif
   29648 
   29649 #ifdef __LITTLE_ENDIAN__
   29650 __ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
   29651   float32x4x2_t __ret;
   29652   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
   29653   return __ret;
   29654 }
   29655 #else
   29656 __ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
   29657   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29658   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29659   float32x4x2_t __ret;
   29660   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   29661 
   29662   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29663   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29664   return __ret;
   29665 }
   29666 #endif
   29667 
   29668 #ifdef __LITTLE_ENDIAN__
   29669 __ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
   29670   int32x4x2_t __ret;
   29671   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
   29672   return __ret;
   29673 }
   29674 #else
   29675 __ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
   29676   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29677   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29678   int32x4x2_t __ret;
   29679   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   29680 
   29681   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29682   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29683   return __ret;
   29684 }
   29685 #endif
   29686 
   29687 #ifdef __LITTLE_ENDIAN__
   29688 __ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
   29689   int16x8x2_t __ret;
   29690   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
   29691   return __ret;
   29692 }
   29693 #else
   29694 __ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
   29695   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29696   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29697   int16x8x2_t __ret;
   29698   __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   29699 
   29700   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29701   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29702   return __ret;
   29703 }
   29704 #endif
   29705 
   29706 #ifdef __LITTLE_ENDIAN__
   29707 __ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
   29708   uint8x8x2_t __ret;
   29709   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
   29710   return __ret;
   29711 }
   29712 #else
   29713 __ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
   29714   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29715   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29716   uint8x8x2_t __ret;
   29717   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   29718 
   29719   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29720   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29721   return __ret;
   29722 }
   29723 #endif
   29724 
   29725 #ifdef __LITTLE_ENDIAN__
   29726 __ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
   29727   uint32x2x2_t __ret;
   29728   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
   29729   return __ret;
   29730 }
   29731 #else
   29732 __ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
   29733   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   29734   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   29735   uint32x2x2_t __ret;
   29736   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   29737 
   29738   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
   29739   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
   29740   return __ret;
   29741 }
   29742 #endif
   29743 
   29744 #ifdef __LITTLE_ENDIAN__
   29745 __ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
   29746   uint16x4x2_t __ret;
   29747   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
   29748   return __ret;
   29749 }
   29750 #else
   29751 __ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
   29752   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29753   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29754   uint16x4x2_t __ret;
   29755   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   29756 
   29757   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29758   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29759   return __ret;
   29760 }
   29761 #endif
   29762 
   29763 #ifdef __LITTLE_ENDIAN__
   29764 __ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
   29765   int8x8x2_t __ret;
   29766   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
   29767   return __ret;
   29768 }
   29769 #else
   29770 __ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
   29771   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   29772   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   29773   int8x8x2_t __ret;
   29774   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   29775 
   29776   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
   29777   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
   29778   return __ret;
   29779 }
   29780 #endif
   29781 
   29782 #ifdef __LITTLE_ENDIAN__
   29783 __ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
   29784   float32x2x2_t __ret;
   29785   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
   29786   return __ret;
   29787 }
   29788 #else
   29789 __ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
   29790   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   29791   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   29792   float32x2x2_t __ret;
   29793   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   29794 
   29795   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
   29796   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
   29797   return __ret;
   29798 }
   29799 #endif
   29800 
   29801 #ifdef __LITTLE_ENDIAN__
   29802 __ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
   29803   int32x2x2_t __ret;
   29804   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
   29805   return __ret;
   29806 }
   29807 #else
   29808 __ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
   29809   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   29810   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   29811   int32x2x2_t __ret;
   29812   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   29813 
   29814   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
   29815   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
   29816   return __ret;
   29817 }
   29818 #endif
   29819 
   29820 #ifdef __LITTLE_ENDIAN__
   29821 __ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
   29822   int16x4x2_t __ret;
   29823   __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
   29824   return __ret;
   29825 }
   29826 #else
   29827 __ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
   29828   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   29829   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   29830   int16x4x2_t __ret;
   29831   __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   29832 
   29833   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
   29834   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
   29835   return __ret;
   29836 }
   29837 #endif
   29838 
   29839 #if !defined(__aarch64__)
   29840 #ifdef __LITTLE_ENDIAN__
   29841 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
   29842   poly8x8_t __ret;
   29843   __ret = (poly8x8_t)(__p0);
   29844   return __ret;
   29845 }
   29846 #else
   29847 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
   29848   poly8x8_t __ret;
   29849   __ret = (poly8x8_t)(__p0);
   29850   return __ret;
   29851 }
   29852 #endif
   29853 
   29854 #ifdef __LITTLE_ENDIAN__
   29855 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
   29856   poly8x8_t __ret;
   29857   __ret = (poly8x8_t)(__p0);
   29858   return __ret;
   29859 }
   29860 #else
   29861 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
   29862   poly8x8_t __ret;
   29863   __ret = (poly8x8_t)(__p0);
   29864   return __ret;
   29865 }
   29866 #endif
   29867 
   29868 #ifdef __LITTLE_ENDIAN__
   29869 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
   29870   poly8x8_t __ret;
   29871   __ret = (poly8x8_t)(__p0);
   29872   return __ret;
   29873 }
   29874 #else
   29875 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
   29876   poly8x8_t __ret;
   29877   __ret = (poly8x8_t)(__p0);
   29878   return __ret;
   29879 }
   29880 #endif
   29881 
   29882 #ifdef __LITTLE_ENDIAN__
   29883 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
   29884   poly8x8_t __ret;
   29885   __ret = (poly8x8_t)(__p0);
   29886   return __ret;
   29887 }
   29888 #else
   29889 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
   29890   poly8x8_t __ret;
   29891   __ret = (poly8x8_t)(__p0);
   29892   return __ret;
   29893 }
   29894 #endif
   29895 
   29896 #ifdef __LITTLE_ENDIAN__
   29897 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
   29898   poly8x8_t __ret;
   29899   __ret = (poly8x8_t)(__p0);
   29900   return __ret;
   29901 }
   29902 #else
   29903 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
   29904   poly8x8_t __ret;
   29905   __ret = (poly8x8_t)(__p0);
   29906   return __ret;
   29907 }
   29908 #endif
   29909 
   29910 #ifdef __LITTLE_ENDIAN__
   29911 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
   29912   poly8x8_t __ret;
   29913   __ret = (poly8x8_t)(__p0);
   29914   return __ret;
   29915 }
   29916 #else
   29917 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
   29918   poly8x8_t __ret;
   29919   __ret = (poly8x8_t)(__p0);
   29920   return __ret;
   29921 }
   29922 #endif
   29923 
   29924 #ifdef __LITTLE_ENDIAN__
   29925 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
   29926   poly8x8_t __ret;
   29927   __ret = (poly8x8_t)(__p0);
   29928   return __ret;
   29929 }
   29930 #else
   29931 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
   29932   poly8x8_t __ret;
   29933   __ret = (poly8x8_t)(__p0);
   29934   return __ret;
   29935 }
   29936 #endif
   29937 
   29938 #ifdef __LITTLE_ENDIAN__
   29939 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
   29940   poly8x8_t __ret;
   29941   __ret = (poly8x8_t)(__p0);
   29942   return __ret;
   29943 }
   29944 #else
   29945 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
   29946   poly8x8_t __ret;
   29947   __ret = (poly8x8_t)(__p0);
   29948   return __ret;
   29949 }
   29950 #endif
   29951 
   29952 #ifdef __LITTLE_ENDIAN__
   29953 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
   29954   poly8x8_t __ret;
   29955   __ret = (poly8x8_t)(__p0);
   29956   return __ret;
   29957 }
   29958 #else
   29959 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
   29960   poly8x8_t __ret;
   29961   __ret = (poly8x8_t)(__p0);
   29962   return __ret;
   29963 }
   29964 #endif
   29965 
   29966 #ifdef __LITTLE_ENDIAN__
   29967 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
   29968   poly8x8_t __ret;
   29969   __ret = (poly8x8_t)(__p0);
   29970   return __ret;
   29971 }
   29972 #else
   29973 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
   29974   poly8x8_t __ret;
   29975   __ret = (poly8x8_t)(__p0);
   29976   return __ret;
   29977 }
   29978 #endif
   29979 
   29980 #ifdef __LITTLE_ENDIAN__
   29981 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
   29982   poly8x8_t __ret;
   29983   __ret = (poly8x8_t)(__p0);
   29984   return __ret;
   29985 }
   29986 #else
   29987 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
   29988   poly8x8_t __ret;
   29989   __ret = (poly8x8_t)(__p0);
   29990   return __ret;
   29991 }
   29992 #endif
   29993 
   29994 #ifdef __LITTLE_ENDIAN__
   29995 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
   29996   poly16x4_t __ret;
   29997   __ret = (poly16x4_t)(__p0);
   29998   return __ret;
   29999 }
   30000 #else
   30001 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
   30002   poly16x4_t __ret;
   30003   __ret = (poly16x4_t)(__p0);
   30004   return __ret;
   30005 }
   30006 #endif
   30007 
   30008 #ifdef __LITTLE_ENDIAN__
   30009 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
   30010   poly16x4_t __ret;
   30011   __ret = (poly16x4_t)(__p0);
   30012   return __ret;
   30013 }
   30014 #else
   30015 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
   30016   poly16x4_t __ret;
   30017   __ret = (poly16x4_t)(__p0);
   30018   return __ret;
   30019 }
   30020 #endif
   30021 
   30022 #ifdef __LITTLE_ENDIAN__
   30023 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
   30024   poly16x4_t __ret;
   30025   __ret = (poly16x4_t)(__p0);
   30026   return __ret;
   30027 }
   30028 #else
   30029 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
   30030   poly16x4_t __ret;
   30031   __ret = (poly16x4_t)(__p0);
   30032   return __ret;
   30033 }
   30034 #endif
   30035 
   30036 #ifdef __LITTLE_ENDIAN__
   30037 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
   30038   poly16x4_t __ret;
   30039   __ret = (poly16x4_t)(__p0);
   30040   return __ret;
   30041 }
   30042 #else
   30043 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
   30044   poly16x4_t __ret;
   30045   __ret = (poly16x4_t)(__p0);
   30046   return __ret;
   30047 }
   30048 #endif
   30049 
   30050 #ifdef __LITTLE_ENDIAN__
   30051 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
   30052   poly16x4_t __ret;
   30053   __ret = (poly16x4_t)(__p0);
   30054   return __ret;
   30055 }
   30056 #else
   30057 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
   30058   poly16x4_t __ret;
   30059   __ret = (poly16x4_t)(__p0);
   30060   return __ret;
   30061 }
   30062 #endif
   30063 
   30064 #ifdef __LITTLE_ENDIAN__
   30065 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
   30066   poly16x4_t __ret;
   30067   __ret = (poly16x4_t)(__p0);
   30068   return __ret;
   30069 }
   30070 #else
   30071 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
   30072   poly16x4_t __ret;
   30073   __ret = (poly16x4_t)(__p0);
   30074   return __ret;
   30075 }
   30076 #endif
   30077 
   30078 #ifdef __LITTLE_ENDIAN__
   30079 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
   30080   poly16x4_t __ret;
   30081   __ret = (poly16x4_t)(__p0);
   30082   return __ret;
   30083 }
   30084 #else
   30085 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
   30086   poly16x4_t __ret;
   30087   __ret = (poly16x4_t)(__p0);
   30088   return __ret;
   30089 }
   30090 #endif
   30091 
   30092 #ifdef __LITTLE_ENDIAN__
   30093 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
   30094   poly16x4_t __ret;
   30095   __ret = (poly16x4_t)(__p0);
   30096   return __ret;
   30097 }
   30098 #else
   30099 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
   30100   poly16x4_t __ret;
   30101   __ret = (poly16x4_t)(__p0);
   30102   return __ret;
   30103 }
   30104 #endif
   30105 
   30106 #ifdef __LITTLE_ENDIAN__
   30107 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
   30108   poly16x4_t __ret;
   30109   __ret = (poly16x4_t)(__p0);
   30110   return __ret;
   30111 }
   30112 #else
   30113 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
   30114   poly16x4_t __ret;
   30115   __ret = (poly16x4_t)(__p0);
   30116   return __ret;
   30117 }
   30118 #endif
   30119 
   30120 #ifdef __LITTLE_ENDIAN__
   30121 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
   30122   poly16x4_t __ret;
   30123   __ret = (poly16x4_t)(__p0);
   30124   return __ret;
   30125 }
   30126 #else
   30127 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
   30128   poly16x4_t __ret;
   30129   __ret = (poly16x4_t)(__p0);
   30130   return __ret;
   30131 }
   30132 #endif
   30133 
   30134 #ifdef __LITTLE_ENDIAN__
   30135 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
   30136   poly16x4_t __ret;
   30137   __ret = (poly16x4_t)(__p0);
   30138   return __ret;
   30139 }
   30140 #else
   30141 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
   30142   poly16x4_t __ret;
   30143   __ret = (poly16x4_t)(__p0);
   30144   return __ret;
   30145 }
   30146 #endif
   30147 
   30148 #ifdef __LITTLE_ENDIAN__
   30149 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
   30150   poly8x16_t __ret;
   30151   __ret = (poly8x16_t)(__p0);
   30152   return __ret;
   30153 }
   30154 #else
   30155 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
   30156   poly8x16_t __ret;
   30157   __ret = (poly8x16_t)(__p0);
   30158   return __ret;
   30159 }
   30160 #endif
   30161 
   30162 #ifdef __LITTLE_ENDIAN__
   30163 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
   30164   poly8x16_t __ret;
   30165   __ret = (poly8x16_t)(__p0);
   30166   return __ret;
   30167 }
   30168 #else
   30169 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
   30170   poly8x16_t __ret;
   30171   __ret = (poly8x16_t)(__p0);
   30172   return __ret;
   30173 }
   30174 #endif
   30175 
   30176 #ifdef __LITTLE_ENDIAN__
   30177 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
   30178   poly8x16_t __ret;
   30179   __ret = (poly8x16_t)(__p0);
   30180   return __ret;
   30181 }
   30182 #else
   30183 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
   30184   poly8x16_t __ret;
   30185   __ret = (poly8x16_t)(__p0);
   30186   return __ret;
   30187 }
   30188 #endif
   30189 
   30190 #ifdef __LITTLE_ENDIAN__
   30191 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
   30192   poly8x16_t __ret;
   30193   __ret = (poly8x16_t)(__p0);
   30194   return __ret;
   30195 }
   30196 #else
   30197 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
   30198   poly8x16_t __ret;
   30199   __ret = (poly8x16_t)(__p0);
   30200   return __ret;
   30201 }
   30202 #endif
   30203 
   30204 #ifdef __LITTLE_ENDIAN__
   30205 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
   30206   poly8x16_t __ret;
   30207   __ret = (poly8x16_t)(__p0);
   30208   return __ret;
   30209 }
   30210 #else
   30211 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
   30212   poly8x16_t __ret;
   30213   __ret = (poly8x16_t)(__p0);
   30214   return __ret;
   30215 }
   30216 #endif
   30217 
   30218 #ifdef __LITTLE_ENDIAN__
   30219 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
   30220   poly8x16_t __ret;
   30221   __ret = (poly8x16_t)(__p0);
   30222   return __ret;
   30223 }
   30224 #else
   30225 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
   30226   poly8x16_t __ret;
   30227   __ret = (poly8x16_t)(__p0);
   30228   return __ret;
   30229 }
   30230 #endif
   30231 
   30232 #ifdef __LITTLE_ENDIAN__
   30233 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
   30234   poly8x16_t __ret;
   30235   __ret = (poly8x16_t)(__p0);
   30236   return __ret;
   30237 }
   30238 #else
   30239 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
   30240   poly8x16_t __ret;
   30241   __ret = (poly8x16_t)(__p0);
   30242   return __ret;
   30243 }
   30244 #endif
   30245 
   30246 #ifdef __LITTLE_ENDIAN__
   30247 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
   30248   poly8x16_t __ret;
   30249   __ret = (poly8x16_t)(__p0);
   30250   return __ret;
   30251 }
   30252 #else
   30253 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
   30254   poly8x16_t __ret;
   30255   __ret = (poly8x16_t)(__p0);
   30256   return __ret;
   30257 }
   30258 #endif
   30259 
   30260 #ifdef __LITTLE_ENDIAN__
   30261 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
   30262   poly8x16_t __ret;
   30263   __ret = (poly8x16_t)(__p0);
   30264   return __ret;
   30265 }
   30266 #else
   30267 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
   30268   poly8x16_t __ret;
   30269   __ret = (poly8x16_t)(__p0);
   30270   return __ret;
   30271 }
   30272 #endif
   30273 
   30274 #ifdef __LITTLE_ENDIAN__
   30275 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
   30276   poly8x16_t __ret;
   30277   __ret = (poly8x16_t)(__p0);
   30278   return __ret;
   30279 }
   30280 #else
   30281 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
   30282   poly8x16_t __ret;
   30283   __ret = (poly8x16_t)(__p0);
   30284   return __ret;
   30285 }
   30286 #endif
   30287 
   30288 #ifdef __LITTLE_ENDIAN__
   30289 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
   30290   poly8x16_t __ret;
   30291   __ret = (poly8x16_t)(__p0);
   30292   return __ret;
   30293 }
   30294 #else
   30295 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
   30296   poly8x16_t __ret;
   30297   __ret = (poly8x16_t)(__p0);
   30298   return __ret;
   30299 }
   30300 #endif
   30301 
   30302 #ifdef __LITTLE_ENDIAN__
   30303 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
   30304   poly16x8_t __ret;
   30305   __ret = (poly16x8_t)(__p0);
   30306   return __ret;
   30307 }
   30308 #else
   30309 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
   30310   poly16x8_t __ret;
   30311   __ret = (poly16x8_t)(__p0);
   30312   return __ret;
   30313 }
   30314 #endif
   30315 
   30316 #ifdef __LITTLE_ENDIAN__
   30317 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
   30318   poly16x8_t __ret;
   30319   __ret = (poly16x8_t)(__p0);
   30320   return __ret;
   30321 }
   30322 #else
   30323 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
   30324   poly16x8_t __ret;
   30325   __ret = (poly16x8_t)(__p0);
   30326   return __ret;
   30327 }
   30328 #endif
   30329 
   30330 #ifdef __LITTLE_ENDIAN__
   30331 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
   30332   poly16x8_t __ret;
   30333   __ret = (poly16x8_t)(__p0);
   30334   return __ret;
   30335 }
   30336 #else
   30337 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
   30338   poly16x8_t __ret;
   30339   __ret = (poly16x8_t)(__p0);
   30340   return __ret;
   30341 }
   30342 #endif
   30343 
   30344 #ifdef __LITTLE_ENDIAN__
   30345 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
   30346   poly16x8_t __ret;
   30347   __ret = (poly16x8_t)(__p0);
   30348   return __ret;
   30349 }
   30350 #else
   30351 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
   30352   poly16x8_t __ret;
   30353   __ret = (poly16x8_t)(__p0);
   30354   return __ret;
   30355 }
   30356 #endif
   30357 
   30358 #ifdef __LITTLE_ENDIAN__
   30359 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
   30360   poly16x8_t __ret;
   30361   __ret = (poly16x8_t)(__p0);
   30362   return __ret;
   30363 }
   30364 #else
   30365 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
   30366   poly16x8_t __ret;
   30367   __ret = (poly16x8_t)(__p0);
   30368   return __ret;
   30369 }
   30370 #endif
   30371 
   30372 #ifdef __LITTLE_ENDIAN__
   30373 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
   30374   poly16x8_t __ret;
   30375   __ret = (poly16x8_t)(__p0);
   30376   return __ret;
   30377 }
   30378 #else
   30379 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
   30380   poly16x8_t __ret;
   30381   __ret = (poly16x8_t)(__p0);
   30382   return __ret;
   30383 }
   30384 #endif
   30385 
   30386 #ifdef __LITTLE_ENDIAN__
   30387 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
   30388   poly16x8_t __ret;
   30389   __ret = (poly16x8_t)(__p0);
   30390   return __ret;
   30391 }
   30392 #else
   30393 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
   30394   poly16x8_t __ret;
   30395   __ret = (poly16x8_t)(__p0);
   30396   return __ret;
   30397 }
   30398 #endif
   30399 
   30400 #ifdef __LITTLE_ENDIAN__
   30401 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
   30402   poly16x8_t __ret;
   30403   __ret = (poly16x8_t)(__p0);
   30404   return __ret;
   30405 }
   30406 #else
   30407 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
   30408   poly16x8_t __ret;
   30409   __ret = (poly16x8_t)(__p0);
   30410   return __ret;
   30411 }
   30412 #endif
   30413 
   30414 #ifdef __LITTLE_ENDIAN__
   30415 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
   30416   poly16x8_t __ret;
   30417   __ret = (poly16x8_t)(__p0);
   30418   return __ret;
   30419 }
   30420 #else
   30421 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
   30422   poly16x8_t __ret;
   30423   __ret = (poly16x8_t)(__p0);
   30424   return __ret;
   30425 }
   30426 #endif
   30427 
   30428 #ifdef __LITTLE_ENDIAN__
   30429 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
   30430   poly16x8_t __ret;
   30431   __ret = (poly16x8_t)(__p0);
   30432   return __ret;
   30433 }
   30434 #else
   30435 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
   30436   poly16x8_t __ret;
   30437   __ret = (poly16x8_t)(__p0);
   30438   return __ret;
   30439 }
   30440 #endif
   30441 
   30442 #ifdef __LITTLE_ENDIAN__
   30443 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
   30444   poly16x8_t __ret;
   30445   __ret = (poly16x8_t)(__p0);
   30446   return __ret;
   30447 }
   30448 #else
   30449 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
   30450   poly16x8_t __ret;
   30451   __ret = (poly16x8_t)(__p0);
   30452   return __ret;
   30453 }
   30454 #endif
   30455 
   30456 #ifdef __LITTLE_ENDIAN__
   30457 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
   30458   uint8x16_t __ret;
   30459   __ret = (uint8x16_t)(__p0);
   30460   return __ret;
   30461 }
   30462 #else
   30463 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
   30464   uint8x16_t __ret;
   30465   __ret = (uint8x16_t)(__p0);
   30466   return __ret;
   30467 }
   30468 #endif
   30469 
   30470 #ifdef __LITTLE_ENDIAN__
   30471 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
   30472   uint8x16_t __ret;
   30473   __ret = (uint8x16_t)(__p0);
   30474   return __ret;
   30475 }
   30476 #else
   30477 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
   30478   uint8x16_t __ret;
   30479   __ret = (uint8x16_t)(__p0);
   30480   return __ret;
   30481 }
   30482 #endif
   30483 
   30484 #ifdef __LITTLE_ENDIAN__
   30485 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
   30486   uint8x16_t __ret;
   30487   __ret = (uint8x16_t)(__p0);
   30488   return __ret;
   30489 }
   30490 #else
   30491 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
   30492   uint8x16_t __ret;
   30493   __ret = (uint8x16_t)(__p0);
   30494   return __ret;
   30495 }
   30496 #endif
   30497 
   30498 #ifdef __LITTLE_ENDIAN__
   30499 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
   30500   uint8x16_t __ret;
   30501   __ret = (uint8x16_t)(__p0);
   30502   return __ret;
   30503 }
   30504 #else
   30505 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
   30506   uint8x16_t __ret;
   30507   __ret = (uint8x16_t)(__p0);
   30508   return __ret;
   30509 }
   30510 #endif
   30511 
   30512 #ifdef __LITTLE_ENDIAN__
   30513 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
   30514   uint8x16_t __ret;
   30515   __ret = (uint8x16_t)(__p0);
   30516   return __ret;
   30517 }
   30518 #else
   30519 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
   30520   uint8x16_t __ret;
   30521   __ret = (uint8x16_t)(__p0);
   30522   return __ret;
   30523 }
   30524 #endif
   30525 
   30526 #ifdef __LITTLE_ENDIAN__
   30527 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
   30528   uint8x16_t __ret;
   30529   __ret = (uint8x16_t)(__p0);
   30530   return __ret;
   30531 }
   30532 #else
   30533 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
   30534   uint8x16_t __ret;
   30535   __ret = (uint8x16_t)(__p0);
   30536   return __ret;
   30537 }
   30538 #endif
   30539 
   30540 #ifdef __LITTLE_ENDIAN__
   30541 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
   30542   uint8x16_t __ret;
   30543   __ret = (uint8x16_t)(__p0);
   30544   return __ret;
   30545 }
   30546 #else
   30547 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
   30548   uint8x16_t __ret;
   30549   __ret = (uint8x16_t)(__p0);
   30550   return __ret;
   30551 }
   30552 #endif
   30553 
   30554 #ifdef __LITTLE_ENDIAN__
   30555 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
   30556   uint8x16_t __ret;
   30557   __ret = (uint8x16_t)(__p0);
   30558   return __ret;
   30559 }
   30560 #else
   30561 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
   30562   uint8x16_t __ret;
   30563   __ret = (uint8x16_t)(__p0);
   30564   return __ret;
   30565 }
   30566 #endif
   30567 
   30568 #ifdef __LITTLE_ENDIAN__
   30569 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
   30570   uint8x16_t __ret;
   30571   __ret = (uint8x16_t)(__p0);
   30572   return __ret;
   30573 }
   30574 #else
   30575 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
   30576   uint8x16_t __ret;
   30577   __ret = (uint8x16_t)(__p0);
   30578   return __ret;
   30579 }
   30580 #endif
   30581 
   30582 #ifdef __LITTLE_ENDIAN__
   30583 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
   30584   uint8x16_t __ret;
   30585   __ret = (uint8x16_t)(__p0);
   30586   return __ret;
   30587 }
   30588 #else
   30589 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
   30590   uint8x16_t __ret;
   30591   __ret = (uint8x16_t)(__p0);
   30592   return __ret;
   30593 }
   30594 #endif
   30595 
   30596 #ifdef __LITTLE_ENDIAN__
   30597 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
   30598   uint8x16_t __ret;
   30599   __ret = (uint8x16_t)(__p0);
   30600   return __ret;
   30601 }
   30602 #else
   30603 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
   30604   uint8x16_t __ret;
   30605   __ret = (uint8x16_t)(__p0);
   30606   return __ret;
   30607 }
   30608 #endif
   30609 
   30610 #ifdef __LITTLE_ENDIAN__
   30611 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
   30612   uint32x4_t __ret;
   30613   __ret = (uint32x4_t)(__p0);
   30614   return __ret;
   30615 }
   30616 #else
   30617 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
   30618   uint32x4_t __ret;
   30619   __ret = (uint32x4_t)(__p0);
   30620   return __ret;
   30621 }
   30622 #endif
   30623 
   30624 #ifdef __LITTLE_ENDIAN__
   30625 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
   30626   uint32x4_t __ret;
   30627   __ret = (uint32x4_t)(__p0);
   30628   return __ret;
   30629 }
   30630 #else
   30631 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
   30632   uint32x4_t __ret;
   30633   __ret = (uint32x4_t)(__p0);
   30634   return __ret;
   30635 }
   30636 #endif
   30637 
   30638 #ifdef __LITTLE_ENDIAN__
   30639 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
   30640   uint32x4_t __ret;
   30641   __ret = (uint32x4_t)(__p0);
   30642   return __ret;
   30643 }
   30644 #else
   30645 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
   30646   uint32x4_t __ret;
   30647   __ret = (uint32x4_t)(__p0);
   30648   return __ret;
   30649 }
   30650 #endif
   30651 
   30652 #ifdef __LITTLE_ENDIAN__
   30653 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
   30654   uint32x4_t __ret;
   30655   __ret = (uint32x4_t)(__p0);
   30656   return __ret;
   30657 }
   30658 #else
   30659 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
   30660   uint32x4_t __ret;
   30661   __ret = (uint32x4_t)(__p0);
   30662   return __ret;
   30663 }
   30664 #endif
   30665 
   30666 #ifdef __LITTLE_ENDIAN__
   30667 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
   30668   uint32x4_t __ret;
   30669   __ret = (uint32x4_t)(__p0);
   30670   return __ret;
   30671 }
   30672 #else
   30673 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
   30674   uint32x4_t __ret;
   30675   __ret = (uint32x4_t)(__p0);
   30676   return __ret;
   30677 }
   30678 #endif
   30679 
   30680 #ifdef __LITTLE_ENDIAN__
   30681 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
   30682   uint32x4_t __ret;
   30683   __ret = (uint32x4_t)(__p0);
   30684   return __ret;
   30685 }
   30686 #else
   30687 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
   30688   uint32x4_t __ret;
   30689   __ret = (uint32x4_t)(__p0);
   30690   return __ret;
   30691 }
   30692 #endif
   30693 
   30694 #ifdef __LITTLE_ENDIAN__
   30695 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
   30696   uint32x4_t __ret;
   30697   __ret = (uint32x4_t)(__p0);
   30698   return __ret;
   30699 }
   30700 #else
   30701 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
   30702   uint32x4_t __ret;
   30703   __ret = (uint32x4_t)(__p0);
   30704   return __ret;
   30705 }
   30706 #endif
   30707 
   30708 #ifdef __LITTLE_ENDIAN__
   30709 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
   30710   uint32x4_t __ret;
   30711   __ret = (uint32x4_t)(__p0);
   30712   return __ret;
   30713 }
   30714 #else
   30715 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
   30716   uint32x4_t __ret;
   30717   __ret = (uint32x4_t)(__p0);
   30718   return __ret;
   30719 }
   30720 #endif
   30721 
   30722 #ifdef __LITTLE_ENDIAN__
   30723 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
   30724   uint32x4_t __ret;
   30725   __ret = (uint32x4_t)(__p0);
   30726   return __ret;
   30727 }
   30728 #else
   30729 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
   30730   uint32x4_t __ret;
   30731   __ret = (uint32x4_t)(__p0);
   30732   return __ret;
   30733 }
   30734 #endif
   30735 
   30736 #ifdef __LITTLE_ENDIAN__
   30737 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
   30738   uint32x4_t __ret;
   30739   __ret = (uint32x4_t)(__p0);
   30740   return __ret;
   30741 }
   30742 #else
   30743 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
   30744   uint32x4_t __ret;
   30745   __ret = (uint32x4_t)(__p0);
   30746   return __ret;
   30747 }
   30748 #endif
   30749 
   30750 #ifdef __LITTLE_ENDIAN__
   30751 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
   30752   uint32x4_t __ret;
   30753   __ret = (uint32x4_t)(__p0);
   30754   return __ret;
   30755 }
   30756 #else
   30757 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
   30758   uint32x4_t __ret;
   30759   __ret = (uint32x4_t)(__p0);
   30760   return __ret;
   30761 }
   30762 #endif
   30763 
   30764 #ifdef __LITTLE_ENDIAN__
   30765 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
   30766   uint64x2_t __ret;
   30767   __ret = (uint64x2_t)(__p0);
   30768   return __ret;
   30769 }
   30770 #else
   30771 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
   30772   uint64x2_t __ret;
   30773   __ret = (uint64x2_t)(__p0);
   30774   return __ret;
   30775 }
   30776 #endif
   30777 
   30778 #ifdef __LITTLE_ENDIAN__
   30779 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
   30780   uint64x2_t __ret;
   30781   __ret = (uint64x2_t)(__p0);
   30782   return __ret;
   30783 }
   30784 #else
   30785 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
   30786   uint64x2_t __ret;
   30787   __ret = (uint64x2_t)(__p0);
   30788   return __ret;
   30789 }
   30790 #endif
   30791 
   30792 #ifdef __LITTLE_ENDIAN__
   30793 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
   30794   uint64x2_t __ret;
   30795   __ret = (uint64x2_t)(__p0);
   30796   return __ret;
   30797 }
   30798 #else
   30799 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
   30800   uint64x2_t __ret;
   30801   __ret = (uint64x2_t)(__p0);
   30802   return __ret;
   30803 }
   30804 #endif
   30805 
   30806 #ifdef __LITTLE_ENDIAN__
   30807 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
   30808   uint64x2_t __ret;
   30809   __ret = (uint64x2_t)(__p0);
   30810   return __ret;
   30811 }
   30812 #else
   30813 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
   30814   uint64x2_t __ret;
   30815   __ret = (uint64x2_t)(__p0);
   30816   return __ret;
   30817 }
   30818 #endif
   30819 
   30820 #ifdef __LITTLE_ENDIAN__
   30821 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
   30822   uint64x2_t __ret;
   30823   __ret = (uint64x2_t)(__p0);
   30824   return __ret;
   30825 }
   30826 #else
   30827 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
   30828   uint64x2_t __ret;
   30829   __ret = (uint64x2_t)(__p0);
   30830   return __ret;
   30831 }
   30832 #endif
   30833 
   30834 #ifdef __LITTLE_ENDIAN__
   30835 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
   30836   uint64x2_t __ret;
   30837   __ret = (uint64x2_t)(__p0);
   30838   return __ret;
   30839 }
   30840 #else
   30841 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
   30842   uint64x2_t __ret;
   30843   __ret = (uint64x2_t)(__p0);
   30844   return __ret;
   30845 }
   30846 #endif
   30847 
   30848 #ifdef __LITTLE_ENDIAN__
   30849 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
   30850   uint64x2_t __ret;
   30851   __ret = (uint64x2_t)(__p0);
   30852   return __ret;
   30853 }
   30854 #else
   30855 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
   30856   uint64x2_t __ret;
   30857   __ret = (uint64x2_t)(__p0);
   30858   return __ret;
   30859 }
   30860 #endif
   30861 
   30862 #ifdef __LITTLE_ENDIAN__
   30863 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
   30864   uint64x2_t __ret;
   30865   __ret = (uint64x2_t)(__p0);
   30866   return __ret;
   30867 }
   30868 #else
   30869 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
   30870   uint64x2_t __ret;
   30871   __ret = (uint64x2_t)(__p0);
   30872   return __ret;
   30873 }
   30874 #endif
   30875 
   30876 #ifdef __LITTLE_ENDIAN__
   30877 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
   30878   uint64x2_t __ret;
   30879   __ret = (uint64x2_t)(__p0);
   30880   return __ret;
   30881 }
   30882 #else
   30883 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
   30884   uint64x2_t __ret;
   30885   __ret = (uint64x2_t)(__p0);
   30886   return __ret;
   30887 }
   30888 #endif
   30889 
   30890 #ifdef __LITTLE_ENDIAN__
   30891 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
   30892   uint64x2_t __ret;
   30893   __ret = (uint64x2_t)(__p0);
   30894   return __ret;
   30895 }
   30896 #else
   30897 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
   30898   uint64x2_t __ret;
   30899   __ret = (uint64x2_t)(__p0);
   30900   return __ret;
   30901 }
   30902 #endif
   30903 
   30904 #ifdef __LITTLE_ENDIAN__
   30905 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
   30906   uint64x2_t __ret;
   30907   __ret = (uint64x2_t)(__p0);
   30908   return __ret;
   30909 }
   30910 #else
   30911 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
   30912   uint64x2_t __ret;
   30913   __ret = (uint64x2_t)(__p0);
   30914   return __ret;
   30915 }
   30916 #endif
   30917 
   30918 #ifdef __LITTLE_ENDIAN__
   30919 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
   30920   uint16x8_t __ret;
   30921   __ret = (uint16x8_t)(__p0);
   30922   return __ret;
   30923 }
   30924 #else
   30925 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
   30926   uint16x8_t __ret;
   30927   __ret = (uint16x8_t)(__p0);
   30928   return __ret;
   30929 }
   30930 #endif
   30931 
   30932 #ifdef __LITTLE_ENDIAN__
   30933 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
   30934   uint16x8_t __ret;
   30935   __ret = (uint16x8_t)(__p0);
   30936   return __ret;
   30937 }
   30938 #else
   30939 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
   30940   uint16x8_t __ret;
   30941   __ret = (uint16x8_t)(__p0);
   30942   return __ret;
   30943 }
   30944 #endif
   30945 
   30946 #ifdef __LITTLE_ENDIAN__
   30947 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
   30948   uint16x8_t __ret;
   30949   __ret = (uint16x8_t)(__p0);
   30950   return __ret;
   30951 }
   30952 #else
   30953 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
   30954   uint16x8_t __ret;
   30955   __ret = (uint16x8_t)(__p0);
   30956   return __ret;
   30957 }
   30958 #endif
   30959 
   30960 #ifdef __LITTLE_ENDIAN__
   30961 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
   30962   uint16x8_t __ret;
   30963   __ret = (uint16x8_t)(__p0);
   30964   return __ret;
   30965 }
   30966 #else
   30967 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
   30968   uint16x8_t __ret;
   30969   __ret = (uint16x8_t)(__p0);
   30970   return __ret;
   30971 }
   30972 #endif
   30973 
   30974 #ifdef __LITTLE_ENDIAN__
   30975 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
   30976   uint16x8_t __ret;
   30977   __ret = (uint16x8_t)(__p0);
   30978   return __ret;
   30979 }
   30980 #else
   30981 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
   30982   uint16x8_t __ret;
   30983   __ret = (uint16x8_t)(__p0);
   30984   return __ret;
   30985 }
   30986 #endif
   30987 
   30988 #ifdef __LITTLE_ENDIAN__
   30989 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
   30990   uint16x8_t __ret;
   30991   __ret = (uint16x8_t)(__p0);
   30992   return __ret;
   30993 }
   30994 #else
   30995 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
   30996   uint16x8_t __ret;
   30997   __ret = (uint16x8_t)(__p0);
   30998   return __ret;
   30999 }
   31000 #endif
   31001 
   31002 #ifdef __LITTLE_ENDIAN__
   31003 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
   31004   uint16x8_t __ret;
   31005   __ret = (uint16x8_t)(__p0);
   31006   return __ret;
   31007 }
   31008 #else
   31009 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
   31010   uint16x8_t __ret;
   31011   __ret = (uint16x8_t)(__p0);
   31012   return __ret;
   31013 }
   31014 #endif
   31015 
   31016 #ifdef __LITTLE_ENDIAN__
   31017 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
   31018   uint16x8_t __ret;
   31019   __ret = (uint16x8_t)(__p0);
   31020   return __ret;
   31021 }
   31022 #else
   31023 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
   31024   uint16x8_t __ret;
   31025   __ret = (uint16x8_t)(__p0);
   31026   return __ret;
   31027 }
   31028 #endif
   31029 
   31030 #ifdef __LITTLE_ENDIAN__
   31031 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
   31032   uint16x8_t __ret;
   31033   __ret = (uint16x8_t)(__p0);
   31034   return __ret;
   31035 }
   31036 #else
   31037 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
   31038   uint16x8_t __ret;
   31039   __ret = (uint16x8_t)(__p0);
   31040   return __ret;
   31041 }
   31042 #endif
   31043 
   31044 #ifdef __LITTLE_ENDIAN__
   31045 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
   31046   uint16x8_t __ret;
   31047   __ret = (uint16x8_t)(__p0);
   31048   return __ret;
   31049 }
   31050 #else
   31051 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
   31052   uint16x8_t __ret;
   31053   __ret = (uint16x8_t)(__p0);
   31054   return __ret;
   31055 }
   31056 #endif
   31057 
   31058 #ifdef __LITTLE_ENDIAN__
   31059 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
   31060   uint16x8_t __ret;
   31061   __ret = (uint16x8_t)(__p0);
   31062   return __ret;
   31063 }
   31064 #else
   31065 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
   31066   uint16x8_t __ret;
   31067   __ret = (uint16x8_t)(__p0);
   31068   return __ret;
   31069 }
   31070 #endif
   31071 
   31072 #ifdef __LITTLE_ENDIAN__
   31073 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
   31074   int8x16_t __ret;
   31075   __ret = (int8x16_t)(__p0);
   31076   return __ret;
   31077 }
   31078 #else
   31079 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
   31080   int8x16_t __ret;
   31081   __ret = (int8x16_t)(__p0);
   31082   return __ret;
   31083 }
   31084 #endif
   31085 
   31086 #ifdef __LITTLE_ENDIAN__
   31087 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
   31088   int8x16_t __ret;
   31089   __ret = (int8x16_t)(__p0);
   31090   return __ret;
   31091 }
   31092 #else
   31093 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
   31094   int8x16_t __ret;
   31095   __ret = (int8x16_t)(__p0);
   31096   return __ret;
   31097 }
   31098 #endif
   31099 
   31100 #ifdef __LITTLE_ENDIAN__
   31101 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
   31102   int8x16_t __ret;
   31103   __ret = (int8x16_t)(__p0);
   31104   return __ret;
   31105 }
   31106 #else
   31107 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
   31108   int8x16_t __ret;
   31109   __ret = (int8x16_t)(__p0);
   31110   return __ret;
   31111 }
   31112 #endif
   31113 
   31114 #ifdef __LITTLE_ENDIAN__
   31115 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
   31116   int8x16_t __ret;
   31117   __ret = (int8x16_t)(__p0);
   31118   return __ret;
   31119 }
   31120 #else
   31121 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
   31122   int8x16_t __ret;
   31123   __ret = (int8x16_t)(__p0);
   31124   return __ret;
   31125 }
   31126 #endif
   31127 
   31128 #ifdef __LITTLE_ENDIAN__
   31129 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
   31130   int8x16_t __ret;
   31131   __ret = (int8x16_t)(__p0);
   31132   return __ret;
   31133 }
   31134 #else
   31135 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
   31136   int8x16_t __ret;
   31137   __ret = (int8x16_t)(__p0);
   31138   return __ret;
   31139 }
   31140 #endif
   31141 
   31142 #ifdef __LITTLE_ENDIAN__
   31143 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
   31144   int8x16_t __ret;
   31145   __ret = (int8x16_t)(__p0);
   31146   return __ret;
   31147 }
   31148 #else
   31149 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
   31150   int8x16_t __ret;
   31151   __ret = (int8x16_t)(__p0);
   31152   return __ret;
   31153 }
   31154 #endif
   31155 
   31156 #ifdef __LITTLE_ENDIAN__
   31157 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
   31158   int8x16_t __ret;
   31159   __ret = (int8x16_t)(__p0);
   31160   return __ret;
   31161 }
   31162 #else
   31163 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
   31164   int8x16_t __ret;
   31165   __ret = (int8x16_t)(__p0);
   31166   return __ret;
   31167 }
   31168 #endif
   31169 
   31170 #ifdef __LITTLE_ENDIAN__
   31171 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
   31172   int8x16_t __ret;
   31173   __ret = (int8x16_t)(__p0);
   31174   return __ret;
   31175 }
   31176 #else
   31177 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
   31178   int8x16_t __ret;
   31179   __ret = (int8x16_t)(__p0);
   31180   return __ret;
   31181 }
   31182 #endif
   31183 
   31184 #ifdef __LITTLE_ENDIAN__
   31185 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
   31186   int8x16_t __ret;
   31187   __ret = (int8x16_t)(__p0);
   31188   return __ret;
   31189 }
   31190 #else
   31191 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
   31192   int8x16_t __ret;
   31193   __ret = (int8x16_t)(__p0);
   31194   return __ret;
   31195 }
   31196 #endif
   31197 
   31198 #ifdef __LITTLE_ENDIAN__
   31199 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
   31200   int8x16_t __ret;
   31201   __ret = (int8x16_t)(__p0);
   31202   return __ret;
   31203 }
   31204 #else
   31205 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
   31206   int8x16_t __ret;
   31207   __ret = (int8x16_t)(__p0);
   31208   return __ret;
   31209 }
   31210 #endif
   31211 
   31212 #ifdef __LITTLE_ENDIAN__
   31213 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
   31214   int8x16_t __ret;
   31215   __ret = (int8x16_t)(__p0);
   31216   return __ret;
   31217 }
   31218 #else
   31219 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
   31220   int8x16_t __ret;
   31221   __ret = (int8x16_t)(__p0);
   31222   return __ret;
   31223 }
   31224 #endif
   31225 
   31226 #ifdef __LITTLE_ENDIAN__
   31227 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
   31228   float32x4_t __ret;
   31229   __ret = (float32x4_t)(__p0);
   31230   return __ret;
   31231 }
   31232 #else
   31233 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
   31234   float32x4_t __ret;
   31235   __ret = (float32x4_t)(__p0);
   31236   return __ret;
   31237 }
   31238 #endif
   31239 
   31240 #ifdef __LITTLE_ENDIAN__
   31241 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
   31242   float32x4_t __ret;
   31243   __ret = (float32x4_t)(__p0);
   31244   return __ret;
   31245 }
   31246 #else
   31247 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
   31248   float32x4_t __ret;
   31249   __ret = (float32x4_t)(__p0);
   31250   return __ret;
   31251 }
   31252 #endif
   31253 
   31254 #ifdef __LITTLE_ENDIAN__
   31255 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
   31256   float32x4_t __ret;
   31257   __ret = (float32x4_t)(__p0);
   31258   return __ret;
   31259 }
   31260 #else
   31261 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
   31262   float32x4_t __ret;
   31263   __ret = (float32x4_t)(__p0);
   31264   return __ret;
   31265 }
   31266 #endif
   31267 
   31268 #ifdef __LITTLE_ENDIAN__
   31269 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
   31270   float32x4_t __ret;
   31271   __ret = (float32x4_t)(__p0);
   31272   return __ret;
   31273 }
   31274 #else
   31275 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
   31276   float32x4_t __ret;
   31277   __ret = (float32x4_t)(__p0);
   31278   return __ret;
   31279 }
   31280 #endif
   31281 
   31282 #ifdef __LITTLE_ENDIAN__
   31283 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
   31284   float32x4_t __ret;
   31285   __ret = (float32x4_t)(__p0);
   31286   return __ret;
   31287 }
   31288 #else
   31289 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
   31290   float32x4_t __ret;
   31291   __ret = (float32x4_t)(__p0);
   31292   return __ret;
   31293 }
   31294 #endif
   31295 
   31296 #ifdef __LITTLE_ENDIAN__
   31297 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
   31298   float32x4_t __ret;
   31299   __ret = (float32x4_t)(__p0);
   31300   return __ret;
   31301 }
   31302 #else
   31303 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
   31304   float32x4_t __ret;
   31305   __ret = (float32x4_t)(__p0);
   31306   return __ret;
   31307 }
   31308 #endif
   31309 
   31310 #ifdef __LITTLE_ENDIAN__
   31311 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
   31312   float32x4_t __ret;
   31313   __ret = (float32x4_t)(__p0);
   31314   return __ret;
   31315 }
   31316 #else
   31317 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
   31318   float32x4_t __ret;
   31319   __ret = (float32x4_t)(__p0);
   31320   return __ret;
   31321 }
   31322 #endif
   31323 
   31324 #ifdef __LITTLE_ENDIAN__
   31325 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
   31326   float32x4_t __ret;
   31327   __ret = (float32x4_t)(__p0);
   31328   return __ret;
   31329 }
   31330 #else
   31331 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
   31332   float32x4_t __ret;
   31333   __ret = (float32x4_t)(__p0);
   31334   return __ret;
   31335 }
   31336 #endif
   31337 
   31338 #ifdef __LITTLE_ENDIAN__
   31339 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
   31340   float32x4_t __ret;
   31341   __ret = (float32x4_t)(__p0);
   31342   return __ret;
   31343 }
   31344 #else
   31345 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
   31346   float32x4_t __ret;
   31347   __ret = (float32x4_t)(__p0);
   31348   return __ret;
   31349 }
   31350 #endif
   31351 
   31352 #ifdef __LITTLE_ENDIAN__
   31353 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
   31354   float32x4_t __ret;
   31355   __ret = (float32x4_t)(__p0);
   31356   return __ret;
   31357 }
   31358 #else
   31359 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
   31360   float32x4_t __ret;
   31361   __ret = (float32x4_t)(__p0);
   31362   return __ret;
   31363 }
   31364 #endif
   31365 
   31366 #ifdef __LITTLE_ENDIAN__
   31367 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
   31368   float32x4_t __ret;
   31369   __ret = (float32x4_t)(__p0);
   31370   return __ret;
   31371 }
   31372 #else
   31373 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
   31374   float32x4_t __ret;
   31375   __ret = (float32x4_t)(__p0);
   31376   return __ret;
   31377 }
   31378 #endif
   31379 
   31380 #ifdef __LITTLE_ENDIAN__
   31381 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
   31382   float16x8_t __ret;
   31383   __ret = (float16x8_t)(__p0);
   31384   return __ret;
   31385 }
   31386 #else
   31387 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
   31388   float16x8_t __ret;
   31389   __ret = (float16x8_t)(__p0);
   31390   return __ret;
   31391 }
   31392 #endif
   31393 
   31394 #ifdef __LITTLE_ENDIAN__
   31395 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
   31396   float16x8_t __ret;
   31397   __ret = (float16x8_t)(__p0);
   31398   return __ret;
   31399 }
   31400 #else
   31401 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
   31402   float16x8_t __ret;
   31403   __ret = (float16x8_t)(__p0);
   31404   return __ret;
   31405 }
   31406 #endif
   31407 
   31408 #ifdef __LITTLE_ENDIAN__
   31409 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
   31410   float16x8_t __ret;
   31411   __ret = (float16x8_t)(__p0);
   31412   return __ret;
   31413 }
   31414 #else
   31415 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
   31416   float16x8_t __ret;
   31417   __ret = (float16x8_t)(__p0);
   31418   return __ret;
   31419 }
   31420 #endif
   31421 
   31422 #ifdef __LITTLE_ENDIAN__
   31423 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
   31424   float16x8_t __ret;
   31425   __ret = (float16x8_t)(__p0);
   31426   return __ret;
   31427 }
   31428 #else
   31429 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
   31430   float16x8_t __ret;
   31431   __ret = (float16x8_t)(__p0);
   31432   return __ret;
   31433 }
   31434 #endif
   31435 
   31436 #ifdef __LITTLE_ENDIAN__
   31437 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
   31438   float16x8_t __ret;
   31439   __ret = (float16x8_t)(__p0);
   31440   return __ret;
   31441 }
   31442 #else
   31443 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
   31444   float16x8_t __ret;
   31445   __ret = (float16x8_t)(__p0);
   31446   return __ret;
   31447 }
   31448 #endif
   31449 
   31450 #ifdef __LITTLE_ENDIAN__
   31451 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
   31452   float16x8_t __ret;
   31453   __ret = (float16x8_t)(__p0);
   31454   return __ret;
   31455 }
   31456 #else
   31457 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
   31458   float16x8_t __ret;
   31459   __ret = (float16x8_t)(__p0);
   31460   return __ret;
   31461 }
   31462 #endif
   31463 
   31464 #ifdef __LITTLE_ENDIAN__
   31465 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
   31466   float16x8_t __ret;
   31467   __ret = (float16x8_t)(__p0);
   31468   return __ret;
   31469 }
   31470 #else
   31471 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
   31472   float16x8_t __ret;
   31473   __ret = (float16x8_t)(__p0);
   31474   return __ret;
   31475 }
   31476 #endif
   31477 
   31478 #ifdef __LITTLE_ENDIAN__
   31479 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
   31480   float16x8_t __ret;
   31481   __ret = (float16x8_t)(__p0);
   31482   return __ret;
   31483 }
   31484 #else
   31485 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
   31486   float16x8_t __ret;
   31487   __ret = (float16x8_t)(__p0);
   31488   return __ret;
   31489 }
   31490 #endif
   31491 
   31492 #ifdef __LITTLE_ENDIAN__
   31493 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
   31494   float16x8_t __ret;
   31495   __ret = (float16x8_t)(__p0);
   31496   return __ret;
   31497 }
   31498 #else
   31499 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
   31500   float16x8_t __ret;
   31501   __ret = (float16x8_t)(__p0);
   31502   return __ret;
   31503 }
   31504 #endif
   31505 
   31506 #ifdef __LITTLE_ENDIAN__
   31507 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
   31508   float16x8_t __ret;
   31509   __ret = (float16x8_t)(__p0);
   31510   return __ret;
   31511 }
   31512 #else
   31513 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
   31514   float16x8_t __ret;
   31515   __ret = (float16x8_t)(__p0);
   31516   return __ret;
   31517 }
   31518 #endif
   31519 
   31520 #ifdef __LITTLE_ENDIAN__
   31521 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
   31522   float16x8_t __ret;
   31523   __ret = (float16x8_t)(__p0);
   31524   return __ret;
   31525 }
   31526 #else
   31527 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
   31528   float16x8_t __ret;
   31529   __ret = (float16x8_t)(__p0);
   31530   return __ret;
   31531 }
   31532 #endif
   31533 
   31534 #ifdef __LITTLE_ENDIAN__
   31535 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
   31536   int32x4_t __ret;
   31537   __ret = (int32x4_t)(__p0);
   31538   return __ret;
   31539 }
   31540 #else
   31541 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
   31542   int32x4_t __ret;
   31543   __ret = (int32x4_t)(__p0);
   31544   return __ret;
   31545 }
   31546 #endif
   31547 
   31548 #ifdef __LITTLE_ENDIAN__
   31549 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
   31550   int32x4_t __ret;
   31551   __ret = (int32x4_t)(__p0);
   31552   return __ret;
   31553 }
   31554 #else
   31555 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
   31556   int32x4_t __ret;
   31557   __ret = (int32x4_t)(__p0);
   31558   return __ret;
   31559 }
   31560 #endif
   31561 
   31562 #ifdef __LITTLE_ENDIAN__
   31563 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
   31564   int32x4_t __ret;
   31565   __ret = (int32x4_t)(__p0);
   31566   return __ret;
   31567 }
   31568 #else
   31569 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
   31570   int32x4_t __ret;
   31571   __ret = (int32x4_t)(__p0);
   31572   return __ret;
   31573 }
   31574 #endif
   31575 
   31576 #ifdef __LITTLE_ENDIAN__
   31577 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
   31578   int32x4_t __ret;
   31579   __ret = (int32x4_t)(__p0);
   31580   return __ret;
   31581 }
   31582 #else
   31583 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
   31584   int32x4_t __ret;
   31585   __ret = (int32x4_t)(__p0);
   31586   return __ret;
   31587 }
   31588 #endif
   31589 
   31590 #ifdef __LITTLE_ENDIAN__
   31591 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
   31592   int32x4_t __ret;
   31593   __ret = (int32x4_t)(__p0);
   31594   return __ret;
   31595 }
   31596 #else
   31597 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
   31598   int32x4_t __ret;
   31599   __ret = (int32x4_t)(__p0);
   31600   return __ret;
   31601 }
   31602 #endif
   31603 
   31604 #ifdef __LITTLE_ENDIAN__
   31605 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
   31606   int32x4_t __ret;
   31607   __ret = (int32x4_t)(__p0);
   31608   return __ret;
   31609 }
   31610 #else
   31611 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
   31612   int32x4_t __ret;
   31613   __ret = (int32x4_t)(__p0);
   31614   return __ret;
   31615 }
   31616 #endif
   31617 
   31618 #ifdef __LITTLE_ENDIAN__
   31619 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
   31620   int32x4_t __ret;
   31621   __ret = (int32x4_t)(__p0);
   31622   return __ret;
   31623 }
   31624 #else
   31625 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
   31626   int32x4_t __ret;
   31627   __ret = (int32x4_t)(__p0);
   31628   return __ret;
   31629 }
   31630 #endif
   31631 
   31632 #ifdef __LITTLE_ENDIAN__
   31633 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
   31634   int32x4_t __ret;
   31635   __ret = (int32x4_t)(__p0);
   31636   return __ret;
   31637 }
   31638 #else
   31639 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
   31640   int32x4_t __ret;
   31641   __ret = (int32x4_t)(__p0);
   31642   return __ret;
   31643 }
   31644 #endif
   31645 
   31646 #ifdef __LITTLE_ENDIAN__
   31647 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
   31648   int32x4_t __ret;
   31649   __ret = (int32x4_t)(__p0);
   31650   return __ret;
   31651 }
   31652 #else
   31653 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
   31654   int32x4_t __ret;
   31655   __ret = (int32x4_t)(__p0);
   31656   return __ret;
   31657 }
   31658 #endif
   31659 
   31660 #ifdef __LITTLE_ENDIAN__
   31661 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
   31662   int32x4_t __ret;
   31663   __ret = (int32x4_t)(__p0);
   31664   return __ret;
   31665 }
   31666 #else
   31667 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
   31668   int32x4_t __ret;
   31669   __ret = (int32x4_t)(__p0);
   31670   return __ret;
   31671 }
   31672 #endif
   31673 
   31674 #ifdef __LITTLE_ENDIAN__
   31675 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
   31676   int32x4_t __ret;
   31677   __ret = (int32x4_t)(__p0);
   31678   return __ret;
   31679 }
   31680 #else
   31681 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
   31682   int32x4_t __ret;
   31683   __ret = (int32x4_t)(__p0);
   31684   return __ret;
   31685 }
   31686 #endif
   31687 
   31688 #ifdef __LITTLE_ENDIAN__
   31689 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
   31690   int64x2_t __ret;
   31691   __ret = (int64x2_t)(__p0);
   31692   return __ret;
   31693 }
   31694 #else
   31695 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
   31696   int64x2_t __ret;
   31697   __ret = (int64x2_t)(__p0);
   31698   return __ret;
   31699 }
   31700 #endif
   31701 
   31702 #ifdef __LITTLE_ENDIAN__
   31703 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
   31704   int64x2_t __ret;
   31705   __ret = (int64x2_t)(__p0);
   31706   return __ret;
   31707 }
   31708 #else
   31709 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
   31710   int64x2_t __ret;
   31711   __ret = (int64x2_t)(__p0);
   31712   return __ret;
   31713 }
   31714 #endif
   31715 
   31716 #ifdef __LITTLE_ENDIAN__
   31717 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
   31718   int64x2_t __ret;
   31719   __ret = (int64x2_t)(__p0);
   31720   return __ret;
   31721 }
   31722 #else
   31723 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
   31724   int64x2_t __ret;
   31725   __ret = (int64x2_t)(__p0);
   31726   return __ret;
   31727 }
   31728 #endif
   31729 
   31730 #ifdef __LITTLE_ENDIAN__
   31731 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
   31732   int64x2_t __ret;
   31733   __ret = (int64x2_t)(__p0);
   31734   return __ret;
   31735 }
   31736 #else
   31737 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
   31738   int64x2_t __ret;
   31739   __ret = (int64x2_t)(__p0);
   31740   return __ret;
   31741 }
   31742 #endif
   31743 
   31744 #ifdef __LITTLE_ENDIAN__
   31745 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
   31746   int64x2_t __ret;
   31747   __ret = (int64x2_t)(__p0);
   31748   return __ret;
   31749 }
   31750 #else
   31751 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
   31752   int64x2_t __ret;
   31753   __ret = (int64x2_t)(__p0);
   31754   return __ret;
   31755 }
   31756 #endif
   31757 
   31758 #ifdef __LITTLE_ENDIAN__
   31759 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
   31760   int64x2_t __ret;
   31761   __ret = (int64x2_t)(__p0);
   31762   return __ret;
   31763 }
   31764 #else
   31765 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
   31766   int64x2_t __ret;
   31767   __ret = (int64x2_t)(__p0);
   31768   return __ret;
   31769 }
   31770 #endif
   31771 
   31772 #ifdef __LITTLE_ENDIAN__
   31773 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
   31774   int64x2_t __ret;
   31775   __ret = (int64x2_t)(__p0);
   31776   return __ret;
   31777 }
   31778 #else
   31779 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
   31780   int64x2_t __ret;
   31781   __ret = (int64x2_t)(__p0);
   31782   return __ret;
   31783 }
   31784 #endif
   31785 
   31786 #ifdef __LITTLE_ENDIAN__
   31787 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
   31788   int64x2_t __ret;
   31789   __ret = (int64x2_t)(__p0);
   31790   return __ret;
   31791 }
   31792 #else
   31793 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
   31794   int64x2_t __ret;
   31795   __ret = (int64x2_t)(__p0);
   31796   return __ret;
   31797 }
   31798 #endif
   31799 
   31800 #ifdef __LITTLE_ENDIAN__
   31801 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
   31802   int64x2_t __ret;
   31803   __ret = (int64x2_t)(__p0);
   31804   return __ret;
   31805 }
   31806 #else
   31807 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
   31808   int64x2_t __ret;
   31809   __ret = (int64x2_t)(__p0);
   31810   return __ret;
   31811 }
   31812 #endif
   31813 
   31814 #ifdef __LITTLE_ENDIAN__
   31815 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
   31816   int64x2_t __ret;
   31817   __ret = (int64x2_t)(__p0);
   31818   return __ret;
   31819 }
   31820 #else
   31821 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
   31822   int64x2_t __ret;
   31823   __ret = (int64x2_t)(__p0);
   31824   return __ret;
   31825 }
   31826 #endif
   31827 
   31828 #ifdef __LITTLE_ENDIAN__
   31829 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
   31830   int64x2_t __ret;
   31831   __ret = (int64x2_t)(__p0);
   31832   return __ret;
   31833 }
   31834 #else
   31835 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
   31836   int64x2_t __ret;
   31837   __ret = (int64x2_t)(__p0);
   31838   return __ret;
   31839 }
   31840 #endif
   31841 
   31842 #ifdef __LITTLE_ENDIAN__
   31843 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
   31844   int16x8_t __ret;
   31845   __ret = (int16x8_t)(__p0);
   31846   return __ret;
   31847 }
   31848 #else
   31849 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
   31850   int16x8_t __ret;
   31851   __ret = (int16x8_t)(__p0);
   31852   return __ret;
   31853 }
   31854 #endif
   31855 
   31856 #ifdef __LITTLE_ENDIAN__
   31857 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
   31858   int16x8_t __ret;
   31859   __ret = (int16x8_t)(__p0);
   31860   return __ret;
   31861 }
   31862 #else
   31863 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
   31864   int16x8_t __ret;
   31865   __ret = (int16x8_t)(__p0);
   31866   return __ret;
   31867 }
   31868 #endif
   31869 
   31870 #ifdef __LITTLE_ENDIAN__
   31871 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
   31872   int16x8_t __ret;
   31873   __ret = (int16x8_t)(__p0);
   31874   return __ret;
   31875 }
   31876 #else
   31877 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
   31878   int16x8_t __ret;
   31879   __ret = (int16x8_t)(__p0);
   31880   return __ret;
   31881 }
   31882 #endif
   31883 
   31884 #ifdef __LITTLE_ENDIAN__
   31885 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
   31886   int16x8_t __ret;
   31887   __ret = (int16x8_t)(__p0);
   31888   return __ret;
   31889 }
   31890 #else
   31891 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
   31892   int16x8_t __ret;
   31893   __ret = (int16x8_t)(__p0);
   31894   return __ret;
   31895 }
   31896 #endif
   31897 
   31898 #ifdef __LITTLE_ENDIAN__
   31899 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
   31900   int16x8_t __ret;
   31901   __ret = (int16x8_t)(__p0);
   31902   return __ret;
   31903 }
   31904 #else
   31905 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
   31906   int16x8_t __ret;
   31907   __ret = (int16x8_t)(__p0);
   31908   return __ret;
   31909 }
   31910 #endif
   31911 
   31912 #ifdef __LITTLE_ENDIAN__
   31913 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
   31914   int16x8_t __ret;
   31915   __ret = (int16x8_t)(__p0);
   31916   return __ret;
   31917 }
   31918 #else
   31919 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
   31920   int16x8_t __ret;
   31921   __ret = (int16x8_t)(__p0);
   31922   return __ret;
   31923 }
   31924 #endif
   31925 
   31926 #ifdef __LITTLE_ENDIAN__
   31927 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
   31928   int16x8_t __ret;
   31929   __ret = (int16x8_t)(__p0);
   31930   return __ret;
   31931 }
   31932 #else
   31933 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
   31934   int16x8_t __ret;
   31935   __ret = (int16x8_t)(__p0);
   31936   return __ret;
   31937 }
   31938 #endif
   31939 
   31940 #ifdef __LITTLE_ENDIAN__
   31941 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
   31942   int16x8_t __ret;
   31943   __ret = (int16x8_t)(__p0);
   31944   return __ret;
   31945 }
   31946 #else
   31947 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
   31948   int16x8_t __ret;
   31949   __ret = (int16x8_t)(__p0);
   31950   return __ret;
   31951 }
   31952 #endif
   31953 
   31954 #ifdef __LITTLE_ENDIAN__
   31955 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
   31956   int16x8_t __ret;
   31957   __ret = (int16x8_t)(__p0);
   31958   return __ret;
   31959 }
   31960 #else
   31961 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
   31962   int16x8_t __ret;
   31963   __ret = (int16x8_t)(__p0);
   31964   return __ret;
   31965 }
   31966 #endif
   31967 
   31968 #ifdef __LITTLE_ENDIAN__
   31969 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
   31970   int16x8_t __ret;
   31971   __ret = (int16x8_t)(__p0);
   31972   return __ret;
   31973 }
   31974 #else
   31975 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
   31976   int16x8_t __ret;
   31977   __ret = (int16x8_t)(__p0);
   31978   return __ret;
   31979 }
   31980 #endif
   31981 
   31982 #ifdef __LITTLE_ENDIAN__
   31983 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
   31984   int16x8_t __ret;
   31985   __ret = (int16x8_t)(__p0);
   31986   return __ret;
   31987 }
   31988 #else
   31989 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
   31990   int16x8_t __ret;
   31991   __ret = (int16x8_t)(__p0);
   31992   return __ret;
   31993 }
   31994 #endif
   31995 
   31996 #ifdef __LITTLE_ENDIAN__
   31997 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
   31998   uint8x8_t __ret;
   31999   __ret = (uint8x8_t)(__p0);
   32000   return __ret;
   32001 }
   32002 #else
   32003 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
   32004   uint8x8_t __ret;
   32005   __ret = (uint8x8_t)(__p0);
   32006   return __ret;
   32007 }
   32008 #endif
   32009 
   32010 #ifdef __LITTLE_ENDIAN__
   32011 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
   32012   uint8x8_t __ret;
   32013   __ret = (uint8x8_t)(__p0);
   32014   return __ret;
   32015 }
   32016 #else
   32017 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
   32018   uint8x8_t __ret;
   32019   __ret = (uint8x8_t)(__p0);
   32020   return __ret;
   32021 }
   32022 #endif
   32023 
   32024 #ifdef __LITTLE_ENDIAN__
   32025 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
   32026   uint8x8_t __ret;
   32027   __ret = (uint8x8_t)(__p0);
   32028   return __ret;
   32029 }
   32030 #else
   32031 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
   32032   uint8x8_t __ret;
   32033   __ret = (uint8x8_t)(__p0);
   32034   return __ret;
   32035 }
   32036 #endif
   32037 
   32038 #ifdef __LITTLE_ENDIAN__
   32039 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
   32040   uint8x8_t __ret;
   32041   __ret = (uint8x8_t)(__p0);
   32042   return __ret;
   32043 }
   32044 #else
   32045 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
   32046   uint8x8_t __ret;
   32047   __ret = (uint8x8_t)(__p0);
   32048   return __ret;
   32049 }
   32050 #endif
   32051 
   32052 #ifdef __LITTLE_ENDIAN__
   32053 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
   32054   uint8x8_t __ret;
   32055   __ret = (uint8x8_t)(__p0);
   32056   return __ret;
   32057 }
   32058 #else
   32059 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
   32060   uint8x8_t __ret;
   32061   __ret = (uint8x8_t)(__p0);
   32062   return __ret;
   32063 }
   32064 #endif
   32065 
   32066 #ifdef __LITTLE_ENDIAN__
   32067 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
   32068   uint8x8_t __ret;
   32069   __ret = (uint8x8_t)(__p0);
   32070   return __ret;
   32071 }
   32072 #else
   32073 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
   32074   uint8x8_t __ret;
   32075   __ret = (uint8x8_t)(__p0);
   32076   return __ret;
   32077 }
   32078 #endif
   32079 
   32080 #ifdef __LITTLE_ENDIAN__
   32081 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
   32082   uint8x8_t __ret;
   32083   __ret = (uint8x8_t)(__p0);
   32084   return __ret;
   32085 }
   32086 #else
   32087 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
   32088   uint8x8_t __ret;
   32089   __ret = (uint8x8_t)(__p0);
   32090   return __ret;
   32091 }
   32092 #endif
   32093 
   32094 #ifdef __LITTLE_ENDIAN__
   32095 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
   32096   uint8x8_t __ret;
   32097   __ret = (uint8x8_t)(__p0);
   32098   return __ret;
   32099 }
   32100 #else
   32101 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
   32102   uint8x8_t __ret;
   32103   __ret = (uint8x8_t)(__p0);
   32104   return __ret;
   32105 }
   32106 #endif
   32107 
   32108 #ifdef __LITTLE_ENDIAN__
   32109 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
   32110   uint8x8_t __ret;
   32111   __ret = (uint8x8_t)(__p0);
   32112   return __ret;
   32113 }
   32114 #else
   32115 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
   32116   uint8x8_t __ret;
   32117   __ret = (uint8x8_t)(__p0);
   32118   return __ret;
   32119 }
   32120 #endif
   32121 
   32122 #ifdef __LITTLE_ENDIAN__
   32123 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
   32124   uint8x8_t __ret;
   32125   __ret = (uint8x8_t)(__p0);
   32126   return __ret;
   32127 }
   32128 #else
   32129 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
   32130   uint8x8_t __ret;
   32131   __ret = (uint8x8_t)(__p0);
   32132   return __ret;
   32133 }
   32134 #endif
   32135 
   32136 #ifdef __LITTLE_ENDIAN__
   32137 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
   32138   uint8x8_t __ret;
   32139   __ret = (uint8x8_t)(__p0);
   32140   return __ret;
   32141 }
   32142 #else
   32143 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
   32144   uint8x8_t __ret;
   32145   __ret = (uint8x8_t)(__p0);
   32146   return __ret;
   32147 }
   32148 #endif
   32149 
   32150 #ifdef __LITTLE_ENDIAN__
   32151 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
   32152   uint32x2_t __ret;
   32153   __ret = (uint32x2_t)(__p0);
   32154   return __ret;
   32155 }
   32156 #else
   32157 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
   32158   uint32x2_t __ret;
   32159   __ret = (uint32x2_t)(__p0);
   32160   return __ret;
   32161 }
   32162 #endif
   32163 
   32164 #ifdef __LITTLE_ENDIAN__
   32165 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
   32166   uint32x2_t __ret;
   32167   __ret = (uint32x2_t)(__p0);
   32168   return __ret;
   32169 }
   32170 #else
   32171 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
   32172   uint32x2_t __ret;
   32173   __ret = (uint32x2_t)(__p0);
   32174   return __ret;
   32175 }
   32176 #endif
   32177 
   32178 #ifdef __LITTLE_ENDIAN__
   32179 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
   32180   uint32x2_t __ret;
   32181   __ret = (uint32x2_t)(__p0);
   32182   return __ret;
   32183 }
   32184 #else
   32185 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
   32186   uint32x2_t __ret;
   32187   __ret = (uint32x2_t)(__p0);
   32188   return __ret;
   32189 }
   32190 #endif
   32191 
   32192 #ifdef __LITTLE_ENDIAN__
   32193 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
   32194   uint32x2_t __ret;
   32195   __ret = (uint32x2_t)(__p0);
   32196   return __ret;
   32197 }
   32198 #else
   32199 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
   32200   uint32x2_t __ret;
   32201   __ret = (uint32x2_t)(__p0);
   32202   return __ret;
   32203 }
   32204 #endif
   32205 
   32206 #ifdef __LITTLE_ENDIAN__
   32207 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
   32208   uint32x2_t __ret;
   32209   __ret = (uint32x2_t)(__p0);
   32210   return __ret;
   32211 }
   32212 #else
   32213 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
   32214   uint32x2_t __ret;
   32215   __ret = (uint32x2_t)(__p0);
   32216   return __ret;
   32217 }
   32218 #endif
   32219 
   32220 #ifdef __LITTLE_ENDIAN__
   32221 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
   32222   uint32x2_t __ret;
   32223   __ret = (uint32x2_t)(__p0);
   32224   return __ret;
   32225 }
   32226 #else
   32227 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
   32228   uint32x2_t __ret;
   32229   __ret = (uint32x2_t)(__p0);
   32230   return __ret;
   32231 }
   32232 #endif
   32233 
   32234 #ifdef __LITTLE_ENDIAN__
   32235 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
   32236   uint32x2_t __ret;
   32237   __ret = (uint32x2_t)(__p0);
   32238   return __ret;
   32239 }
   32240 #else
   32241 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
   32242   uint32x2_t __ret;
   32243   __ret = (uint32x2_t)(__p0);
   32244   return __ret;
   32245 }
   32246 #endif
   32247 
   32248 #ifdef __LITTLE_ENDIAN__
   32249 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
   32250   uint32x2_t __ret;
   32251   __ret = (uint32x2_t)(__p0);
   32252   return __ret;
   32253 }
   32254 #else
   32255 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
   32256   uint32x2_t __ret;
   32257   __ret = (uint32x2_t)(__p0);
   32258   return __ret;
   32259 }
   32260 #endif
   32261 
   32262 #ifdef __LITTLE_ENDIAN__
   32263 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
   32264   uint32x2_t __ret;
   32265   __ret = (uint32x2_t)(__p0);
   32266   return __ret;
   32267 }
   32268 #else
   32269 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
   32270   uint32x2_t __ret;
   32271   __ret = (uint32x2_t)(__p0);
   32272   return __ret;
   32273 }
   32274 #endif
   32275 
   32276 #ifdef __LITTLE_ENDIAN__
   32277 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
   32278   uint32x2_t __ret;
   32279   __ret = (uint32x2_t)(__p0);
   32280   return __ret;
   32281 }
   32282 #else
   32283 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
   32284   uint32x2_t __ret;
   32285   __ret = (uint32x2_t)(__p0);
   32286   return __ret;
   32287 }
   32288 #endif
   32289 
   32290 #ifdef __LITTLE_ENDIAN__
   32291 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
   32292   uint32x2_t __ret;
   32293   __ret = (uint32x2_t)(__p0);
   32294   return __ret;
   32295 }
   32296 #else
   32297 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
   32298   uint32x2_t __ret;
   32299   __ret = (uint32x2_t)(__p0);
   32300   return __ret;
   32301 }
   32302 #endif
   32303 
   32304 #ifdef __LITTLE_ENDIAN__
   32305 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
   32306   uint64x1_t __ret;
   32307   __ret = (uint64x1_t)(__p0);
   32308   return __ret;
   32309 }
   32310 #else
   32311 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
   32312   uint64x1_t __ret;
   32313   __ret = (uint64x1_t)(__p0);
   32314   return __ret;
   32315 }
   32316 #endif
   32317 
   32318 #ifdef __LITTLE_ENDIAN__
   32319 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
   32320   uint64x1_t __ret;
   32321   __ret = (uint64x1_t)(__p0);
   32322   return __ret;
   32323 }
   32324 #else
   32325 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
   32326   uint64x1_t __ret;
   32327   __ret = (uint64x1_t)(__p0);
   32328   return __ret;
   32329 }
   32330 #endif
   32331 
   32332 #ifdef __LITTLE_ENDIAN__
   32333 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
   32334   uint64x1_t __ret;
   32335   __ret = (uint64x1_t)(__p0);
   32336   return __ret;
   32337 }
   32338 #else
   32339 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
   32340   uint64x1_t __ret;
   32341   __ret = (uint64x1_t)(__p0);
   32342   return __ret;
   32343 }
   32344 #endif
   32345 
   32346 #ifdef __LITTLE_ENDIAN__
   32347 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
   32348   uint64x1_t __ret;
   32349   __ret = (uint64x1_t)(__p0);
   32350   return __ret;
   32351 }
   32352 #else
   32353 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
   32354   uint64x1_t __ret;
   32355   __ret = (uint64x1_t)(__p0);
   32356   return __ret;
   32357 }
   32358 #endif
   32359 
   32360 #ifdef __LITTLE_ENDIAN__
   32361 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
   32362   uint64x1_t __ret;
   32363   __ret = (uint64x1_t)(__p0);
   32364   return __ret;
   32365 }
   32366 #else
   32367 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
   32368   uint64x1_t __ret;
   32369   __ret = (uint64x1_t)(__p0);
   32370   return __ret;
   32371 }
   32372 #endif
   32373 
   32374 #ifdef __LITTLE_ENDIAN__
   32375 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
   32376   uint64x1_t __ret;
   32377   __ret = (uint64x1_t)(__p0);
   32378   return __ret;
   32379 }
   32380 #else
   32381 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
   32382   uint64x1_t __ret;
   32383   __ret = (uint64x1_t)(__p0);
   32384   return __ret;
   32385 }
   32386 #endif
   32387 
   32388 #ifdef __LITTLE_ENDIAN__
   32389 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
   32390   uint64x1_t __ret;
   32391   __ret = (uint64x1_t)(__p0);
   32392   return __ret;
   32393 }
   32394 #else
   32395 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
   32396   uint64x1_t __ret;
   32397   __ret = (uint64x1_t)(__p0);
   32398   return __ret;
   32399 }
   32400 #endif
   32401 
   32402 #ifdef __LITTLE_ENDIAN__
   32403 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
   32404   uint64x1_t __ret;
   32405   __ret = (uint64x1_t)(__p0);
   32406   return __ret;
   32407 }
   32408 #else
   32409 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
   32410   uint64x1_t __ret;
   32411   __ret = (uint64x1_t)(__p0);
   32412   return __ret;
   32413 }
   32414 #endif
   32415 
   32416 #ifdef __LITTLE_ENDIAN__
   32417 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
   32418   uint64x1_t __ret;
   32419   __ret = (uint64x1_t)(__p0);
   32420   return __ret;
   32421 }
   32422 #else
   32423 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
   32424   uint64x1_t __ret;
   32425   __ret = (uint64x1_t)(__p0);
   32426   return __ret;
   32427 }
   32428 #endif
   32429 
   32430 #ifdef __LITTLE_ENDIAN__
   32431 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
   32432   uint64x1_t __ret;
   32433   __ret = (uint64x1_t)(__p0);
   32434   return __ret;
   32435 }
   32436 #else
   32437 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
   32438   uint64x1_t __ret;
   32439   __ret = (uint64x1_t)(__p0);
   32440   return __ret;
   32441 }
   32442 #endif
   32443 
   32444 #ifdef __LITTLE_ENDIAN__
   32445 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
   32446   uint64x1_t __ret;
   32447   __ret = (uint64x1_t)(__p0);
   32448   return __ret;
   32449 }
   32450 #else
   32451 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
   32452   uint64x1_t __ret;
   32453   __ret = (uint64x1_t)(__p0);
   32454   return __ret;
   32455 }
   32456 #endif
   32457 
   32458 #ifdef __LITTLE_ENDIAN__
   32459 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
   32460   uint16x4_t __ret;
   32461   __ret = (uint16x4_t)(__p0);
   32462   return __ret;
   32463 }
   32464 #else
   32465 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
   32466   uint16x4_t __ret;
   32467   __ret = (uint16x4_t)(__p0);
   32468   return __ret;
   32469 }
   32470 #endif
   32471 
   32472 #ifdef __LITTLE_ENDIAN__
   32473 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
   32474   uint16x4_t __ret;
   32475   __ret = (uint16x4_t)(__p0);
   32476   return __ret;
   32477 }
   32478 #else
   32479 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
   32480   uint16x4_t __ret;
   32481   __ret = (uint16x4_t)(__p0);
   32482   return __ret;
   32483 }
   32484 #endif
   32485 
   32486 #ifdef __LITTLE_ENDIAN__
   32487 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
   32488   uint16x4_t __ret;
   32489   __ret = (uint16x4_t)(__p0);
   32490   return __ret;
   32491 }
   32492 #else
   32493 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
   32494   uint16x4_t __ret;
   32495   __ret = (uint16x4_t)(__p0);
   32496   return __ret;
   32497 }
   32498 #endif
   32499 
   32500 #ifdef __LITTLE_ENDIAN__
   32501 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
   32502   uint16x4_t __ret;
   32503   __ret = (uint16x4_t)(__p0);
   32504   return __ret;
   32505 }
   32506 #else
   32507 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
   32508   uint16x4_t __ret;
   32509   __ret = (uint16x4_t)(__p0);
   32510   return __ret;
   32511 }
   32512 #endif
   32513 
   32514 #ifdef __LITTLE_ENDIAN__
   32515 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
   32516   uint16x4_t __ret;
   32517   __ret = (uint16x4_t)(__p0);
   32518   return __ret;
   32519 }
   32520 #else
   32521 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
   32522   uint16x4_t __ret;
   32523   __ret = (uint16x4_t)(__p0);
   32524   return __ret;
   32525 }
   32526 #endif
   32527 
   32528 #ifdef __LITTLE_ENDIAN__
   32529 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
   32530   uint16x4_t __ret;
   32531   __ret = (uint16x4_t)(__p0);
   32532   return __ret;
   32533 }
   32534 #else
   32535 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
   32536   uint16x4_t __ret;
   32537   __ret = (uint16x4_t)(__p0);
   32538   return __ret;
   32539 }
   32540 #endif
   32541 
   32542 #ifdef __LITTLE_ENDIAN__
   32543 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
   32544   uint16x4_t __ret;
   32545   __ret = (uint16x4_t)(__p0);
   32546   return __ret;
   32547 }
   32548 #else
   32549 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
   32550   uint16x4_t __ret;
   32551   __ret = (uint16x4_t)(__p0);
   32552   return __ret;
   32553 }
   32554 #endif
   32555 
   32556 #ifdef __LITTLE_ENDIAN__
   32557 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
   32558   uint16x4_t __ret;
   32559   __ret = (uint16x4_t)(__p0);
   32560   return __ret;
   32561 }
   32562 #else
   32563 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
   32564   uint16x4_t __ret;
   32565   __ret = (uint16x4_t)(__p0);
   32566   return __ret;
   32567 }
   32568 #endif
   32569 
   32570 #ifdef __LITTLE_ENDIAN__
   32571 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
   32572   uint16x4_t __ret;
   32573   __ret = (uint16x4_t)(__p0);
   32574   return __ret;
   32575 }
   32576 #else
   32577 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
   32578   uint16x4_t __ret;
   32579   __ret = (uint16x4_t)(__p0);
   32580   return __ret;
   32581 }
   32582 #endif
   32583 
   32584 #ifdef __LITTLE_ENDIAN__
   32585 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
   32586   uint16x4_t __ret;
   32587   __ret = (uint16x4_t)(__p0);
   32588   return __ret;
   32589 }
   32590 #else
   32591 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
   32592   uint16x4_t __ret;
   32593   __ret = (uint16x4_t)(__p0);
   32594   return __ret;
   32595 }
   32596 #endif
   32597 
   32598 #ifdef __LITTLE_ENDIAN__
   32599 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
   32600   uint16x4_t __ret;
   32601   __ret = (uint16x4_t)(__p0);
   32602   return __ret;
   32603 }
   32604 #else
   32605 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
   32606   uint16x4_t __ret;
   32607   __ret = (uint16x4_t)(__p0);
   32608   return __ret;
   32609 }
   32610 #endif
   32611 
   32612 #ifdef __LITTLE_ENDIAN__
   32613 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
   32614   int8x8_t __ret;
   32615   __ret = (int8x8_t)(__p0);
   32616   return __ret;
   32617 }
   32618 #else
   32619 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
   32620   int8x8_t __ret;
   32621   __ret = (int8x8_t)(__p0);
   32622   return __ret;
   32623 }
   32624 #endif
   32625 
   32626 #ifdef __LITTLE_ENDIAN__
   32627 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
   32628   int8x8_t __ret;
   32629   __ret = (int8x8_t)(__p0);
   32630   return __ret;
   32631 }
   32632 #else
   32633 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
   32634   int8x8_t __ret;
   32635   __ret = (int8x8_t)(__p0);
   32636   return __ret;
   32637 }
   32638 #endif
   32639 
   32640 #ifdef __LITTLE_ENDIAN__
   32641 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
   32642   int8x8_t __ret;
   32643   __ret = (int8x8_t)(__p0);
   32644   return __ret;
   32645 }
   32646 #else
   32647 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
   32648   int8x8_t __ret;
   32649   __ret = (int8x8_t)(__p0);
   32650   return __ret;
   32651 }
   32652 #endif
   32653 
   32654 #ifdef __LITTLE_ENDIAN__
   32655 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
   32656   int8x8_t __ret;
   32657   __ret = (int8x8_t)(__p0);
   32658   return __ret;
   32659 }
   32660 #else
   32661 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
   32662   int8x8_t __ret;
   32663   __ret = (int8x8_t)(__p0);
   32664   return __ret;
   32665 }
   32666 #endif
   32667 
   32668 #ifdef __LITTLE_ENDIAN__
   32669 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
   32670   int8x8_t __ret;
   32671   __ret = (int8x8_t)(__p0);
   32672   return __ret;
   32673 }
   32674 #else
   32675 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
   32676   int8x8_t __ret;
   32677   __ret = (int8x8_t)(__p0);
   32678   return __ret;
   32679 }
   32680 #endif
   32681 
   32682 #ifdef __LITTLE_ENDIAN__
   32683 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
   32684   int8x8_t __ret;
   32685   __ret = (int8x8_t)(__p0);
   32686   return __ret;
   32687 }
   32688 #else
   32689 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
   32690   int8x8_t __ret;
   32691   __ret = (int8x8_t)(__p0);
   32692   return __ret;
   32693 }
   32694 #endif
   32695 
   32696 #ifdef __LITTLE_ENDIAN__
   32697 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
   32698   int8x8_t __ret;
   32699   __ret = (int8x8_t)(__p0);
   32700   return __ret;
   32701 }
   32702 #else
   32703 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
   32704   int8x8_t __ret;
   32705   __ret = (int8x8_t)(__p0);
   32706   return __ret;
   32707 }
   32708 #endif
   32709 
   32710 #ifdef __LITTLE_ENDIAN__
   32711 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
   32712   int8x8_t __ret;
   32713   __ret = (int8x8_t)(__p0);
   32714   return __ret;
   32715 }
   32716 #else
   32717 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
   32718   int8x8_t __ret;
   32719   __ret = (int8x8_t)(__p0);
   32720   return __ret;
   32721 }
   32722 #endif
   32723 
   32724 #ifdef __LITTLE_ENDIAN__
   32725 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
   32726   int8x8_t __ret;
   32727   __ret = (int8x8_t)(__p0);
   32728   return __ret;
   32729 }
   32730 #else
   32731 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
   32732   int8x8_t __ret;
   32733   __ret = (int8x8_t)(__p0);
   32734   return __ret;
   32735 }
   32736 #endif
   32737 
   32738 #ifdef __LITTLE_ENDIAN__
   32739 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
   32740   int8x8_t __ret;
   32741   __ret = (int8x8_t)(__p0);
   32742   return __ret;
   32743 }
   32744 #else
   32745 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
   32746   int8x8_t __ret;
   32747   __ret = (int8x8_t)(__p0);
   32748   return __ret;
   32749 }
   32750 #endif
   32751 
   32752 #ifdef __LITTLE_ENDIAN__
   32753 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
   32754   int8x8_t __ret;
   32755   __ret = (int8x8_t)(__p0);
   32756   return __ret;
   32757 }
   32758 #else
   32759 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
   32760   int8x8_t __ret;
   32761   __ret = (int8x8_t)(__p0);
   32762   return __ret;
   32763 }
   32764 #endif
   32765 
   32766 #ifdef __LITTLE_ENDIAN__
   32767 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
   32768   float32x2_t __ret;
   32769   __ret = (float32x2_t)(__p0);
   32770   return __ret;
   32771 }
   32772 #else
   32773 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
   32774   float32x2_t __ret;
   32775   __ret = (float32x2_t)(__p0);
   32776   return __ret;
   32777 }
   32778 #endif
   32779 
   32780 #ifdef __LITTLE_ENDIAN__
   32781 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
   32782   float32x2_t __ret;
   32783   __ret = (float32x2_t)(__p0);
   32784   return __ret;
   32785 }
   32786 #else
   32787 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
   32788   float32x2_t __ret;
   32789   __ret = (float32x2_t)(__p0);
   32790   return __ret;
   32791 }
   32792 #endif
   32793 
   32794 #ifdef __LITTLE_ENDIAN__
   32795 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
   32796   float32x2_t __ret;
   32797   __ret = (float32x2_t)(__p0);
   32798   return __ret;
   32799 }
   32800 #else
   32801 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
   32802   float32x2_t __ret;
   32803   __ret = (float32x2_t)(__p0);
   32804   return __ret;
   32805 }
   32806 #endif
   32807 
   32808 #ifdef __LITTLE_ENDIAN__
   32809 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
   32810   float32x2_t __ret;
   32811   __ret = (float32x2_t)(__p0);
   32812   return __ret;
   32813 }
   32814 #else
   32815 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
   32816   float32x2_t __ret;
   32817   __ret = (float32x2_t)(__p0);
   32818   return __ret;
   32819 }
   32820 #endif
   32821 
   32822 #ifdef __LITTLE_ENDIAN__
   32823 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
   32824   float32x2_t __ret;
   32825   __ret = (float32x2_t)(__p0);
   32826   return __ret;
   32827 }
   32828 #else
   32829 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
   32830   float32x2_t __ret;
   32831   __ret = (float32x2_t)(__p0);
   32832   return __ret;
   32833 }
   32834 #endif
   32835 
   32836 #ifdef __LITTLE_ENDIAN__
   32837 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
   32838   float32x2_t __ret;
   32839   __ret = (float32x2_t)(__p0);
   32840   return __ret;
   32841 }
   32842 #else
   32843 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
   32844   float32x2_t __ret;
   32845   __ret = (float32x2_t)(__p0);
   32846   return __ret;
   32847 }
   32848 #endif
   32849 
   32850 #ifdef __LITTLE_ENDIAN__
   32851 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
   32852   float32x2_t __ret;
   32853   __ret = (float32x2_t)(__p0);
   32854   return __ret;
   32855 }
   32856 #else
   32857 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
   32858   float32x2_t __ret;
   32859   __ret = (float32x2_t)(__p0);
   32860   return __ret;
   32861 }
   32862 #endif
   32863 
   32864 #ifdef __LITTLE_ENDIAN__
   32865 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
   32866   float32x2_t __ret;
   32867   __ret = (float32x2_t)(__p0);
   32868   return __ret;
   32869 }
   32870 #else
   32871 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
   32872   float32x2_t __ret;
   32873   __ret = (float32x2_t)(__p0);
   32874   return __ret;
   32875 }
   32876 #endif
   32877 
   32878 #ifdef __LITTLE_ENDIAN__
   32879 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
   32880   float32x2_t __ret;
   32881   __ret = (float32x2_t)(__p0);
   32882   return __ret;
   32883 }
   32884 #else
   32885 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
   32886   float32x2_t __ret;
   32887   __ret = (float32x2_t)(__p0);
   32888   return __ret;
   32889 }
   32890 #endif
   32891 
   32892 #ifdef __LITTLE_ENDIAN__
   32893 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
   32894   float32x2_t __ret;
   32895   __ret = (float32x2_t)(__p0);
   32896   return __ret;
   32897 }
   32898 #else
   32899 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
   32900   float32x2_t __ret;
   32901   __ret = (float32x2_t)(__p0);
   32902   return __ret;
   32903 }
   32904 #endif
   32905 
   32906 #ifdef __LITTLE_ENDIAN__
   32907 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
   32908   float32x2_t __ret;
   32909   __ret = (float32x2_t)(__p0);
   32910   return __ret;
   32911 }
   32912 #else
   32913 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
   32914   float32x2_t __ret;
   32915   __ret = (float32x2_t)(__p0);
   32916   return __ret;
   32917 }
   32918 #endif
   32919 
   32920 #ifdef __LITTLE_ENDIAN__
   32921 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
   32922   float16x4_t __ret;
   32923   __ret = (float16x4_t)(__p0);
   32924   return __ret;
   32925 }
   32926 #else
   32927 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
   32928   float16x4_t __ret;
   32929   __ret = (float16x4_t)(__p0);
   32930   return __ret;
   32931 }
   32932 #endif
   32933 
   32934 #ifdef __LITTLE_ENDIAN__
   32935 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
   32936   float16x4_t __ret;
   32937   __ret = (float16x4_t)(__p0);
   32938   return __ret;
   32939 }
   32940 #else
   32941 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
   32942   float16x4_t __ret;
   32943   __ret = (float16x4_t)(__p0);
   32944   return __ret;
   32945 }
   32946 #endif
   32947 
   32948 #ifdef __LITTLE_ENDIAN__
   32949 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
   32950   float16x4_t __ret;
   32951   __ret = (float16x4_t)(__p0);
   32952   return __ret;
   32953 }
   32954 #else
   32955 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
   32956   float16x4_t __ret;
   32957   __ret = (float16x4_t)(__p0);
   32958   return __ret;
   32959 }
   32960 #endif
   32961 
   32962 #ifdef __LITTLE_ENDIAN__
   32963 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
   32964   float16x4_t __ret;
   32965   __ret = (float16x4_t)(__p0);
   32966   return __ret;
   32967 }
   32968 #else
   32969 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
   32970   float16x4_t __ret;
   32971   __ret = (float16x4_t)(__p0);
   32972   return __ret;
   32973 }
   32974 #endif
   32975 
   32976 #ifdef __LITTLE_ENDIAN__
   32977 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
   32978   float16x4_t __ret;
   32979   __ret = (float16x4_t)(__p0);
   32980   return __ret;
   32981 }
   32982 #else
   32983 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
   32984   float16x4_t __ret;
   32985   __ret = (float16x4_t)(__p0);
   32986   return __ret;
   32987 }
   32988 #endif
   32989 
   32990 #ifdef __LITTLE_ENDIAN__
   32991 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
   32992   float16x4_t __ret;
   32993   __ret = (float16x4_t)(__p0);
   32994   return __ret;
   32995 }
   32996 #else
   32997 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
   32998   float16x4_t __ret;
   32999   __ret = (float16x4_t)(__p0);
   33000   return __ret;
   33001 }
   33002 #endif
   33003 
   33004 #ifdef __LITTLE_ENDIAN__
   33005 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
   33006   float16x4_t __ret;
   33007   __ret = (float16x4_t)(__p0);
   33008   return __ret;
   33009 }
   33010 #else
   33011 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
   33012   float16x4_t __ret;
   33013   __ret = (float16x4_t)(__p0);
   33014   return __ret;
   33015 }
   33016 #endif
   33017 
   33018 #ifdef __LITTLE_ENDIAN__
   33019 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
   33020   float16x4_t __ret;
   33021   __ret = (float16x4_t)(__p0);
   33022   return __ret;
   33023 }
   33024 #else
   33025 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
   33026   float16x4_t __ret;
   33027   __ret = (float16x4_t)(__p0);
   33028   return __ret;
   33029 }
   33030 #endif
   33031 
   33032 #ifdef __LITTLE_ENDIAN__
   33033 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
   33034   float16x4_t __ret;
   33035   __ret = (float16x4_t)(__p0);
   33036   return __ret;
   33037 }
   33038 #else
   33039 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
   33040   float16x4_t __ret;
   33041   __ret = (float16x4_t)(__p0);
   33042   return __ret;
   33043 }
   33044 #endif
   33045 
   33046 #ifdef __LITTLE_ENDIAN__
   33047 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
   33048   float16x4_t __ret;
   33049   __ret = (float16x4_t)(__p0);
   33050   return __ret;
   33051 }
   33052 #else
   33053 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
   33054   float16x4_t __ret;
   33055   __ret = (float16x4_t)(__p0);
   33056   return __ret;
   33057 }
   33058 #endif
   33059 
   33060 #ifdef __LITTLE_ENDIAN__
   33061 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
   33062   float16x4_t __ret;
   33063   __ret = (float16x4_t)(__p0);
   33064   return __ret;
   33065 }
   33066 #else
   33067 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
   33068   float16x4_t __ret;
   33069   __ret = (float16x4_t)(__p0);
   33070   return __ret;
   33071 }
   33072 #endif
   33073 
   33074 #ifdef __LITTLE_ENDIAN__
   33075 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
   33076   int32x2_t __ret;
   33077   __ret = (int32x2_t)(__p0);
   33078   return __ret;
   33079 }
   33080 #else
   33081 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
   33082   int32x2_t __ret;
   33083   __ret = (int32x2_t)(__p0);
   33084   return __ret;
   33085 }
   33086 #endif
   33087 
   33088 #ifdef __LITTLE_ENDIAN__
   33089 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
   33090   int32x2_t __ret;
   33091   __ret = (int32x2_t)(__p0);
   33092   return __ret;
   33093 }
   33094 #else
   33095 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
   33096   int32x2_t __ret;
   33097   __ret = (int32x2_t)(__p0);
   33098   return __ret;
   33099 }
   33100 #endif
   33101 
   33102 #ifdef __LITTLE_ENDIAN__
   33103 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
   33104   int32x2_t __ret;
   33105   __ret = (int32x2_t)(__p0);
   33106   return __ret;
   33107 }
   33108 #else
   33109 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
   33110   int32x2_t __ret;
   33111   __ret = (int32x2_t)(__p0);
   33112   return __ret;
   33113 }
   33114 #endif
   33115 
   33116 #ifdef __LITTLE_ENDIAN__
   33117 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
   33118   int32x2_t __ret;
   33119   __ret = (int32x2_t)(__p0);
   33120   return __ret;
   33121 }
   33122 #else
   33123 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
   33124   int32x2_t __ret;
   33125   __ret = (int32x2_t)(__p0);
   33126   return __ret;
   33127 }
   33128 #endif
   33129 
   33130 #ifdef __LITTLE_ENDIAN__
   33131 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
   33132   int32x2_t __ret;
   33133   __ret = (int32x2_t)(__p0);
   33134   return __ret;
   33135 }
   33136 #else
   33137 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
   33138   int32x2_t __ret;
   33139   __ret = (int32x2_t)(__p0);
   33140   return __ret;
   33141 }
   33142 #endif
   33143 
   33144 #ifdef __LITTLE_ENDIAN__
   33145 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
   33146   int32x2_t __ret;
   33147   __ret = (int32x2_t)(__p0);
   33148   return __ret;
   33149 }
   33150 #else
   33151 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
   33152   int32x2_t __ret;
   33153   __ret = (int32x2_t)(__p0);
   33154   return __ret;
   33155 }
   33156 #endif
   33157 
   33158 #ifdef __LITTLE_ENDIAN__
   33159 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
   33160   int32x2_t __ret;
   33161   __ret = (int32x2_t)(__p0);
   33162   return __ret;
   33163 }
   33164 #else
   33165 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
   33166   int32x2_t __ret;
   33167   __ret = (int32x2_t)(__p0);
   33168   return __ret;
   33169 }
   33170 #endif
   33171 
   33172 #ifdef __LITTLE_ENDIAN__
   33173 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
   33174   int32x2_t __ret;
   33175   __ret = (int32x2_t)(__p0);
   33176   return __ret;
   33177 }
   33178 #else
   33179 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
   33180   int32x2_t __ret;
   33181   __ret = (int32x2_t)(__p0);
   33182   return __ret;
   33183 }
   33184 #endif
   33185 
   33186 #ifdef __LITTLE_ENDIAN__
   33187 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
   33188   int32x2_t __ret;
   33189   __ret = (int32x2_t)(__p0);
   33190   return __ret;
   33191 }
   33192 #else
   33193 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
   33194   int32x2_t __ret;
   33195   __ret = (int32x2_t)(__p0);
   33196   return __ret;
   33197 }
   33198 #endif
   33199 
   33200 #ifdef __LITTLE_ENDIAN__
   33201 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
   33202   int32x2_t __ret;
   33203   __ret = (int32x2_t)(__p0);
   33204   return __ret;
   33205 }
   33206 #else
   33207 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
   33208   int32x2_t __ret;
   33209   __ret = (int32x2_t)(__p0);
   33210   return __ret;
   33211 }
   33212 #endif
   33213 
   33214 #ifdef __LITTLE_ENDIAN__
   33215 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
   33216   int32x2_t __ret;
   33217   __ret = (int32x2_t)(__p0);
   33218   return __ret;
   33219 }
   33220 #else
   33221 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
   33222   int32x2_t __ret;
   33223   __ret = (int32x2_t)(__p0);
   33224   return __ret;
   33225 }
   33226 #endif
   33227 
   33228 #ifdef __LITTLE_ENDIAN__
   33229 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
   33230   int64x1_t __ret;
   33231   __ret = (int64x1_t)(__p0);
   33232   return __ret;
   33233 }
   33234 #else
   33235 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
   33236   int64x1_t __ret;
   33237   __ret = (int64x1_t)(__p0);
   33238   return __ret;
   33239 }
   33240 #endif
   33241 
   33242 #ifdef __LITTLE_ENDIAN__
   33243 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
   33244   int64x1_t __ret;
   33245   __ret = (int64x1_t)(__p0);
   33246   return __ret;
   33247 }
   33248 #else
   33249 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
   33250   int64x1_t __ret;
   33251   __ret = (int64x1_t)(__p0);
   33252   return __ret;
   33253 }
   33254 #endif
   33255 
   33256 #ifdef __LITTLE_ENDIAN__
   33257 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
   33258   int64x1_t __ret;
   33259   __ret = (int64x1_t)(__p0);
   33260   return __ret;
   33261 }
   33262 #else
   33263 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
   33264   int64x1_t __ret;
   33265   __ret = (int64x1_t)(__p0);
   33266   return __ret;
   33267 }
   33268 #endif
   33269 
   33270 #ifdef __LITTLE_ENDIAN__
   33271 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
   33272   int64x1_t __ret;
   33273   __ret = (int64x1_t)(__p0);
   33274   return __ret;
   33275 }
   33276 #else
   33277 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
   33278   int64x1_t __ret;
   33279   __ret = (int64x1_t)(__p0);
   33280   return __ret;
   33281 }
   33282 #endif
   33283 
   33284 #ifdef __LITTLE_ENDIAN__
   33285 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
   33286   int64x1_t __ret;
   33287   __ret = (int64x1_t)(__p0);
   33288   return __ret;
   33289 }
   33290 #else
   33291 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
   33292   int64x1_t __ret;
   33293   __ret = (int64x1_t)(__p0);
   33294   return __ret;
   33295 }
   33296 #endif
   33297 
   33298 #ifdef __LITTLE_ENDIAN__
   33299 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
   33300   int64x1_t __ret;
   33301   __ret = (int64x1_t)(__p0);
   33302   return __ret;
   33303 }
   33304 #else
   33305 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
   33306   int64x1_t __ret;
   33307   __ret = (int64x1_t)(__p0);
   33308   return __ret;
   33309 }
   33310 #endif
   33311 
   33312 #ifdef __LITTLE_ENDIAN__
   33313 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
   33314   int64x1_t __ret;
   33315   __ret = (int64x1_t)(__p0);
   33316   return __ret;
   33317 }
   33318 #else
   33319 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
   33320   int64x1_t __ret;
   33321   __ret = (int64x1_t)(__p0);
   33322   return __ret;
   33323 }
   33324 #endif
   33325 
   33326 #ifdef __LITTLE_ENDIAN__
   33327 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
   33328   int64x1_t __ret;
   33329   __ret = (int64x1_t)(__p0);
   33330   return __ret;
   33331 }
   33332 #else
   33333 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
   33334   int64x1_t __ret;
   33335   __ret = (int64x1_t)(__p0);
   33336   return __ret;
   33337 }
   33338 #endif
   33339 
   33340 #ifdef __LITTLE_ENDIAN__
   33341 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
   33342   int64x1_t __ret;
   33343   __ret = (int64x1_t)(__p0);
   33344   return __ret;
   33345 }
   33346 #else
   33347 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
   33348   int64x1_t __ret;
   33349   __ret = (int64x1_t)(__p0);
   33350   return __ret;
   33351 }
   33352 #endif
   33353 
   33354 #ifdef __LITTLE_ENDIAN__
   33355 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
   33356   int64x1_t __ret;
   33357   __ret = (int64x1_t)(__p0);
   33358   return __ret;
   33359 }
   33360 #else
   33361 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
   33362   int64x1_t __ret;
   33363   __ret = (int64x1_t)(__p0);
   33364   return __ret;
   33365 }
   33366 #endif
   33367 
   33368 #ifdef __LITTLE_ENDIAN__
   33369 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
   33370   int64x1_t __ret;
   33371   __ret = (int64x1_t)(__p0);
   33372   return __ret;
   33373 }
   33374 #else
   33375 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
   33376   int64x1_t __ret;
   33377   __ret = (int64x1_t)(__p0);
   33378   return __ret;
   33379 }
   33380 #endif
   33381 
   33382 #ifdef __LITTLE_ENDIAN__
   33383 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
   33384   int16x4_t __ret;
   33385   __ret = (int16x4_t)(__p0);
   33386   return __ret;
   33387 }
   33388 #else
   33389 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
   33390   int16x4_t __ret;
   33391   __ret = (int16x4_t)(__p0);
   33392   return __ret;
   33393 }
   33394 #endif
   33395 
   33396 #ifdef __LITTLE_ENDIAN__
   33397 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
   33398   int16x4_t __ret;
   33399   __ret = (int16x4_t)(__p0);
   33400   return __ret;
   33401 }
   33402 #else
   33403 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
   33404   int16x4_t __ret;
   33405   __ret = (int16x4_t)(__p0);
   33406   return __ret;
   33407 }
   33408 #endif
   33409 
   33410 #ifdef __LITTLE_ENDIAN__
   33411 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
   33412   int16x4_t __ret;
   33413   __ret = (int16x4_t)(__p0);
   33414   return __ret;
   33415 }
   33416 #else
   33417 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
   33418   int16x4_t __ret;
   33419   __ret = (int16x4_t)(__p0);
   33420   return __ret;
   33421 }
   33422 #endif
   33423 
   33424 #ifdef __LITTLE_ENDIAN__
   33425 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
   33426   int16x4_t __ret;
   33427   __ret = (int16x4_t)(__p0);
   33428   return __ret;
   33429 }
   33430 #else
   33431 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
   33432   int16x4_t __ret;
   33433   __ret = (int16x4_t)(__p0);
   33434   return __ret;
   33435 }
   33436 #endif
   33437 
   33438 #ifdef __LITTLE_ENDIAN__
   33439 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
   33440   int16x4_t __ret;
   33441   __ret = (int16x4_t)(__p0);
   33442   return __ret;
   33443 }
   33444 #else
   33445 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
   33446   int16x4_t __ret;
   33447   __ret = (int16x4_t)(__p0);
   33448   return __ret;
   33449 }
   33450 #endif
   33451 
   33452 #ifdef __LITTLE_ENDIAN__
   33453 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
   33454   int16x4_t __ret;
   33455   __ret = (int16x4_t)(__p0);
   33456   return __ret;
   33457 }
   33458 #else
   33459 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
   33460   int16x4_t __ret;
   33461   __ret = (int16x4_t)(__p0);
   33462   return __ret;
   33463 }
   33464 #endif
   33465 
   33466 #ifdef __LITTLE_ENDIAN__
   33467 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
   33468   int16x4_t __ret;
   33469   __ret = (int16x4_t)(__p0);
   33470   return __ret;
   33471 }
   33472 #else
   33473 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
   33474   int16x4_t __ret;
   33475   __ret = (int16x4_t)(__p0);
   33476   return __ret;
   33477 }
   33478 #endif
   33479 
   33480 #ifdef __LITTLE_ENDIAN__
   33481 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
   33482   int16x4_t __ret;
   33483   __ret = (int16x4_t)(__p0);
   33484   return __ret;
   33485 }
   33486 #else
   33487 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
   33488   int16x4_t __ret;
   33489   __ret = (int16x4_t)(__p0);
   33490   return __ret;
   33491 }
   33492 #endif
   33493 
   33494 #ifdef __LITTLE_ENDIAN__
   33495 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
   33496   int16x4_t __ret;
   33497   __ret = (int16x4_t)(__p0);
   33498   return __ret;
   33499 }
   33500 #else
   33501 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
   33502   int16x4_t __ret;
   33503   __ret = (int16x4_t)(__p0);
   33504   return __ret;
   33505 }
   33506 #endif
   33507 
   33508 #ifdef __LITTLE_ENDIAN__
   33509 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
   33510   int16x4_t __ret;
   33511   __ret = (int16x4_t)(__p0);
   33512   return __ret;
   33513 }
   33514 #else
   33515 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
   33516   int16x4_t __ret;
   33517   __ret = (int16x4_t)(__p0);
   33518   return __ret;
   33519 }
   33520 #endif
   33521 
   33522 #ifdef __LITTLE_ENDIAN__
   33523 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
   33524   int16x4_t __ret;
   33525   __ret = (int16x4_t)(__p0);
   33526   return __ret;
   33527 }
   33528 #else
   33529 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
   33530   int16x4_t __ret;
   33531   __ret = (int16x4_t)(__p0);
   33532   return __ret;
   33533 }
   33534 #endif
   33535 
   33536 #endif
   33537 #if (__ARM_FP & 2)
   33538 #ifdef __LITTLE_ENDIAN__
   33539 __ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
   33540   float16x4_t __ret;
   33541   __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 8);
   33542   return __ret;
   33543 }
   33544 #else
   33545 __ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
   33546   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33547   float16x4_t __ret;
   33548   __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__rev0, 8);
   33549   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33550   return __ret;
   33551 }
   33552 __ai float16x4_t __noswap_vcvt_f16_f32(float32x4_t __p0) {
   33553   float16x4_t __ret;
   33554   __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 8);
   33555   return __ret;
   33556 }
   33557 #endif
   33558 
   33559 #ifdef __LITTLE_ENDIAN__
   33560 __ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
   33561   float32x4_t __ret;
   33562   __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 41);
   33563   return __ret;
   33564 }
   33565 #else
   33566 __ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
   33567   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33568   float32x4_t __ret;
   33569   __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__rev0, 41);
   33570   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33571   return __ret;
   33572 }
   33573 __ai float32x4_t __noswap_vcvt_f32_f16(float16x4_t __p0) {
   33574   float32x4_t __ret;
   33575   __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 41);
   33576   return __ret;
   33577 }
   33578 #endif
   33579 
   33580 #endif
   33581 #if __ARM_ARCH >= 8
   33582 #ifdef __LITTLE_ENDIAN__
   33583 __ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
   33584   int32x4_t __ret;
   33585   __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__p0, 34);
   33586   return __ret;
   33587 }
   33588 #else
   33589 __ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
   33590   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33591   int32x4_t __ret;
   33592   __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__rev0, 34);
   33593   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33594   return __ret;
   33595 }
   33596 #endif
   33597 
   33598 #ifdef __LITTLE_ENDIAN__
   33599 __ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
   33600   int32x2_t __ret;
   33601   __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__p0, 2);
   33602   return __ret;
   33603 }
   33604 #else
   33605 __ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
   33606   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33607   int32x2_t __ret;
   33608   __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__rev0, 2);
   33609   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33610   return __ret;
   33611 }
   33612 #endif
   33613 
   33614 #ifdef __LITTLE_ENDIAN__
   33615 __ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
   33616   uint32x4_t __ret;
   33617   __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__p0, 50);
   33618   return __ret;
   33619 }
   33620 #else
   33621 __ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
   33622   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33623   uint32x4_t __ret;
   33624   __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__rev0, 50);
   33625   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33626   return __ret;
   33627 }
   33628 #endif
   33629 
   33630 #ifdef __LITTLE_ENDIAN__
   33631 __ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
   33632   uint32x2_t __ret;
   33633   __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__p0, 18);
   33634   return __ret;
   33635 }
   33636 #else
   33637 __ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
   33638   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33639   uint32x2_t __ret;
   33640   __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__rev0, 18);
   33641   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33642   return __ret;
   33643 }
   33644 #endif
   33645 
   33646 #ifdef __LITTLE_ENDIAN__
   33647 __ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
   33648   int32x4_t __ret;
   33649   __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__p0, 34);
   33650   return __ret;
   33651 }
   33652 #else
   33653 __ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
   33654   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33655   int32x4_t __ret;
   33656   __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__rev0, 34);
   33657   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33658   return __ret;
   33659 }
   33660 #endif
   33661 
   33662 #ifdef __LITTLE_ENDIAN__
   33663 __ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
   33664   int32x2_t __ret;
   33665   __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__p0, 2);
   33666   return __ret;
   33667 }
   33668 #else
   33669 __ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
   33670   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33671   int32x2_t __ret;
   33672   __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__rev0, 2);
   33673   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33674   return __ret;
   33675 }
   33676 #endif
   33677 
   33678 #ifdef __LITTLE_ENDIAN__
   33679 __ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
   33680   uint32x4_t __ret;
   33681   __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__p0, 50);
   33682   return __ret;
   33683 }
   33684 #else
   33685 __ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
   33686   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33687   uint32x4_t __ret;
   33688   __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__rev0, 50);
   33689   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33690   return __ret;
   33691 }
   33692 #endif
   33693 
   33694 #ifdef __LITTLE_ENDIAN__
   33695 __ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
   33696   uint32x2_t __ret;
   33697   __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__p0, 18);
   33698   return __ret;
   33699 }
   33700 #else
   33701 __ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
   33702   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33703   uint32x2_t __ret;
   33704   __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__rev0, 18);
   33705   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33706   return __ret;
   33707 }
   33708 #endif
   33709 
   33710 #ifdef __LITTLE_ENDIAN__
   33711 __ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
   33712   int32x4_t __ret;
   33713   __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__p0, 34);
   33714   return __ret;
   33715 }
   33716 #else
   33717 __ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
   33718   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33719   int32x4_t __ret;
   33720   __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__rev0, 34);
   33721   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33722   return __ret;
   33723 }
   33724 #endif
   33725 
   33726 #ifdef __LITTLE_ENDIAN__
   33727 __ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
   33728   int32x2_t __ret;
   33729   __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__p0, 2);
   33730   return __ret;
   33731 }
   33732 #else
   33733 __ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
   33734   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33735   int32x2_t __ret;
   33736   __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__rev0, 2);
   33737   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33738   return __ret;
   33739 }
   33740 #endif
   33741 
   33742 #ifdef __LITTLE_ENDIAN__
   33743 __ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
   33744   uint32x4_t __ret;
   33745   __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__p0, 50);
   33746   return __ret;
   33747 }
   33748 #else
   33749 __ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
   33750   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33751   uint32x4_t __ret;
   33752   __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__rev0, 50);
   33753   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33754   return __ret;
   33755 }
   33756 #endif
   33757 
   33758 #ifdef __LITTLE_ENDIAN__
   33759 __ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
   33760   uint32x2_t __ret;
   33761   __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__p0, 18);
   33762   return __ret;
   33763 }
   33764 #else
   33765 __ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
   33766   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33767   uint32x2_t __ret;
   33768   __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__rev0, 18);
   33769   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33770   return __ret;
   33771 }
   33772 #endif
   33773 
   33774 #ifdef __LITTLE_ENDIAN__
   33775 __ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
   33776   int32x4_t __ret;
   33777   __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__p0, 34);
   33778   return __ret;
   33779 }
   33780 #else
   33781 __ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
   33782   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33783   int32x4_t __ret;
   33784   __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__rev0, 34);
   33785   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33786   return __ret;
   33787 }
   33788 #endif
   33789 
   33790 #ifdef __LITTLE_ENDIAN__
   33791 __ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
   33792   int32x2_t __ret;
   33793   __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__p0, 2);
   33794   return __ret;
   33795 }
   33796 #else
   33797 __ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
   33798   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33799   int32x2_t __ret;
   33800   __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__rev0, 2);
   33801   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33802   return __ret;
   33803 }
   33804 #endif
   33805 
   33806 #ifdef __LITTLE_ENDIAN__
   33807 __ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
   33808   uint32x4_t __ret;
   33809   __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__p0, 50);
   33810   return __ret;
   33811 }
   33812 #else
   33813 __ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
   33814   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33815   uint32x4_t __ret;
   33816   __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__rev0, 50);
   33817   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33818   return __ret;
   33819 }
   33820 #endif
   33821 
   33822 #ifdef __LITTLE_ENDIAN__
   33823 __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
   33824   uint32x2_t __ret;
   33825   __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__p0, 18);
   33826   return __ret;
   33827 }
   33828 #else
   33829 __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
   33830   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33831   uint32x2_t __ret;
   33832   __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__rev0, 18);
   33833   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33834   return __ret;
   33835 }
   33836 #endif
   33837 
   33838 #endif
   33839 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
   33840 #ifdef __LITTLE_ENDIAN__
   33841 __ai float32x4_t vrndq_f32(float32x4_t __p0) {
   33842   float32x4_t __ret;
   33843   __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 41);
   33844   return __ret;
   33845 }
   33846 #else
   33847 __ai float32x4_t vrndq_f32(float32x4_t __p0) {
   33848   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33849   float32x4_t __ret;
   33850   __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 41);
   33851   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33852   return __ret;
   33853 }
   33854 #endif
   33855 
   33856 #ifdef __LITTLE_ENDIAN__
   33857 __ai float32x2_t vrnd_f32(float32x2_t __p0) {
   33858   float32x2_t __ret;
   33859   __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 9);
   33860   return __ret;
   33861 }
   33862 #else
   33863 __ai float32x2_t vrnd_f32(float32x2_t __p0) {
   33864   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33865   float32x2_t __ret;
   33866   __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__rev0, 9);
   33867   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33868   return __ret;
   33869 }
   33870 #endif
   33871 
   33872 #ifdef __LITTLE_ENDIAN__
   33873 __ai float32x4_t vrndaq_f32(float32x4_t __p0) {
   33874   float32x4_t __ret;
   33875   __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 41);
   33876   return __ret;
   33877 }
   33878 #else
   33879 __ai float32x4_t vrndaq_f32(float32x4_t __p0) {
   33880   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33881   float32x4_t __ret;
   33882   __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 41);
   33883   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33884   return __ret;
   33885 }
   33886 #endif
   33887 
   33888 #ifdef __LITTLE_ENDIAN__
   33889 __ai float32x2_t vrnda_f32(float32x2_t __p0) {
   33890   float32x2_t __ret;
   33891   __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 9);
   33892   return __ret;
   33893 }
   33894 #else
   33895 __ai float32x2_t vrnda_f32(float32x2_t __p0) {
   33896   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33897   float32x2_t __ret;
   33898   __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__rev0, 9);
   33899   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33900   return __ret;
   33901 }
   33902 #endif
   33903 
   33904 #ifdef __LITTLE_ENDIAN__
   33905 __ai float32x4_t vrndmq_f32(float32x4_t __p0) {
   33906   float32x4_t __ret;
   33907   __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 41);
   33908   return __ret;
   33909 }
   33910 #else
   33911 __ai float32x4_t vrndmq_f32(float32x4_t __p0) {
   33912   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33913   float32x4_t __ret;
   33914   __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 41);
   33915   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33916   return __ret;
   33917 }
   33918 #endif
   33919 
   33920 #ifdef __LITTLE_ENDIAN__
   33921 __ai float32x2_t vrndm_f32(float32x2_t __p0) {
   33922   float32x2_t __ret;
   33923   __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 9);
   33924   return __ret;
   33925 }
   33926 #else
   33927 __ai float32x2_t vrndm_f32(float32x2_t __p0) {
   33928   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33929   float32x2_t __ret;
   33930   __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__rev0, 9);
   33931   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33932   return __ret;
   33933 }
   33934 #endif
   33935 
   33936 #ifdef __LITTLE_ENDIAN__
   33937 __ai float32x4_t vrndnq_f32(float32x4_t __p0) {
   33938   float32x4_t __ret;
   33939   __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 41);
   33940   return __ret;
   33941 }
   33942 #else
   33943 __ai float32x4_t vrndnq_f32(float32x4_t __p0) {
   33944   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33945   float32x4_t __ret;
   33946   __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 41);
   33947   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33948   return __ret;
   33949 }
   33950 #endif
   33951 
   33952 #ifdef __LITTLE_ENDIAN__
   33953 __ai float32x2_t vrndn_f32(float32x2_t __p0) {
   33954   float32x2_t __ret;
   33955   __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 9);
   33956   return __ret;
   33957 }
   33958 #else
   33959 __ai float32x2_t vrndn_f32(float32x2_t __p0) {
   33960   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33961   float32x2_t __ret;
   33962   __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__rev0, 9);
   33963   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33964   return __ret;
   33965 }
   33966 #endif
   33967 
   33968 #ifdef __LITTLE_ENDIAN__
   33969 __ai float32x4_t vrndpq_f32(float32x4_t __p0) {
   33970   float32x4_t __ret;
   33971   __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 41);
   33972   return __ret;
   33973 }
   33974 #else
   33975 __ai float32x4_t vrndpq_f32(float32x4_t __p0) {
   33976   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   33977   float32x4_t __ret;
   33978   __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 41);
   33979   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   33980   return __ret;
   33981 }
   33982 #endif
   33983 
   33984 #ifdef __LITTLE_ENDIAN__
   33985 __ai float32x2_t vrndp_f32(float32x2_t __p0) {
   33986   float32x2_t __ret;
   33987   __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 9);
   33988   return __ret;
   33989 }
   33990 #else
   33991 __ai float32x2_t vrndp_f32(float32x2_t __p0) {
   33992   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   33993   float32x2_t __ret;
   33994   __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__rev0, 9);
   33995   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   33996   return __ret;
   33997 }
   33998 #endif
   33999 
   34000 #ifdef __LITTLE_ENDIAN__
   34001 __ai float32x4_t vrndxq_f32(float32x4_t __p0) {
   34002   float32x4_t __ret;
   34003   __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 41);
   34004   return __ret;
   34005 }
   34006 #else
   34007 __ai float32x4_t vrndxq_f32(float32x4_t __p0) {
   34008   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   34009   float32x4_t __ret;
   34010   __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 41);
   34011   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   34012   return __ret;
   34013 }
   34014 #endif
   34015 
   34016 #ifdef __LITTLE_ENDIAN__
   34017 __ai float32x2_t vrndx_f32(float32x2_t __p0) {
   34018   float32x2_t __ret;
   34019   __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 9);
   34020   return __ret;
   34021 }
   34022 #else
   34023 __ai float32x2_t vrndx_f32(float32x2_t __p0) {
   34024   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34025   float32x2_t __ret;
   34026   __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__rev0, 9);
   34027   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34028   return __ret;
   34029 }
   34030 #endif
   34031 
   34032 #endif
   34033 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
   34034 #ifdef __LITTLE_ENDIAN__
   34035 __ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
   34036   float32x4_t __ret;
   34037   __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   34038   return __ret;
   34039 }
   34040 #else
   34041 __ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
   34042   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   34043   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   34044   float32x4_t __ret;
   34045   __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   34046   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   34047   return __ret;
   34048 }
   34049 #endif
   34050 
   34051 #ifdef __LITTLE_ENDIAN__
   34052 __ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
   34053   float32x2_t __ret;
   34054   __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   34055   return __ret;
   34056 }
   34057 #else
   34058 __ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
   34059   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34060   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   34061   float32x2_t __ret;
   34062   __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   34063   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34064   return __ret;
   34065 }
   34066 #endif
   34067 
   34068 #ifdef __LITTLE_ENDIAN__
   34069 __ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
   34070   float32x4_t __ret;
   34071   __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   34072   return __ret;
   34073 }
   34074 #else
   34075 __ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
   34076   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   34077   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   34078   float32x4_t __ret;
   34079   __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   34080   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   34081   return __ret;
   34082 }
   34083 #endif
   34084 
   34085 #ifdef __LITTLE_ENDIAN__
   34086 __ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
   34087   float32x2_t __ret;
   34088   __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   34089   return __ret;
   34090 }
   34091 #else
   34092 __ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
   34093   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34094   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   34095   float32x2_t __ret;
   34096   __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   34097   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34098   return __ret;
   34099 }
   34100 #endif
   34101 
   34102 #endif
   34103 #if __ARM_ARCH >= 8 && defined(__aarch64__)
   34104 #ifdef __LITTLE_ENDIAN__
   34105 __ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
   34106   int64x2_t __ret;
   34107   __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__p0, 35);
   34108   return __ret;
   34109 }
   34110 #else
   34111 __ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
   34112   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34113   int64x2_t __ret;
   34114   __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__rev0, 35);
   34115   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34116   return __ret;
   34117 }
   34118 #endif
   34119 
   34120 #ifdef __LITTLE_ENDIAN__
   34121 __ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
   34122   int64x1_t __ret;
   34123   __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
   34124   return __ret;
   34125 }
   34126 #else
   34127 __ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
   34128   int64x1_t __ret;
   34129   __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
   34130   return __ret;
   34131 }
   34132 #endif
   34133 
   34134 #ifdef __LITTLE_ENDIAN__
   34135 __ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
   34136   uint64x2_t __ret;
   34137   __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__p0, 51);
   34138   return __ret;
   34139 }
   34140 #else
   34141 __ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
   34142   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34143   uint64x2_t __ret;
   34144   __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__rev0, 51);
   34145   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34146   return __ret;
   34147 }
   34148 #endif
   34149 
   34150 #ifdef __LITTLE_ENDIAN__
   34151 __ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
   34152   uint64x1_t __ret;
   34153   __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
   34154   return __ret;
   34155 }
   34156 #else
   34157 __ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
   34158   uint64x1_t __ret;
   34159   __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
   34160   return __ret;
   34161 }
   34162 #endif
   34163 
   34164 #ifdef __LITTLE_ENDIAN__
   34165 __ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
   34166   int64x2_t __ret;
   34167   __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__p0, 35);
   34168   return __ret;
   34169 }
   34170 #else
   34171 __ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
   34172   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34173   int64x2_t __ret;
   34174   __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__rev0, 35);
   34175   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34176   return __ret;
   34177 }
   34178 #endif
   34179 
   34180 #ifdef __LITTLE_ENDIAN__
   34181 __ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
   34182   int64x1_t __ret;
   34183   __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
   34184   return __ret;
   34185 }
   34186 #else
   34187 __ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
   34188   int64x1_t __ret;
   34189   __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
   34190   return __ret;
   34191 }
   34192 #endif
   34193 
   34194 #ifdef __LITTLE_ENDIAN__
   34195 __ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
   34196   uint64x2_t __ret;
   34197   __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__p0, 51);
   34198   return __ret;
   34199 }
   34200 #else
   34201 __ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
   34202   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34203   uint64x2_t __ret;
   34204   __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__rev0, 51);
   34205   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34206   return __ret;
   34207 }
   34208 #endif
   34209 
   34210 #ifdef __LITTLE_ENDIAN__
   34211 __ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
   34212   uint64x1_t __ret;
   34213   __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
   34214   return __ret;
   34215 }
   34216 #else
   34217 __ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
   34218   uint64x1_t __ret;
   34219   __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
   34220   return __ret;
   34221 }
   34222 #endif
   34223 
   34224 #ifdef __LITTLE_ENDIAN__
   34225 __ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
   34226   int64x2_t __ret;
   34227   __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__p0, 35);
   34228   return __ret;
   34229 }
   34230 #else
   34231 __ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
   34232   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34233   int64x2_t __ret;
   34234   __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__rev0, 35);
   34235   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34236   return __ret;
   34237 }
   34238 #endif
   34239 
   34240 #ifdef __LITTLE_ENDIAN__
   34241 __ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
   34242   int64x1_t __ret;
   34243   __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
   34244   return __ret;
   34245 }
   34246 #else
   34247 __ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
   34248   int64x1_t __ret;
   34249   __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
   34250   return __ret;
   34251 }
   34252 #endif
   34253 
   34254 #ifdef __LITTLE_ENDIAN__
   34255 __ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
   34256   uint64x2_t __ret;
   34257   __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__p0, 51);
   34258   return __ret;
   34259 }
   34260 #else
   34261 __ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
   34262   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34263   uint64x2_t __ret;
   34264   __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__rev0, 51);
   34265   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34266   return __ret;
   34267 }
   34268 #endif
   34269 
   34270 #ifdef __LITTLE_ENDIAN__
   34271 __ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
   34272   uint64x1_t __ret;
   34273   __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
   34274   return __ret;
   34275 }
   34276 #else
   34277 __ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
   34278   uint64x1_t __ret;
   34279   __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
   34280   return __ret;
   34281 }
   34282 #endif
   34283 
   34284 #ifdef __LITTLE_ENDIAN__
   34285 __ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
   34286   int64x2_t __ret;
   34287   __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__p0, 35);
   34288   return __ret;
   34289 }
   34290 #else
   34291 __ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
   34292   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34293   int64x2_t __ret;
   34294   __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__rev0, 35);
   34295   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34296   return __ret;
   34297 }
   34298 #endif
   34299 
   34300 #ifdef __LITTLE_ENDIAN__
   34301 __ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
   34302   int64x1_t __ret;
   34303   __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
   34304   return __ret;
   34305 }
   34306 #else
   34307 __ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
   34308   int64x1_t __ret;
   34309   __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
   34310   return __ret;
   34311 }
   34312 #endif
   34313 
   34314 #ifdef __LITTLE_ENDIAN__
   34315 __ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
   34316   uint64x2_t __ret;
   34317   __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__p0, 51);
   34318   return __ret;
   34319 }
   34320 #else
   34321 __ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
   34322   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   34323   uint64x2_t __ret;
   34324   __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__rev0, 51);
   34325   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   34326   return __ret;
   34327 }
   34328 #endif
   34329 
   34330 #ifdef __LITTLE_ENDIAN__
   34331 __ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
   34332   uint64x1_t __ret;
   34333   __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
   34334   return __ret;
   34335 }
   34336 #else
   34337 __ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
   34338   uint64x1_t __ret;
   34339   __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
   34340   return __ret;
   34341 }
   34342 #endif
   34343 
   34344 #ifdef __LITTLE_ENDIAN__
   34345 __ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
   34346   poly8x8_t __ret;
   34347   __ret = (poly8x8_t)(__p0);
   34348   return __ret;
   34349 }
   34350 #else
   34351 __ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
   34352   poly8x8_t __ret;
   34353   __ret = (poly8x8_t)(__p0);
   34354   return __ret;
   34355 }
   34356 #endif
   34357 
   34358 #ifdef __LITTLE_ENDIAN__
   34359 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
   34360   poly8x8_t __ret;
   34361   __ret = (poly8x8_t)(__p0);
   34362   return __ret;
   34363 }
   34364 #else
   34365 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
   34366   poly8x8_t __ret;
   34367   __ret = (poly8x8_t)(__p0);
   34368   return __ret;
   34369 }
   34370 #endif
   34371 
   34372 #ifdef __LITTLE_ENDIAN__
   34373 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
   34374   poly8x8_t __ret;
   34375   __ret = (poly8x8_t)(__p0);
   34376   return __ret;
   34377 }
   34378 #else
   34379 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
   34380   poly8x8_t __ret;
   34381   __ret = (poly8x8_t)(__p0);
   34382   return __ret;
   34383 }
   34384 #endif
   34385 
   34386 #ifdef __LITTLE_ENDIAN__
   34387 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
   34388   poly8x8_t __ret;
   34389   __ret = (poly8x8_t)(__p0);
   34390   return __ret;
   34391 }
   34392 #else
   34393 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
   34394   poly8x8_t __ret;
   34395   __ret = (poly8x8_t)(__p0);
   34396   return __ret;
   34397 }
   34398 #endif
   34399 
   34400 #ifdef __LITTLE_ENDIAN__
   34401 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
   34402   poly8x8_t __ret;
   34403   __ret = (poly8x8_t)(__p0);
   34404   return __ret;
   34405 }
   34406 #else
   34407 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
   34408   poly8x8_t __ret;
   34409   __ret = (poly8x8_t)(__p0);
   34410   return __ret;
   34411 }
   34412 #endif
   34413 
   34414 #ifdef __LITTLE_ENDIAN__
   34415 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
   34416   poly8x8_t __ret;
   34417   __ret = (poly8x8_t)(__p0);
   34418   return __ret;
   34419 }
   34420 #else
   34421 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
   34422   poly8x8_t __ret;
   34423   __ret = (poly8x8_t)(__p0);
   34424   return __ret;
   34425 }
   34426 #endif
   34427 
   34428 #ifdef __LITTLE_ENDIAN__
   34429 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
   34430   poly8x8_t __ret;
   34431   __ret = (poly8x8_t)(__p0);
   34432   return __ret;
   34433 }
   34434 #else
   34435 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
   34436   poly8x8_t __ret;
   34437   __ret = (poly8x8_t)(__p0);
   34438   return __ret;
   34439 }
   34440 #endif
   34441 
   34442 #ifdef __LITTLE_ENDIAN__
   34443 __ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
   34444   poly8x8_t __ret;
   34445   __ret = (poly8x8_t)(__p0);
   34446   return __ret;
   34447 }
   34448 #else
   34449 __ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
   34450   poly8x8_t __ret;
   34451   __ret = (poly8x8_t)(__p0);
   34452   return __ret;
   34453 }
   34454 #endif
   34455 
   34456 #ifdef __LITTLE_ENDIAN__
   34457 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
   34458   poly8x8_t __ret;
   34459   __ret = (poly8x8_t)(__p0);
   34460   return __ret;
   34461 }
   34462 #else
   34463 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
   34464   poly8x8_t __ret;
   34465   __ret = (poly8x8_t)(__p0);
   34466   return __ret;
   34467 }
   34468 #endif
   34469 
   34470 #ifdef __LITTLE_ENDIAN__
   34471 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
   34472   poly8x8_t __ret;
   34473   __ret = (poly8x8_t)(__p0);
   34474   return __ret;
   34475 }
   34476 #else
   34477 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
   34478   poly8x8_t __ret;
   34479   __ret = (poly8x8_t)(__p0);
   34480   return __ret;
   34481 }
   34482 #endif
   34483 
   34484 #ifdef __LITTLE_ENDIAN__
   34485 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
   34486   poly8x8_t __ret;
   34487   __ret = (poly8x8_t)(__p0);
   34488   return __ret;
   34489 }
   34490 #else
   34491 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
   34492   poly8x8_t __ret;
   34493   __ret = (poly8x8_t)(__p0);
   34494   return __ret;
   34495 }
   34496 #endif
   34497 
   34498 #ifdef __LITTLE_ENDIAN__
   34499 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
   34500   poly8x8_t __ret;
   34501   __ret = (poly8x8_t)(__p0);
   34502   return __ret;
   34503 }
   34504 #else
   34505 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
   34506   poly8x8_t __ret;
   34507   __ret = (poly8x8_t)(__p0);
   34508   return __ret;
   34509 }
   34510 #endif
   34511 
   34512 #ifdef __LITTLE_ENDIAN__
   34513 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
   34514   poly8x8_t __ret;
   34515   __ret = (poly8x8_t)(__p0);
   34516   return __ret;
   34517 }
   34518 #else
   34519 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
   34520   poly8x8_t __ret;
   34521   __ret = (poly8x8_t)(__p0);
   34522   return __ret;
   34523 }
   34524 #endif
   34525 
   34526 #ifdef __LITTLE_ENDIAN__
   34527 __ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
   34528   poly64x1_t __ret;
   34529   __ret = (poly64x1_t)(__p0);
   34530   return __ret;
   34531 }
   34532 #else
   34533 __ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
   34534   poly64x1_t __ret;
   34535   __ret = (poly64x1_t)(__p0);
   34536   return __ret;
   34537 }
   34538 #endif
   34539 
   34540 #ifdef __LITTLE_ENDIAN__
   34541 __ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
   34542   poly64x1_t __ret;
   34543   __ret = (poly64x1_t)(__p0);
   34544   return __ret;
   34545 }
   34546 #else
   34547 __ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
   34548   poly64x1_t __ret;
   34549   __ret = (poly64x1_t)(__p0);
   34550   return __ret;
   34551 }
   34552 #endif
   34553 
   34554 #ifdef __LITTLE_ENDIAN__
   34555 __ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
   34556   poly64x1_t __ret;
   34557   __ret = (poly64x1_t)(__p0);
   34558   return __ret;
   34559 }
   34560 #else
   34561 __ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
   34562   poly64x1_t __ret;
   34563   __ret = (poly64x1_t)(__p0);
   34564   return __ret;
   34565 }
   34566 #endif
   34567 
   34568 #ifdef __LITTLE_ENDIAN__
   34569 __ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
   34570   poly64x1_t __ret;
   34571   __ret = (poly64x1_t)(__p0);
   34572   return __ret;
   34573 }
   34574 #else
   34575 __ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
   34576   poly64x1_t __ret;
   34577   __ret = (poly64x1_t)(__p0);
   34578   return __ret;
   34579 }
   34580 #endif
   34581 
   34582 #ifdef __LITTLE_ENDIAN__
   34583 __ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
   34584   poly64x1_t __ret;
   34585   __ret = (poly64x1_t)(__p0);
   34586   return __ret;
   34587 }
   34588 #else
   34589 __ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
   34590   poly64x1_t __ret;
   34591   __ret = (poly64x1_t)(__p0);
   34592   return __ret;
   34593 }
   34594 #endif
   34595 
   34596 #ifdef __LITTLE_ENDIAN__
   34597 __ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
   34598   poly64x1_t __ret;
   34599   __ret = (poly64x1_t)(__p0);
   34600   return __ret;
   34601 }
   34602 #else
   34603 __ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
   34604   poly64x1_t __ret;
   34605   __ret = (poly64x1_t)(__p0);
   34606   return __ret;
   34607 }
   34608 #endif
   34609 
   34610 #ifdef __LITTLE_ENDIAN__
   34611 __ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
   34612   poly64x1_t __ret;
   34613   __ret = (poly64x1_t)(__p0);
   34614   return __ret;
   34615 }
   34616 #else
   34617 __ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
   34618   poly64x1_t __ret;
   34619   __ret = (poly64x1_t)(__p0);
   34620   return __ret;
   34621 }
   34622 #endif
   34623 
   34624 #ifdef __LITTLE_ENDIAN__
   34625 __ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
   34626   poly64x1_t __ret;
   34627   __ret = (poly64x1_t)(__p0);
   34628   return __ret;
   34629 }
   34630 #else
   34631 __ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
   34632   poly64x1_t __ret;
   34633   __ret = (poly64x1_t)(__p0);
   34634   return __ret;
   34635 }
   34636 #endif
   34637 
   34638 #ifdef __LITTLE_ENDIAN__
   34639 __ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
   34640   poly64x1_t __ret;
   34641   __ret = (poly64x1_t)(__p0);
   34642   return __ret;
   34643 }
   34644 #else
   34645 __ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
   34646   poly64x1_t __ret;
   34647   __ret = (poly64x1_t)(__p0);
   34648   return __ret;
   34649 }
   34650 #endif
   34651 
   34652 #ifdef __LITTLE_ENDIAN__
   34653 __ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
   34654   poly64x1_t __ret;
   34655   __ret = (poly64x1_t)(__p0);
   34656   return __ret;
   34657 }
   34658 #else
   34659 __ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
   34660   poly64x1_t __ret;
   34661   __ret = (poly64x1_t)(__p0);
   34662   return __ret;
   34663 }
   34664 #endif
   34665 
   34666 #ifdef __LITTLE_ENDIAN__
   34667 __ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
   34668   poly64x1_t __ret;
   34669   __ret = (poly64x1_t)(__p0);
   34670   return __ret;
   34671 }
   34672 #else
   34673 __ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
   34674   poly64x1_t __ret;
   34675   __ret = (poly64x1_t)(__p0);
   34676   return __ret;
   34677 }
   34678 #endif
   34679 
   34680 #ifdef __LITTLE_ENDIAN__
   34681 __ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
   34682   poly64x1_t __ret;
   34683   __ret = (poly64x1_t)(__p0);
   34684   return __ret;
   34685 }
   34686 #else
   34687 __ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
   34688   poly64x1_t __ret;
   34689   __ret = (poly64x1_t)(__p0);
   34690   return __ret;
   34691 }
   34692 #endif
   34693 
   34694 #ifdef __LITTLE_ENDIAN__
   34695 __ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
   34696   poly64x1_t __ret;
   34697   __ret = (poly64x1_t)(__p0);
   34698   return __ret;
   34699 }
   34700 #else
   34701 __ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
   34702   poly64x1_t __ret;
   34703   __ret = (poly64x1_t)(__p0);
   34704   return __ret;
   34705 }
   34706 #endif
   34707 
   34708 #ifdef __LITTLE_ENDIAN__
   34709 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
   34710   poly16x4_t __ret;
   34711   __ret = (poly16x4_t)(__p0);
   34712   return __ret;
   34713 }
   34714 #else
   34715 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
   34716   poly16x4_t __ret;
   34717   __ret = (poly16x4_t)(__p0);
   34718   return __ret;
   34719 }
   34720 #endif
   34721 
   34722 #ifdef __LITTLE_ENDIAN__
   34723 __ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
   34724   poly16x4_t __ret;
   34725   __ret = (poly16x4_t)(__p0);
   34726   return __ret;
   34727 }
   34728 #else
   34729 __ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
   34730   poly16x4_t __ret;
   34731   __ret = (poly16x4_t)(__p0);
   34732   return __ret;
   34733 }
   34734 #endif
   34735 
   34736 #ifdef __LITTLE_ENDIAN__
   34737 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
   34738   poly16x4_t __ret;
   34739   __ret = (poly16x4_t)(__p0);
   34740   return __ret;
   34741 }
   34742 #else
   34743 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
   34744   poly16x4_t __ret;
   34745   __ret = (poly16x4_t)(__p0);
   34746   return __ret;
   34747 }
   34748 #endif
   34749 
   34750 #ifdef __LITTLE_ENDIAN__
   34751 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
   34752   poly16x4_t __ret;
   34753   __ret = (poly16x4_t)(__p0);
   34754   return __ret;
   34755 }
   34756 #else
   34757 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
   34758   poly16x4_t __ret;
   34759   __ret = (poly16x4_t)(__p0);
   34760   return __ret;
   34761 }
   34762 #endif
   34763 
   34764 #ifdef __LITTLE_ENDIAN__
   34765 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
   34766   poly16x4_t __ret;
   34767   __ret = (poly16x4_t)(__p0);
   34768   return __ret;
   34769 }
   34770 #else
   34771 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
   34772   poly16x4_t __ret;
   34773   __ret = (poly16x4_t)(__p0);
   34774   return __ret;
   34775 }
   34776 #endif
   34777 
   34778 #ifdef __LITTLE_ENDIAN__
   34779 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
   34780   poly16x4_t __ret;
   34781   __ret = (poly16x4_t)(__p0);
   34782   return __ret;
   34783 }
   34784 #else
   34785 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
   34786   poly16x4_t __ret;
   34787   __ret = (poly16x4_t)(__p0);
   34788   return __ret;
   34789 }
   34790 #endif
   34791 
   34792 #ifdef __LITTLE_ENDIAN__
   34793 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
   34794   poly16x4_t __ret;
   34795   __ret = (poly16x4_t)(__p0);
   34796   return __ret;
   34797 }
   34798 #else
   34799 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
   34800   poly16x4_t __ret;
   34801   __ret = (poly16x4_t)(__p0);
   34802   return __ret;
   34803 }
   34804 #endif
   34805 
   34806 #ifdef __LITTLE_ENDIAN__
   34807 __ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
   34808   poly16x4_t __ret;
   34809   __ret = (poly16x4_t)(__p0);
   34810   return __ret;
   34811 }
   34812 #else
   34813 __ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
   34814   poly16x4_t __ret;
   34815   __ret = (poly16x4_t)(__p0);
   34816   return __ret;
   34817 }
   34818 #endif
   34819 
   34820 #ifdef __LITTLE_ENDIAN__
   34821 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
   34822   poly16x4_t __ret;
   34823   __ret = (poly16x4_t)(__p0);
   34824   return __ret;
   34825 }
   34826 #else
   34827 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
   34828   poly16x4_t __ret;
   34829   __ret = (poly16x4_t)(__p0);
   34830   return __ret;
   34831 }
   34832 #endif
   34833 
   34834 #ifdef __LITTLE_ENDIAN__
   34835 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
   34836   poly16x4_t __ret;
   34837   __ret = (poly16x4_t)(__p0);
   34838   return __ret;
   34839 }
   34840 #else
   34841 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
   34842   poly16x4_t __ret;
   34843   __ret = (poly16x4_t)(__p0);
   34844   return __ret;
   34845 }
   34846 #endif
   34847 
   34848 #ifdef __LITTLE_ENDIAN__
   34849 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
   34850   poly16x4_t __ret;
   34851   __ret = (poly16x4_t)(__p0);
   34852   return __ret;
   34853 }
   34854 #else
   34855 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
   34856   poly16x4_t __ret;
   34857   __ret = (poly16x4_t)(__p0);
   34858   return __ret;
   34859 }
   34860 #endif
   34861 
   34862 #ifdef __LITTLE_ENDIAN__
   34863 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
   34864   poly16x4_t __ret;
   34865   __ret = (poly16x4_t)(__p0);
   34866   return __ret;
   34867 }
   34868 #else
   34869 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
   34870   poly16x4_t __ret;
   34871   __ret = (poly16x4_t)(__p0);
   34872   return __ret;
   34873 }
   34874 #endif
   34875 
   34876 #ifdef __LITTLE_ENDIAN__
   34877 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
   34878   poly16x4_t __ret;
   34879   __ret = (poly16x4_t)(__p0);
   34880   return __ret;
   34881 }
   34882 #else
   34883 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
   34884   poly16x4_t __ret;
   34885   __ret = (poly16x4_t)(__p0);
   34886   return __ret;
   34887 }
   34888 #endif
   34889 
   34890 #ifdef __LITTLE_ENDIAN__
   34891 __ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
   34892   poly8x16_t __ret;
   34893   __ret = (poly8x16_t)(__p0);
   34894   return __ret;
   34895 }
   34896 #else
   34897 __ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
   34898   poly8x16_t __ret;
   34899   __ret = (poly8x16_t)(__p0);
   34900   return __ret;
   34901 }
   34902 #endif
   34903 
   34904 #ifdef __LITTLE_ENDIAN__
   34905 __ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
   34906   poly8x16_t __ret;
   34907   __ret = (poly8x16_t)(__p0);
   34908   return __ret;
   34909 }
   34910 #else
   34911 __ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
   34912   poly8x16_t __ret;
   34913   __ret = (poly8x16_t)(__p0);
   34914   return __ret;
   34915 }
   34916 #endif
   34917 
   34918 #ifdef __LITTLE_ENDIAN__
   34919 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
   34920   poly8x16_t __ret;
   34921   __ret = (poly8x16_t)(__p0);
   34922   return __ret;
   34923 }
   34924 #else
   34925 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
   34926   poly8x16_t __ret;
   34927   __ret = (poly8x16_t)(__p0);
   34928   return __ret;
   34929 }
   34930 #endif
   34931 
   34932 #ifdef __LITTLE_ENDIAN__
   34933 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
   34934   poly8x16_t __ret;
   34935   __ret = (poly8x16_t)(__p0);
   34936   return __ret;
   34937 }
   34938 #else
   34939 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
   34940   poly8x16_t __ret;
   34941   __ret = (poly8x16_t)(__p0);
   34942   return __ret;
   34943 }
   34944 #endif
   34945 
   34946 #ifdef __LITTLE_ENDIAN__
   34947 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
   34948   poly8x16_t __ret;
   34949   __ret = (poly8x16_t)(__p0);
   34950   return __ret;
   34951 }
   34952 #else
   34953 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
   34954   poly8x16_t __ret;
   34955   __ret = (poly8x16_t)(__p0);
   34956   return __ret;
   34957 }
   34958 #endif
   34959 
   34960 #ifdef __LITTLE_ENDIAN__
   34961 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
   34962   poly8x16_t __ret;
   34963   __ret = (poly8x16_t)(__p0);
   34964   return __ret;
   34965 }
   34966 #else
   34967 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
   34968   poly8x16_t __ret;
   34969   __ret = (poly8x16_t)(__p0);
   34970   return __ret;
   34971 }
   34972 #endif
   34973 
   34974 #ifdef __LITTLE_ENDIAN__
   34975 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
   34976   poly8x16_t __ret;
   34977   __ret = (poly8x16_t)(__p0);
   34978   return __ret;
   34979 }
   34980 #else
   34981 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
   34982   poly8x16_t __ret;
   34983   __ret = (poly8x16_t)(__p0);
   34984   return __ret;
   34985 }
   34986 #endif
   34987 
   34988 #ifdef __LITTLE_ENDIAN__
   34989 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
   34990   poly8x16_t __ret;
   34991   __ret = (poly8x16_t)(__p0);
   34992   return __ret;
   34993 }
   34994 #else
   34995 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
   34996   poly8x16_t __ret;
   34997   __ret = (poly8x16_t)(__p0);
   34998   return __ret;
   34999 }
   35000 #endif
   35001 
   35002 #ifdef __LITTLE_ENDIAN__
   35003 __ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
   35004   poly8x16_t __ret;
   35005   __ret = (poly8x16_t)(__p0);
   35006   return __ret;
   35007 }
   35008 #else
   35009 __ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
   35010   poly8x16_t __ret;
   35011   __ret = (poly8x16_t)(__p0);
   35012   return __ret;
   35013 }
   35014 #endif
   35015 
   35016 #ifdef __LITTLE_ENDIAN__
   35017 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
   35018   poly8x16_t __ret;
   35019   __ret = (poly8x16_t)(__p0);
   35020   return __ret;
   35021 }
   35022 #else
   35023 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
   35024   poly8x16_t __ret;
   35025   __ret = (poly8x16_t)(__p0);
   35026   return __ret;
   35027 }
   35028 #endif
   35029 
   35030 #ifdef __LITTLE_ENDIAN__
   35031 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
   35032   poly8x16_t __ret;
   35033   __ret = (poly8x16_t)(__p0);
   35034   return __ret;
   35035 }
   35036 #else
   35037 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
   35038   poly8x16_t __ret;
   35039   __ret = (poly8x16_t)(__p0);
   35040   return __ret;
   35041 }
   35042 #endif
   35043 
   35044 #ifdef __LITTLE_ENDIAN__
   35045 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
   35046   poly8x16_t __ret;
   35047   __ret = (poly8x16_t)(__p0);
   35048   return __ret;
   35049 }
   35050 #else
   35051 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
   35052   poly8x16_t __ret;
   35053   __ret = (poly8x16_t)(__p0);
   35054   return __ret;
   35055 }
   35056 #endif
   35057 
   35058 #ifdef __LITTLE_ENDIAN__
   35059 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
   35060   poly8x16_t __ret;
   35061   __ret = (poly8x16_t)(__p0);
   35062   return __ret;
   35063 }
   35064 #else
   35065 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
   35066   poly8x16_t __ret;
   35067   __ret = (poly8x16_t)(__p0);
   35068   return __ret;
   35069 }
   35070 #endif
   35071 
   35072 #ifdef __LITTLE_ENDIAN__
   35073 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
   35074   poly8x16_t __ret;
   35075   __ret = (poly8x16_t)(__p0);
   35076   return __ret;
   35077 }
   35078 #else
   35079 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
   35080   poly8x16_t __ret;
   35081   __ret = (poly8x16_t)(__p0);
   35082   return __ret;
   35083 }
   35084 #endif
   35085 
   35086 #ifdef __LITTLE_ENDIAN__
   35087 __ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
   35088   poly128_t __ret;
   35089   __ret = (poly128_t)(__p0);
   35090   return __ret;
   35091 }
   35092 #else
   35093 __ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
   35094   poly128_t __ret;
   35095   __ret = (poly128_t)(__p0);
   35096   return __ret;
   35097 }
   35098 #endif
   35099 
   35100 #ifdef __LITTLE_ENDIAN__
   35101 __ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
   35102   poly128_t __ret;
   35103   __ret = (poly128_t)(__p0);
   35104   return __ret;
   35105 }
   35106 #else
   35107 __ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
   35108   poly128_t __ret;
   35109   __ret = (poly128_t)(__p0);
   35110   return __ret;
   35111 }
   35112 #endif
   35113 
   35114 #ifdef __LITTLE_ENDIAN__
   35115 __ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
   35116   poly128_t __ret;
   35117   __ret = (poly128_t)(__p0);
   35118   return __ret;
   35119 }
   35120 #else
   35121 __ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
   35122   poly128_t __ret;
   35123   __ret = (poly128_t)(__p0);
   35124   return __ret;
   35125 }
   35126 #endif
   35127 
   35128 #ifdef __LITTLE_ENDIAN__
   35129 __ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
   35130   poly128_t __ret;
   35131   __ret = (poly128_t)(__p0);
   35132   return __ret;
   35133 }
   35134 #else
   35135 __ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
   35136   poly128_t __ret;
   35137   __ret = (poly128_t)(__p0);
   35138   return __ret;
   35139 }
   35140 #endif
   35141 
   35142 #ifdef __LITTLE_ENDIAN__
   35143 __ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
   35144   poly128_t __ret;
   35145   __ret = (poly128_t)(__p0);
   35146   return __ret;
   35147 }
   35148 #else
   35149 __ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
   35150   poly128_t __ret;
   35151   __ret = (poly128_t)(__p0);
   35152   return __ret;
   35153 }
   35154 #endif
   35155 
   35156 #ifdef __LITTLE_ENDIAN__
   35157 __ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
   35158   poly128_t __ret;
   35159   __ret = (poly128_t)(__p0);
   35160   return __ret;
   35161 }
   35162 #else
   35163 __ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
   35164   poly128_t __ret;
   35165   __ret = (poly128_t)(__p0);
   35166   return __ret;
   35167 }
   35168 #endif
   35169 
   35170 #ifdef __LITTLE_ENDIAN__
   35171 __ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
   35172   poly128_t __ret;
   35173   __ret = (poly128_t)(__p0);
   35174   return __ret;
   35175 }
   35176 #else
   35177 __ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
   35178   poly128_t __ret;
   35179   __ret = (poly128_t)(__p0);
   35180   return __ret;
   35181 }
   35182 #endif
   35183 
   35184 #ifdef __LITTLE_ENDIAN__
   35185 __ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
   35186   poly128_t __ret;
   35187   __ret = (poly128_t)(__p0);
   35188   return __ret;
   35189 }
   35190 #else
   35191 __ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
   35192   poly128_t __ret;
   35193   __ret = (poly128_t)(__p0);
   35194   return __ret;
   35195 }
   35196 #endif
   35197 
   35198 #ifdef __LITTLE_ENDIAN__
   35199 __ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
   35200   poly128_t __ret;
   35201   __ret = (poly128_t)(__p0);
   35202   return __ret;
   35203 }
   35204 #else
   35205 __ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
   35206   poly128_t __ret;
   35207   __ret = (poly128_t)(__p0);
   35208   return __ret;
   35209 }
   35210 #endif
   35211 
   35212 #ifdef __LITTLE_ENDIAN__
   35213 __ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
   35214   poly128_t __ret;
   35215   __ret = (poly128_t)(__p0);
   35216   return __ret;
   35217 }
   35218 #else
   35219 __ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
   35220   poly128_t __ret;
   35221   __ret = (poly128_t)(__p0);
   35222   return __ret;
   35223 }
   35224 #endif
   35225 
   35226 #ifdef __LITTLE_ENDIAN__
   35227 __ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
   35228   poly128_t __ret;
   35229   __ret = (poly128_t)(__p0);
   35230   return __ret;
   35231 }
   35232 #else
   35233 __ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
   35234   poly128_t __ret;
   35235   __ret = (poly128_t)(__p0);
   35236   return __ret;
   35237 }
   35238 #endif
   35239 
   35240 #ifdef __LITTLE_ENDIAN__
   35241 __ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
   35242   poly128_t __ret;
   35243   __ret = (poly128_t)(__p0);
   35244   return __ret;
   35245 }
   35246 #else
   35247 __ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
   35248   poly128_t __ret;
   35249   __ret = (poly128_t)(__p0);
   35250   return __ret;
   35251 }
   35252 #endif
   35253 
   35254 #ifdef __LITTLE_ENDIAN__
   35255 __ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
   35256   poly128_t __ret;
   35257   __ret = (poly128_t)(__p0);
   35258   return __ret;
   35259 }
   35260 #else
   35261 __ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
   35262   poly128_t __ret;
   35263   __ret = (poly128_t)(__p0);
   35264   return __ret;
   35265 }
   35266 #endif
   35267 
   35268 #ifdef __LITTLE_ENDIAN__
   35269 __ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
   35270   poly128_t __ret;
   35271   __ret = (poly128_t)(__p0);
   35272   return __ret;
   35273 }
   35274 #else
   35275 __ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
   35276   poly128_t __ret;
   35277   __ret = (poly128_t)(__p0);
   35278   return __ret;
   35279 }
   35280 #endif
   35281 
   35282 #ifdef __LITTLE_ENDIAN__
   35283 __ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
   35284   poly64x2_t __ret;
   35285   __ret = (poly64x2_t)(__p0);
   35286   return __ret;
   35287 }
   35288 #else
   35289 __ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
   35290   poly64x2_t __ret;
   35291   __ret = (poly64x2_t)(__p0);
   35292   return __ret;
   35293 }
   35294 #endif
   35295 
   35296 #ifdef __LITTLE_ENDIAN__
   35297 __ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
   35298   poly64x2_t __ret;
   35299   __ret = (poly64x2_t)(__p0);
   35300   return __ret;
   35301 }
   35302 #else
   35303 __ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
   35304   poly64x2_t __ret;
   35305   __ret = (poly64x2_t)(__p0);
   35306   return __ret;
   35307 }
   35308 #endif
   35309 
   35310 #ifdef __LITTLE_ENDIAN__
   35311 __ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
   35312   poly64x2_t __ret;
   35313   __ret = (poly64x2_t)(__p0);
   35314   return __ret;
   35315 }
   35316 #else
   35317 __ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
   35318   poly64x2_t __ret;
   35319   __ret = (poly64x2_t)(__p0);
   35320   return __ret;
   35321 }
   35322 #endif
   35323 
   35324 #ifdef __LITTLE_ENDIAN__
   35325 __ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
   35326   poly64x2_t __ret;
   35327   __ret = (poly64x2_t)(__p0);
   35328   return __ret;
   35329 }
   35330 #else
   35331 __ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
   35332   poly64x2_t __ret;
   35333   __ret = (poly64x2_t)(__p0);
   35334   return __ret;
   35335 }
   35336 #endif
   35337 
   35338 #ifdef __LITTLE_ENDIAN__
   35339 __ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
   35340   poly64x2_t __ret;
   35341   __ret = (poly64x2_t)(__p0);
   35342   return __ret;
   35343 }
   35344 #else
   35345 __ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
   35346   poly64x2_t __ret;
   35347   __ret = (poly64x2_t)(__p0);
   35348   return __ret;
   35349 }
   35350 #endif
   35351 
   35352 #ifdef __LITTLE_ENDIAN__
   35353 __ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
   35354   poly64x2_t __ret;
   35355   __ret = (poly64x2_t)(__p0);
   35356   return __ret;
   35357 }
   35358 #else
   35359 __ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
   35360   poly64x2_t __ret;
   35361   __ret = (poly64x2_t)(__p0);
   35362   return __ret;
   35363 }
   35364 #endif
   35365 
   35366 #ifdef __LITTLE_ENDIAN__
   35367 __ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
   35368   poly64x2_t __ret;
   35369   __ret = (poly64x2_t)(__p0);
   35370   return __ret;
   35371 }
   35372 #else
   35373 __ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
   35374   poly64x2_t __ret;
   35375   __ret = (poly64x2_t)(__p0);
   35376   return __ret;
   35377 }
   35378 #endif
   35379 
   35380 #ifdef __LITTLE_ENDIAN__
   35381 __ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
   35382   poly64x2_t __ret;
   35383   __ret = (poly64x2_t)(__p0);
   35384   return __ret;
   35385 }
   35386 #else
   35387 __ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
   35388   poly64x2_t __ret;
   35389   __ret = (poly64x2_t)(__p0);
   35390   return __ret;
   35391 }
   35392 #endif
   35393 
   35394 #ifdef __LITTLE_ENDIAN__
   35395 __ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
   35396   poly64x2_t __ret;
   35397   __ret = (poly64x2_t)(__p0);
   35398   return __ret;
   35399 }
   35400 #else
   35401 __ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
   35402   poly64x2_t __ret;
   35403   __ret = (poly64x2_t)(__p0);
   35404   return __ret;
   35405 }
   35406 #endif
   35407 
   35408 #ifdef __LITTLE_ENDIAN__
   35409 __ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
   35410   poly64x2_t __ret;
   35411   __ret = (poly64x2_t)(__p0);
   35412   return __ret;
   35413 }
   35414 #else
   35415 __ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
   35416   poly64x2_t __ret;
   35417   __ret = (poly64x2_t)(__p0);
   35418   return __ret;
   35419 }
   35420 #endif
   35421 
   35422 #ifdef __LITTLE_ENDIAN__
   35423 __ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
   35424   poly64x2_t __ret;
   35425   __ret = (poly64x2_t)(__p0);
   35426   return __ret;
   35427 }
   35428 #else
   35429 __ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
   35430   poly64x2_t __ret;
   35431   __ret = (poly64x2_t)(__p0);
   35432   return __ret;
   35433 }
   35434 #endif
   35435 
   35436 #ifdef __LITTLE_ENDIAN__
   35437 __ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
   35438   poly64x2_t __ret;
   35439   __ret = (poly64x2_t)(__p0);
   35440   return __ret;
   35441 }
   35442 #else
   35443 __ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
   35444   poly64x2_t __ret;
   35445   __ret = (poly64x2_t)(__p0);
   35446   return __ret;
   35447 }
   35448 #endif
   35449 
   35450 #ifdef __LITTLE_ENDIAN__
   35451 __ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
   35452   poly64x2_t __ret;
   35453   __ret = (poly64x2_t)(__p0);
   35454   return __ret;
   35455 }
   35456 #else
   35457 __ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
   35458   poly64x2_t __ret;
   35459   __ret = (poly64x2_t)(__p0);
   35460   return __ret;
   35461 }
   35462 #endif
   35463 
   35464 #ifdef __LITTLE_ENDIAN__
   35465 __ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
   35466   poly64x2_t __ret;
   35467   __ret = (poly64x2_t)(__p0);
   35468   return __ret;
   35469 }
   35470 #else
   35471 __ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
   35472   poly64x2_t __ret;
   35473   __ret = (poly64x2_t)(__p0);
   35474   return __ret;
   35475 }
   35476 #endif
   35477 
   35478 #ifdef __LITTLE_ENDIAN__
   35479 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
   35480   poly16x8_t __ret;
   35481   __ret = (poly16x8_t)(__p0);
   35482   return __ret;
   35483 }
   35484 #else
   35485 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
   35486   poly16x8_t __ret;
   35487   __ret = (poly16x8_t)(__p0);
   35488   return __ret;
   35489 }
   35490 #endif
   35491 
   35492 #ifdef __LITTLE_ENDIAN__
   35493 __ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
   35494   poly16x8_t __ret;
   35495   __ret = (poly16x8_t)(__p0);
   35496   return __ret;
   35497 }
   35498 #else
   35499 __ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
   35500   poly16x8_t __ret;
   35501   __ret = (poly16x8_t)(__p0);
   35502   return __ret;
   35503 }
   35504 #endif
   35505 
   35506 #ifdef __LITTLE_ENDIAN__
   35507 __ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
   35508   poly16x8_t __ret;
   35509   __ret = (poly16x8_t)(__p0);
   35510   return __ret;
   35511 }
   35512 #else
   35513 __ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
   35514   poly16x8_t __ret;
   35515   __ret = (poly16x8_t)(__p0);
   35516   return __ret;
   35517 }
   35518 #endif
   35519 
   35520 #ifdef __LITTLE_ENDIAN__
   35521 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
   35522   poly16x8_t __ret;
   35523   __ret = (poly16x8_t)(__p0);
   35524   return __ret;
   35525 }
   35526 #else
   35527 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
   35528   poly16x8_t __ret;
   35529   __ret = (poly16x8_t)(__p0);
   35530   return __ret;
   35531 }
   35532 #endif
   35533 
   35534 #ifdef __LITTLE_ENDIAN__
   35535 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
   35536   poly16x8_t __ret;
   35537   __ret = (poly16x8_t)(__p0);
   35538   return __ret;
   35539 }
   35540 #else
   35541 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
   35542   poly16x8_t __ret;
   35543   __ret = (poly16x8_t)(__p0);
   35544   return __ret;
   35545 }
   35546 #endif
   35547 
   35548 #ifdef __LITTLE_ENDIAN__
   35549 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
   35550   poly16x8_t __ret;
   35551   __ret = (poly16x8_t)(__p0);
   35552   return __ret;
   35553 }
   35554 #else
   35555 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
   35556   poly16x8_t __ret;
   35557   __ret = (poly16x8_t)(__p0);
   35558   return __ret;
   35559 }
   35560 #endif
   35561 
   35562 #ifdef __LITTLE_ENDIAN__
   35563 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
   35564   poly16x8_t __ret;
   35565   __ret = (poly16x8_t)(__p0);
   35566   return __ret;
   35567 }
   35568 #else
   35569 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
   35570   poly16x8_t __ret;
   35571   __ret = (poly16x8_t)(__p0);
   35572   return __ret;
   35573 }
   35574 #endif
   35575 
   35576 #ifdef __LITTLE_ENDIAN__
   35577 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
   35578   poly16x8_t __ret;
   35579   __ret = (poly16x8_t)(__p0);
   35580   return __ret;
   35581 }
   35582 #else
   35583 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
   35584   poly16x8_t __ret;
   35585   __ret = (poly16x8_t)(__p0);
   35586   return __ret;
   35587 }
   35588 #endif
   35589 
   35590 #ifdef __LITTLE_ENDIAN__
   35591 __ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
   35592   poly16x8_t __ret;
   35593   __ret = (poly16x8_t)(__p0);
   35594   return __ret;
   35595 }
   35596 #else
   35597 __ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
   35598   poly16x8_t __ret;
   35599   __ret = (poly16x8_t)(__p0);
   35600   return __ret;
   35601 }
   35602 #endif
   35603 
   35604 #ifdef __LITTLE_ENDIAN__
   35605 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
   35606   poly16x8_t __ret;
   35607   __ret = (poly16x8_t)(__p0);
   35608   return __ret;
   35609 }
   35610 #else
   35611 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
   35612   poly16x8_t __ret;
   35613   __ret = (poly16x8_t)(__p0);
   35614   return __ret;
   35615 }
   35616 #endif
   35617 
   35618 #ifdef __LITTLE_ENDIAN__
   35619 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
   35620   poly16x8_t __ret;
   35621   __ret = (poly16x8_t)(__p0);
   35622   return __ret;
   35623 }
   35624 #else
   35625 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
   35626   poly16x8_t __ret;
   35627   __ret = (poly16x8_t)(__p0);
   35628   return __ret;
   35629 }
   35630 #endif
   35631 
   35632 #ifdef __LITTLE_ENDIAN__
   35633 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
   35634   poly16x8_t __ret;
   35635   __ret = (poly16x8_t)(__p0);
   35636   return __ret;
   35637 }
   35638 #else
   35639 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
   35640   poly16x8_t __ret;
   35641   __ret = (poly16x8_t)(__p0);
   35642   return __ret;
   35643 }
   35644 #endif
   35645 
   35646 #ifdef __LITTLE_ENDIAN__
   35647 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
   35648   poly16x8_t __ret;
   35649   __ret = (poly16x8_t)(__p0);
   35650   return __ret;
   35651 }
   35652 #else
   35653 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
   35654   poly16x8_t __ret;
   35655   __ret = (poly16x8_t)(__p0);
   35656   return __ret;
   35657 }
   35658 #endif
   35659 
   35660 #ifdef __LITTLE_ENDIAN__
   35661 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
   35662   poly16x8_t __ret;
   35663   __ret = (poly16x8_t)(__p0);
   35664   return __ret;
   35665 }
   35666 #else
   35667 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
   35668   poly16x8_t __ret;
   35669   __ret = (poly16x8_t)(__p0);
   35670   return __ret;
   35671 }
   35672 #endif
   35673 
   35674 #ifdef __LITTLE_ENDIAN__
   35675 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
   35676   uint8x16_t __ret;
   35677   __ret = (uint8x16_t)(__p0);
   35678   return __ret;
   35679 }
   35680 #else
   35681 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
   35682   uint8x16_t __ret;
   35683   __ret = (uint8x16_t)(__p0);
   35684   return __ret;
   35685 }
   35686 #endif
   35687 
   35688 #ifdef __LITTLE_ENDIAN__
   35689 __ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
   35690   uint8x16_t __ret;
   35691   __ret = (uint8x16_t)(__p0);
   35692   return __ret;
   35693 }
   35694 #else
   35695 __ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
   35696   uint8x16_t __ret;
   35697   __ret = (uint8x16_t)(__p0);
   35698   return __ret;
   35699 }
   35700 #endif
   35701 
   35702 #ifdef __LITTLE_ENDIAN__
   35703 __ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
   35704   uint8x16_t __ret;
   35705   __ret = (uint8x16_t)(__p0);
   35706   return __ret;
   35707 }
   35708 #else
   35709 __ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
   35710   uint8x16_t __ret;
   35711   __ret = (uint8x16_t)(__p0);
   35712   return __ret;
   35713 }
   35714 #endif
   35715 
   35716 #ifdef __LITTLE_ENDIAN__
   35717 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
   35718   uint8x16_t __ret;
   35719   __ret = (uint8x16_t)(__p0);
   35720   return __ret;
   35721 }
   35722 #else
   35723 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
   35724   uint8x16_t __ret;
   35725   __ret = (uint8x16_t)(__p0);
   35726   return __ret;
   35727 }
   35728 #endif
   35729 
   35730 #ifdef __LITTLE_ENDIAN__
   35731 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
   35732   uint8x16_t __ret;
   35733   __ret = (uint8x16_t)(__p0);
   35734   return __ret;
   35735 }
   35736 #else
   35737 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
   35738   uint8x16_t __ret;
   35739   __ret = (uint8x16_t)(__p0);
   35740   return __ret;
   35741 }
   35742 #endif
   35743 
   35744 #ifdef __LITTLE_ENDIAN__
   35745 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
   35746   uint8x16_t __ret;
   35747   __ret = (uint8x16_t)(__p0);
   35748   return __ret;
   35749 }
   35750 #else
   35751 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
   35752   uint8x16_t __ret;
   35753   __ret = (uint8x16_t)(__p0);
   35754   return __ret;
   35755 }
   35756 #endif
   35757 
   35758 #ifdef __LITTLE_ENDIAN__
   35759 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
   35760   uint8x16_t __ret;
   35761   __ret = (uint8x16_t)(__p0);
   35762   return __ret;
   35763 }
   35764 #else
   35765 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
   35766   uint8x16_t __ret;
   35767   __ret = (uint8x16_t)(__p0);
   35768   return __ret;
   35769 }
   35770 #endif
   35771 
   35772 #ifdef __LITTLE_ENDIAN__
   35773 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
   35774   uint8x16_t __ret;
   35775   __ret = (uint8x16_t)(__p0);
   35776   return __ret;
   35777 }
   35778 #else
   35779 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
   35780   uint8x16_t __ret;
   35781   __ret = (uint8x16_t)(__p0);
   35782   return __ret;
   35783 }
   35784 #endif
   35785 
   35786 #ifdef __LITTLE_ENDIAN__
   35787 __ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
   35788   uint8x16_t __ret;
   35789   __ret = (uint8x16_t)(__p0);
   35790   return __ret;
   35791 }
   35792 #else
   35793 __ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
   35794   uint8x16_t __ret;
   35795   __ret = (uint8x16_t)(__p0);
   35796   return __ret;
   35797 }
   35798 #endif
   35799 
   35800 #ifdef __LITTLE_ENDIAN__
   35801 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
   35802   uint8x16_t __ret;
   35803   __ret = (uint8x16_t)(__p0);
   35804   return __ret;
   35805 }
   35806 #else
   35807 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
   35808   uint8x16_t __ret;
   35809   __ret = (uint8x16_t)(__p0);
   35810   return __ret;
   35811 }
   35812 #endif
   35813 
   35814 #ifdef __LITTLE_ENDIAN__
   35815 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
   35816   uint8x16_t __ret;
   35817   __ret = (uint8x16_t)(__p0);
   35818   return __ret;
   35819 }
   35820 #else
   35821 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
   35822   uint8x16_t __ret;
   35823   __ret = (uint8x16_t)(__p0);
   35824   return __ret;
   35825 }
   35826 #endif
   35827 
   35828 #ifdef __LITTLE_ENDIAN__
   35829 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
   35830   uint8x16_t __ret;
   35831   __ret = (uint8x16_t)(__p0);
   35832   return __ret;
   35833 }
   35834 #else
   35835 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
   35836   uint8x16_t __ret;
   35837   __ret = (uint8x16_t)(__p0);
   35838   return __ret;
   35839 }
   35840 #endif
   35841 
   35842 #ifdef __LITTLE_ENDIAN__
   35843 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
   35844   uint8x16_t __ret;
   35845   __ret = (uint8x16_t)(__p0);
   35846   return __ret;
   35847 }
   35848 #else
   35849 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
   35850   uint8x16_t __ret;
   35851   __ret = (uint8x16_t)(__p0);
   35852   return __ret;
   35853 }
   35854 #endif
   35855 
   35856 #ifdef __LITTLE_ENDIAN__
   35857 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
   35858   uint8x16_t __ret;
   35859   __ret = (uint8x16_t)(__p0);
   35860   return __ret;
   35861 }
   35862 #else
   35863 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
   35864   uint8x16_t __ret;
   35865   __ret = (uint8x16_t)(__p0);
   35866   return __ret;
   35867 }
   35868 #endif
   35869 
   35870 #ifdef __LITTLE_ENDIAN__
   35871 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
   35872   uint32x4_t __ret;
   35873   __ret = (uint32x4_t)(__p0);
   35874   return __ret;
   35875 }
   35876 #else
   35877 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
   35878   uint32x4_t __ret;
   35879   __ret = (uint32x4_t)(__p0);
   35880   return __ret;
   35881 }
   35882 #endif
   35883 
   35884 #ifdef __LITTLE_ENDIAN__
   35885 __ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
   35886   uint32x4_t __ret;
   35887   __ret = (uint32x4_t)(__p0);
   35888   return __ret;
   35889 }
   35890 #else
   35891 __ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
   35892   uint32x4_t __ret;
   35893   __ret = (uint32x4_t)(__p0);
   35894   return __ret;
   35895 }
   35896 #endif
   35897 
   35898 #ifdef __LITTLE_ENDIAN__
   35899 __ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
   35900   uint32x4_t __ret;
   35901   __ret = (uint32x4_t)(__p0);
   35902   return __ret;
   35903 }
   35904 #else
   35905 __ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
   35906   uint32x4_t __ret;
   35907   __ret = (uint32x4_t)(__p0);
   35908   return __ret;
   35909 }
   35910 #endif
   35911 
   35912 #ifdef __LITTLE_ENDIAN__
   35913 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
   35914   uint32x4_t __ret;
   35915   __ret = (uint32x4_t)(__p0);
   35916   return __ret;
   35917 }
   35918 #else
   35919 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
   35920   uint32x4_t __ret;
   35921   __ret = (uint32x4_t)(__p0);
   35922   return __ret;
   35923 }
   35924 #endif
   35925 
   35926 #ifdef __LITTLE_ENDIAN__
   35927 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
   35928   uint32x4_t __ret;
   35929   __ret = (uint32x4_t)(__p0);
   35930   return __ret;
   35931 }
   35932 #else
   35933 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
   35934   uint32x4_t __ret;
   35935   __ret = (uint32x4_t)(__p0);
   35936   return __ret;
   35937 }
   35938 #endif
   35939 
   35940 #ifdef __LITTLE_ENDIAN__
   35941 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
   35942   uint32x4_t __ret;
   35943   __ret = (uint32x4_t)(__p0);
   35944   return __ret;
   35945 }
   35946 #else
   35947 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
   35948   uint32x4_t __ret;
   35949   __ret = (uint32x4_t)(__p0);
   35950   return __ret;
   35951 }
   35952 #endif
   35953 
   35954 #ifdef __LITTLE_ENDIAN__
   35955 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
   35956   uint32x4_t __ret;
   35957   __ret = (uint32x4_t)(__p0);
   35958   return __ret;
   35959 }
   35960 #else
   35961 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
   35962   uint32x4_t __ret;
   35963   __ret = (uint32x4_t)(__p0);
   35964   return __ret;
   35965 }
   35966 #endif
   35967 
   35968 #ifdef __LITTLE_ENDIAN__
   35969 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
   35970   uint32x4_t __ret;
   35971   __ret = (uint32x4_t)(__p0);
   35972   return __ret;
   35973 }
   35974 #else
   35975 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
   35976   uint32x4_t __ret;
   35977   __ret = (uint32x4_t)(__p0);
   35978   return __ret;
   35979 }
   35980 #endif
   35981 
   35982 #ifdef __LITTLE_ENDIAN__
   35983 __ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
   35984   uint32x4_t __ret;
   35985   __ret = (uint32x4_t)(__p0);
   35986   return __ret;
   35987 }
   35988 #else
   35989 __ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
   35990   uint32x4_t __ret;
   35991   __ret = (uint32x4_t)(__p0);
   35992   return __ret;
   35993 }
   35994 #endif
   35995 
   35996 #ifdef __LITTLE_ENDIAN__
   35997 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
   35998   uint32x4_t __ret;
   35999   __ret = (uint32x4_t)(__p0);
   36000   return __ret;
   36001 }
   36002 #else
   36003 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
   36004   uint32x4_t __ret;
   36005   __ret = (uint32x4_t)(__p0);
   36006   return __ret;
   36007 }
   36008 #endif
   36009 
   36010 #ifdef __LITTLE_ENDIAN__
   36011 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
   36012   uint32x4_t __ret;
   36013   __ret = (uint32x4_t)(__p0);
   36014   return __ret;
   36015 }
   36016 #else
   36017 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
   36018   uint32x4_t __ret;
   36019   __ret = (uint32x4_t)(__p0);
   36020   return __ret;
   36021 }
   36022 #endif
   36023 
   36024 #ifdef __LITTLE_ENDIAN__
   36025 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
   36026   uint32x4_t __ret;
   36027   __ret = (uint32x4_t)(__p0);
   36028   return __ret;
   36029 }
   36030 #else
   36031 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
   36032   uint32x4_t __ret;
   36033   __ret = (uint32x4_t)(__p0);
   36034   return __ret;
   36035 }
   36036 #endif
   36037 
   36038 #ifdef __LITTLE_ENDIAN__
   36039 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
   36040   uint32x4_t __ret;
   36041   __ret = (uint32x4_t)(__p0);
   36042   return __ret;
   36043 }
   36044 #else
   36045 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
   36046   uint32x4_t __ret;
   36047   __ret = (uint32x4_t)(__p0);
   36048   return __ret;
   36049 }
   36050 #endif
   36051 
   36052 #ifdef __LITTLE_ENDIAN__
   36053 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
   36054   uint32x4_t __ret;
   36055   __ret = (uint32x4_t)(__p0);
   36056   return __ret;
   36057 }
   36058 #else
   36059 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
   36060   uint32x4_t __ret;
   36061   __ret = (uint32x4_t)(__p0);
   36062   return __ret;
   36063 }
   36064 #endif
   36065 
   36066 #ifdef __LITTLE_ENDIAN__
   36067 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
   36068   uint64x2_t __ret;
   36069   __ret = (uint64x2_t)(__p0);
   36070   return __ret;
   36071 }
   36072 #else
   36073 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
   36074   uint64x2_t __ret;
   36075   __ret = (uint64x2_t)(__p0);
   36076   return __ret;
   36077 }
   36078 #endif
   36079 
   36080 #ifdef __LITTLE_ENDIAN__
   36081 __ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
   36082   uint64x2_t __ret;
   36083   __ret = (uint64x2_t)(__p0);
   36084   return __ret;
   36085 }
   36086 #else
   36087 __ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
   36088   uint64x2_t __ret;
   36089   __ret = (uint64x2_t)(__p0);
   36090   return __ret;
   36091 }
   36092 #endif
   36093 
   36094 #ifdef __LITTLE_ENDIAN__
   36095 __ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
   36096   uint64x2_t __ret;
   36097   __ret = (uint64x2_t)(__p0);
   36098   return __ret;
   36099 }
   36100 #else
   36101 __ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
   36102   uint64x2_t __ret;
   36103   __ret = (uint64x2_t)(__p0);
   36104   return __ret;
   36105 }
   36106 #endif
   36107 
   36108 #ifdef __LITTLE_ENDIAN__
   36109 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
   36110   uint64x2_t __ret;
   36111   __ret = (uint64x2_t)(__p0);
   36112   return __ret;
   36113 }
   36114 #else
   36115 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
   36116   uint64x2_t __ret;
   36117   __ret = (uint64x2_t)(__p0);
   36118   return __ret;
   36119 }
   36120 #endif
   36121 
   36122 #ifdef __LITTLE_ENDIAN__
   36123 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
   36124   uint64x2_t __ret;
   36125   __ret = (uint64x2_t)(__p0);
   36126   return __ret;
   36127 }
   36128 #else
   36129 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
   36130   uint64x2_t __ret;
   36131   __ret = (uint64x2_t)(__p0);
   36132   return __ret;
   36133 }
   36134 #endif
   36135 
   36136 #ifdef __LITTLE_ENDIAN__
   36137 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
   36138   uint64x2_t __ret;
   36139   __ret = (uint64x2_t)(__p0);
   36140   return __ret;
   36141 }
   36142 #else
   36143 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
   36144   uint64x2_t __ret;
   36145   __ret = (uint64x2_t)(__p0);
   36146   return __ret;
   36147 }
   36148 #endif
   36149 
   36150 #ifdef __LITTLE_ENDIAN__
   36151 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
   36152   uint64x2_t __ret;
   36153   __ret = (uint64x2_t)(__p0);
   36154   return __ret;
   36155 }
   36156 #else
   36157 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
   36158   uint64x2_t __ret;
   36159   __ret = (uint64x2_t)(__p0);
   36160   return __ret;
   36161 }
   36162 #endif
   36163 
   36164 #ifdef __LITTLE_ENDIAN__
   36165 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
   36166   uint64x2_t __ret;
   36167   __ret = (uint64x2_t)(__p0);
   36168   return __ret;
   36169 }
   36170 #else
   36171 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
   36172   uint64x2_t __ret;
   36173   __ret = (uint64x2_t)(__p0);
   36174   return __ret;
   36175 }
   36176 #endif
   36177 
   36178 #ifdef __LITTLE_ENDIAN__
   36179 __ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
   36180   uint64x2_t __ret;
   36181   __ret = (uint64x2_t)(__p0);
   36182   return __ret;
   36183 }
   36184 #else
   36185 __ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
   36186   uint64x2_t __ret;
   36187   __ret = (uint64x2_t)(__p0);
   36188   return __ret;
   36189 }
   36190 #endif
   36191 
   36192 #ifdef __LITTLE_ENDIAN__
   36193 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
   36194   uint64x2_t __ret;
   36195   __ret = (uint64x2_t)(__p0);
   36196   return __ret;
   36197 }
   36198 #else
   36199 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
   36200   uint64x2_t __ret;
   36201   __ret = (uint64x2_t)(__p0);
   36202   return __ret;
   36203 }
   36204 #endif
   36205 
   36206 #ifdef __LITTLE_ENDIAN__
   36207 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
   36208   uint64x2_t __ret;
   36209   __ret = (uint64x2_t)(__p0);
   36210   return __ret;
   36211 }
   36212 #else
   36213 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
   36214   uint64x2_t __ret;
   36215   __ret = (uint64x2_t)(__p0);
   36216   return __ret;
   36217 }
   36218 #endif
   36219 
   36220 #ifdef __LITTLE_ENDIAN__
   36221 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
   36222   uint64x2_t __ret;
   36223   __ret = (uint64x2_t)(__p0);
   36224   return __ret;
   36225 }
   36226 #else
   36227 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
   36228   uint64x2_t __ret;
   36229   __ret = (uint64x2_t)(__p0);
   36230   return __ret;
   36231 }
   36232 #endif
   36233 
   36234 #ifdef __LITTLE_ENDIAN__
   36235 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
   36236   uint64x2_t __ret;
   36237   __ret = (uint64x2_t)(__p0);
   36238   return __ret;
   36239 }
   36240 #else
   36241 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
   36242   uint64x2_t __ret;
   36243   __ret = (uint64x2_t)(__p0);
   36244   return __ret;
   36245 }
   36246 #endif
   36247 
   36248 #ifdef __LITTLE_ENDIAN__
   36249 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
   36250   uint64x2_t __ret;
   36251   __ret = (uint64x2_t)(__p0);
   36252   return __ret;
   36253 }
   36254 #else
   36255 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
   36256   uint64x2_t __ret;
   36257   __ret = (uint64x2_t)(__p0);
   36258   return __ret;
   36259 }
   36260 #endif
   36261 
   36262 #ifdef __LITTLE_ENDIAN__
   36263 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
   36264   uint16x8_t __ret;
   36265   __ret = (uint16x8_t)(__p0);
   36266   return __ret;
   36267 }
   36268 #else
   36269 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
   36270   uint16x8_t __ret;
   36271   __ret = (uint16x8_t)(__p0);
   36272   return __ret;
   36273 }
   36274 #endif
   36275 
   36276 #ifdef __LITTLE_ENDIAN__
   36277 __ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
   36278   uint16x8_t __ret;
   36279   __ret = (uint16x8_t)(__p0);
   36280   return __ret;
   36281 }
   36282 #else
   36283 __ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
   36284   uint16x8_t __ret;
   36285   __ret = (uint16x8_t)(__p0);
   36286   return __ret;
   36287 }
   36288 #endif
   36289 
   36290 #ifdef __LITTLE_ENDIAN__
   36291 __ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
   36292   uint16x8_t __ret;
   36293   __ret = (uint16x8_t)(__p0);
   36294   return __ret;
   36295 }
   36296 #else
   36297 __ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
   36298   uint16x8_t __ret;
   36299   __ret = (uint16x8_t)(__p0);
   36300   return __ret;
   36301 }
   36302 #endif
   36303 
   36304 #ifdef __LITTLE_ENDIAN__
   36305 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
   36306   uint16x8_t __ret;
   36307   __ret = (uint16x8_t)(__p0);
   36308   return __ret;
   36309 }
   36310 #else
   36311 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
   36312   uint16x8_t __ret;
   36313   __ret = (uint16x8_t)(__p0);
   36314   return __ret;
   36315 }
   36316 #endif
   36317 
   36318 #ifdef __LITTLE_ENDIAN__
   36319 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
   36320   uint16x8_t __ret;
   36321   __ret = (uint16x8_t)(__p0);
   36322   return __ret;
   36323 }
   36324 #else
   36325 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
   36326   uint16x8_t __ret;
   36327   __ret = (uint16x8_t)(__p0);
   36328   return __ret;
   36329 }
   36330 #endif
   36331 
   36332 #ifdef __LITTLE_ENDIAN__
   36333 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
   36334   uint16x8_t __ret;
   36335   __ret = (uint16x8_t)(__p0);
   36336   return __ret;
   36337 }
   36338 #else
   36339 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
   36340   uint16x8_t __ret;
   36341   __ret = (uint16x8_t)(__p0);
   36342   return __ret;
   36343 }
   36344 #endif
   36345 
   36346 #ifdef __LITTLE_ENDIAN__
   36347 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
   36348   uint16x8_t __ret;
   36349   __ret = (uint16x8_t)(__p0);
   36350   return __ret;
   36351 }
   36352 #else
   36353 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
   36354   uint16x8_t __ret;
   36355   __ret = (uint16x8_t)(__p0);
   36356   return __ret;
   36357 }
   36358 #endif
   36359 
   36360 #ifdef __LITTLE_ENDIAN__
   36361 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
   36362   uint16x8_t __ret;
   36363   __ret = (uint16x8_t)(__p0);
   36364   return __ret;
   36365 }
   36366 #else
   36367 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
   36368   uint16x8_t __ret;
   36369   __ret = (uint16x8_t)(__p0);
   36370   return __ret;
   36371 }
   36372 #endif
   36373 
   36374 #ifdef __LITTLE_ENDIAN__
   36375 __ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
   36376   uint16x8_t __ret;
   36377   __ret = (uint16x8_t)(__p0);
   36378   return __ret;
   36379 }
   36380 #else
   36381 __ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
   36382   uint16x8_t __ret;
   36383   __ret = (uint16x8_t)(__p0);
   36384   return __ret;
   36385 }
   36386 #endif
   36387 
   36388 #ifdef __LITTLE_ENDIAN__
   36389 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
   36390   uint16x8_t __ret;
   36391   __ret = (uint16x8_t)(__p0);
   36392   return __ret;
   36393 }
   36394 #else
   36395 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
   36396   uint16x8_t __ret;
   36397   __ret = (uint16x8_t)(__p0);
   36398   return __ret;
   36399 }
   36400 #endif
   36401 
   36402 #ifdef __LITTLE_ENDIAN__
   36403 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
   36404   uint16x8_t __ret;
   36405   __ret = (uint16x8_t)(__p0);
   36406   return __ret;
   36407 }
   36408 #else
   36409 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
   36410   uint16x8_t __ret;
   36411   __ret = (uint16x8_t)(__p0);
   36412   return __ret;
   36413 }
   36414 #endif
   36415 
   36416 #ifdef __LITTLE_ENDIAN__
   36417 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
   36418   uint16x8_t __ret;
   36419   __ret = (uint16x8_t)(__p0);
   36420   return __ret;
   36421 }
   36422 #else
   36423 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
   36424   uint16x8_t __ret;
   36425   __ret = (uint16x8_t)(__p0);
   36426   return __ret;
   36427 }
   36428 #endif
   36429 
   36430 #ifdef __LITTLE_ENDIAN__
   36431 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
   36432   uint16x8_t __ret;
   36433   __ret = (uint16x8_t)(__p0);
   36434   return __ret;
   36435 }
   36436 #else
   36437 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
   36438   uint16x8_t __ret;
   36439   __ret = (uint16x8_t)(__p0);
   36440   return __ret;
   36441 }
   36442 #endif
   36443 
   36444 #ifdef __LITTLE_ENDIAN__
   36445 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
   36446   uint16x8_t __ret;
   36447   __ret = (uint16x8_t)(__p0);
   36448   return __ret;
   36449 }
   36450 #else
   36451 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
   36452   uint16x8_t __ret;
   36453   __ret = (uint16x8_t)(__p0);
   36454   return __ret;
   36455 }
   36456 #endif
   36457 
   36458 #ifdef __LITTLE_ENDIAN__
   36459 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
   36460   int8x16_t __ret;
   36461   __ret = (int8x16_t)(__p0);
   36462   return __ret;
   36463 }
   36464 #else
   36465 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
   36466   int8x16_t __ret;
   36467   __ret = (int8x16_t)(__p0);
   36468   return __ret;
   36469 }
   36470 #endif
   36471 
   36472 #ifdef __LITTLE_ENDIAN__
   36473 __ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
   36474   int8x16_t __ret;
   36475   __ret = (int8x16_t)(__p0);
   36476   return __ret;
   36477 }
   36478 #else
   36479 __ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
   36480   int8x16_t __ret;
   36481   __ret = (int8x16_t)(__p0);
   36482   return __ret;
   36483 }
   36484 #endif
   36485 
   36486 #ifdef __LITTLE_ENDIAN__
   36487 __ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
   36488   int8x16_t __ret;
   36489   __ret = (int8x16_t)(__p0);
   36490   return __ret;
   36491 }
   36492 #else
   36493 __ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
   36494   int8x16_t __ret;
   36495   __ret = (int8x16_t)(__p0);
   36496   return __ret;
   36497 }
   36498 #endif
   36499 
   36500 #ifdef __LITTLE_ENDIAN__
   36501 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
   36502   int8x16_t __ret;
   36503   __ret = (int8x16_t)(__p0);
   36504   return __ret;
   36505 }
   36506 #else
   36507 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
   36508   int8x16_t __ret;
   36509   __ret = (int8x16_t)(__p0);
   36510   return __ret;
   36511 }
   36512 #endif
   36513 
   36514 #ifdef __LITTLE_ENDIAN__
   36515 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
   36516   int8x16_t __ret;
   36517   __ret = (int8x16_t)(__p0);
   36518   return __ret;
   36519 }
   36520 #else
   36521 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
   36522   int8x16_t __ret;
   36523   __ret = (int8x16_t)(__p0);
   36524   return __ret;
   36525 }
   36526 #endif
   36527 
   36528 #ifdef __LITTLE_ENDIAN__
   36529 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
   36530   int8x16_t __ret;
   36531   __ret = (int8x16_t)(__p0);
   36532   return __ret;
   36533 }
   36534 #else
   36535 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
   36536   int8x16_t __ret;
   36537   __ret = (int8x16_t)(__p0);
   36538   return __ret;
   36539 }
   36540 #endif
   36541 
   36542 #ifdef __LITTLE_ENDIAN__
   36543 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
   36544   int8x16_t __ret;
   36545   __ret = (int8x16_t)(__p0);
   36546   return __ret;
   36547 }
   36548 #else
   36549 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
   36550   int8x16_t __ret;
   36551   __ret = (int8x16_t)(__p0);
   36552   return __ret;
   36553 }
   36554 #endif
   36555 
   36556 #ifdef __LITTLE_ENDIAN__
   36557 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
   36558   int8x16_t __ret;
   36559   __ret = (int8x16_t)(__p0);
   36560   return __ret;
   36561 }
   36562 #else
   36563 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
   36564   int8x16_t __ret;
   36565   __ret = (int8x16_t)(__p0);
   36566   return __ret;
   36567 }
   36568 #endif
   36569 
   36570 #ifdef __LITTLE_ENDIAN__
   36571 __ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
   36572   int8x16_t __ret;
   36573   __ret = (int8x16_t)(__p0);
   36574   return __ret;
   36575 }
   36576 #else
   36577 __ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
   36578   int8x16_t __ret;
   36579   __ret = (int8x16_t)(__p0);
   36580   return __ret;
   36581 }
   36582 #endif
   36583 
   36584 #ifdef __LITTLE_ENDIAN__
   36585 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
   36586   int8x16_t __ret;
   36587   __ret = (int8x16_t)(__p0);
   36588   return __ret;
   36589 }
   36590 #else
   36591 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
   36592   int8x16_t __ret;
   36593   __ret = (int8x16_t)(__p0);
   36594   return __ret;
   36595 }
   36596 #endif
   36597 
   36598 #ifdef __LITTLE_ENDIAN__
   36599 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
   36600   int8x16_t __ret;
   36601   __ret = (int8x16_t)(__p0);
   36602   return __ret;
   36603 }
   36604 #else
   36605 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
   36606   int8x16_t __ret;
   36607   __ret = (int8x16_t)(__p0);
   36608   return __ret;
   36609 }
   36610 #endif
   36611 
   36612 #ifdef __LITTLE_ENDIAN__
   36613 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
   36614   int8x16_t __ret;
   36615   __ret = (int8x16_t)(__p0);
   36616   return __ret;
   36617 }
   36618 #else
   36619 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
   36620   int8x16_t __ret;
   36621   __ret = (int8x16_t)(__p0);
   36622   return __ret;
   36623 }
   36624 #endif
   36625 
   36626 #ifdef __LITTLE_ENDIAN__
   36627 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
   36628   int8x16_t __ret;
   36629   __ret = (int8x16_t)(__p0);
   36630   return __ret;
   36631 }
   36632 #else
   36633 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
   36634   int8x16_t __ret;
   36635   __ret = (int8x16_t)(__p0);
   36636   return __ret;
   36637 }
   36638 #endif
   36639 
   36640 #ifdef __LITTLE_ENDIAN__
   36641 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
   36642   int8x16_t __ret;
   36643   __ret = (int8x16_t)(__p0);
   36644   return __ret;
   36645 }
   36646 #else
   36647 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
   36648   int8x16_t __ret;
   36649   __ret = (int8x16_t)(__p0);
   36650   return __ret;
   36651 }
   36652 #endif
   36653 
   36654 #ifdef __LITTLE_ENDIAN__
   36655 __ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
   36656   float64x2_t __ret;
   36657   __ret = (float64x2_t)(__p0);
   36658   return __ret;
   36659 }
   36660 #else
   36661 __ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
   36662   float64x2_t __ret;
   36663   __ret = (float64x2_t)(__p0);
   36664   return __ret;
   36665 }
   36666 #endif
   36667 
   36668 #ifdef __LITTLE_ENDIAN__
   36669 __ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
   36670   float64x2_t __ret;
   36671   __ret = (float64x2_t)(__p0);
   36672   return __ret;
   36673 }
   36674 #else
   36675 __ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
   36676   float64x2_t __ret;
   36677   __ret = (float64x2_t)(__p0);
   36678   return __ret;
   36679 }
   36680 #endif
   36681 
   36682 #ifdef __LITTLE_ENDIAN__
   36683 __ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
   36684   float64x2_t __ret;
   36685   __ret = (float64x2_t)(__p0);
   36686   return __ret;
   36687 }
   36688 #else
   36689 __ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
   36690   float64x2_t __ret;
   36691   __ret = (float64x2_t)(__p0);
   36692   return __ret;
   36693 }
   36694 #endif
   36695 
   36696 #ifdef __LITTLE_ENDIAN__
   36697 __ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
   36698   float64x2_t __ret;
   36699   __ret = (float64x2_t)(__p0);
   36700   return __ret;
   36701 }
   36702 #else
   36703 __ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
   36704   float64x2_t __ret;
   36705   __ret = (float64x2_t)(__p0);
   36706   return __ret;
   36707 }
   36708 #endif
   36709 
   36710 #ifdef __LITTLE_ENDIAN__
   36711 __ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
   36712   float64x2_t __ret;
   36713   __ret = (float64x2_t)(__p0);
   36714   return __ret;
   36715 }
   36716 #else
   36717 __ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
   36718   float64x2_t __ret;
   36719   __ret = (float64x2_t)(__p0);
   36720   return __ret;
   36721 }
   36722 #endif
   36723 
   36724 #ifdef __LITTLE_ENDIAN__
   36725 __ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
   36726   float64x2_t __ret;
   36727   __ret = (float64x2_t)(__p0);
   36728   return __ret;
   36729 }
   36730 #else
   36731 __ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
   36732   float64x2_t __ret;
   36733   __ret = (float64x2_t)(__p0);
   36734   return __ret;
   36735 }
   36736 #endif
   36737 
   36738 #ifdef __LITTLE_ENDIAN__
   36739 __ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
   36740   float64x2_t __ret;
   36741   __ret = (float64x2_t)(__p0);
   36742   return __ret;
   36743 }
   36744 #else
   36745 __ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
   36746   float64x2_t __ret;
   36747   __ret = (float64x2_t)(__p0);
   36748   return __ret;
   36749 }
   36750 #endif
   36751 
   36752 #ifdef __LITTLE_ENDIAN__
   36753 __ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
   36754   float64x2_t __ret;
   36755   __ret = (float64x2_t)(__p0);
   36756   return __ret;
   36757 }
   36758 #else
   36759 __ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
   36760   float64x2_t __ret;
   36761   __ret = (float64x2_t)(__p0);
   36762   return __ret;
   36763 }
   36764 #endif
   36765 
   36766 #ifdef __LITTLE_ENDIAN__
   36767 __ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
   36768   float64x2_t __ret;
   36769   __ret = (float64x2_t)(__p0);
   36770   return __ret;
   36771 }
   36772 #else
   36773 __ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
   36774   float64x2_t __ret;
   36775   __ret = (float64x2_t)(__p0);
   36776   return __ret;
   36777 }
   36778 #endif
   36779 
   36780 #ifdef __LITTLE_ENDIAN__
   36781 __ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
   36782   float64x2_t __ret;
   36783   __ret = (float64x2_t)(__p0);
   36784   return __ret;
   36785 }
   36786 #else
   36787 __ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
   36788   float64x2_t __ret;
   36789   __ret = (float64x2_t)(__p0);
   36790   return __ret;
   36791 }
   36792 #endif
   36793 
   36794 #ifdef __LITTLE_ENDIAN__
   36795 __ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
   36796   float64x2_t __ret;
   36797   __ret = (float64x2_t)(__p0);
   36798   return __ret;
   36799 }
   36800 #else
   36801 __ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
   36802   float64x2_t __ret;
   36803   __ret = (float64x2_t)(__p0);
   36804   return __ret;
   36805 }
   36806 #endif
   36807 
   36808 #ifdef __LITTLE_ENDIAN__
   36809 __ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
   36810   float64x2_t __ret;
   36811   __ret = (float64x2_t)(__p0);
   36812   return __ret;
   36813 }
   36814 #else
   36815 __ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
   36816   float64x2_t __ret;
   36817   __ret = (float64x2_t)(__p0);
   36818   return __ret;
   36819 }
   36820 #endif
   36821 
   36822 #ifdef __LITTLE_ENDIAN__
   36823 __ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
   36824   float64x2_t __ret;
   36825   __ret = (float64x2_t)(__p0);
   36826   return __ret;
   36827 }
   36828 #else
   36829 __ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
   36830   float64x2_t __ret;
   36831   __ret = (float64x2_t)(__p0);
   36832   return __ret;
   36833 }
   36834 #endif
   36835 
   36836 #ifdef __LITTLE_ENDIAN__
   36837 __ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
   36838   float64x2_t __ret;
   36839   __ret = (float64x2_t)(__p0);
   36840   return __ret;
   36841 }
   36842 #else
   36843 __ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
   36844   float64x2_t __ret;
   36845   __ret = (float64x2_t)(__p0);
   36846   return __ret;
   36847 }
   36848 #endif
   36849 
   36850 #ifdef __LITTLE_ENDIAN__
   36851 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
   36852   float32x4_t __ret;
   36853   __ret = (float32x4_t)(__p0);
   36854   return __ret;
   36855 }
   36856 #else
   36857 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
   36858   float32x4_t __ret;
   36859   __ret = (float32x4_t)(__p0);
   36860   return __ret;
   36861 }
   36862 #endif
   36863 
   36864 #ifdef __LITTLE_ENDIAN__
   36865 __ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
   36866   float32x4_t __ret;
   36867   __ret = (float32x4_t)(__p0);
   36868   return __ret;
   36869 }
   36870 #else
   36871 __ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
   36872   float32x4_t __ret;
   36873   __ret = (float32x4_t)(__p0);
   36874   return __ret;
   36875 }
   36876 #endif
   36877 
   36878 #ifdef __LITTLE_ENDIAN__
   36879 __ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
   36880   float32x4_t __ret;
   36881   __ret = (float32x4_t)(__p0);
   36882   return __ret;
   36883 }
   36884 #else
   36885 __ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
   36886   float32x4_t __ret;
   36887   __ret = (float32x4_t)(__p0);
   36888   return __ret;
   36889 }
   36890 #endif
   36891 
   36892 #ifdef __LITTLE_ENDIAN__
   36893 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
   36894   float32x4_t __ret;
   36895   __ret = (float32x4_t)(__p0);
   36896   return __ret;
   36897 }
   36898 #else
   36899 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
   36900   float32x4_t __ret;
   36901   __ret = (float32x4_t)(__p0);
   36902   return __ret;
   36903 }
   36904 #endif
   36905 
   36906 #ifdef __LITTLE_ENDIAN__
   36907 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
   36908   float32x4_t __ret;
   36909   __ret = (float32x4_t)(__p0);
   36910   return __ret;
   36911 }
   36912 #else
   36913 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
   36914   float32x4_t __ret;
   36915   __ret = (float32x4_t)(__p0);
   36916   return __ret;
   36917 }
   36918 #endif
   36919 
   36920 #ifdef __LITTLE_ENDIAN__
   36921 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
   36922   float32x4_t __ret;
   36923   __ret = (float32x4_t)(__p0);
   36924   return __ret;
   36925 }
   36926 #else
   36927 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
   36928   float32x4_t __ret;
   36929   __ret = (float32x4_t)(__p0);
   36930   return __ret;
   36931 }
   36932 #endif
   36933 
   36934 #ifdef __LITTLE_ENDIAN__
   36935 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
   36936   float32x4_t __ret;
   36937   __ret = (float32x4_t)(__p0);
   36938   return __ret;
   36939 }
   36940 #else
   36941 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
   36942   float32x4_t __ret;
   36943   __ret = (float32x4_t)(__p0);
   36944   return __ret;
   36945 }
   36946 #endif
   36947 
   36948 #ifdef __LITTLE_ENDIAN__
   36949 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
   36950   float32x4_t __ret;
   36951   __ret = (float32x4_t)(__p0);
   36952   return __ret;
   36953 }
   36954 #else
   36955 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
   36956   float32x4_t __ret;
   36957   __ret = (float32x4_t)(__p0);
   36958   return __ret;
   36959 }
   36960 #endif
   36961 
   36962 #ifdef __LITTLE_ENDIAN__
   36963 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
   36964   float32x4_t __ret;
   36965   __ret = (float32x4_t)(__p0);
   36966   return __ret;
   36967 }
   36968 #else
   36969 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
   36970   float32x4_t __ret;
   36971   __ret = (float32x4_t)(__p0);
   36972   return __ret;
   36973 }
   36974 #endif
   36975 
   36976 #ifdef __LITTLE_ENDIAN__
   36977 __ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
   36978   float32x4_t __ret;
   36979   __ret = (float32x4_t)(__p0);
   36980   return __ret;
   36981 }
   36982 #else
   36983 __ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
   36984   float32x4_t __ret;
   36985   __ret = (float32x4_t)(__p0);
   36986   return __ret;
   36987 }
   36988 #endif
   36989 
   36990 #ifdef __LITTLE_ENDIAN__
   36991 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
   36992   float32x4_t __ret;
   36993   __ret = (float32x4_t)(__p0);
   36994   return __ret;
   36995 }
   36996 #else
   36997 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
   36998   float32x4_t __ret;
   36999   __ret = (float32x4_t)(__p0);
   37000   return __ret;
   37001 }
   37002 #endif
   37003 
   37004 #ifdef __LITTLE_ENDIAN__
   37005 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
   37006   float32x4_t __ret;
   37007   __ret = (float32x4_t)(__p0);
   37008   return __ret;
   37009 }
   37010 #else
   37011 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
   37012   float32x4_t __ret;
   37013   __ret = (float32x4_t)(__p0);
   37014   return __ret;
   37015 }
   37016 #endif
   37017 
   37018 #ifdef __LITTLE_ENDIAN__
   37019 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
   37020   float32x4_t __ret;
   37021   __ret = (float32x4_t)(__p0);
   37022   return __ret;
   37023 }
   37024 #else
   37025 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
   37026   float32x4_t __ret;
   37027   __ret = (float32x4_t)(__p0);
   37028   return __ret;
   37029 }
   37030 #endif
   37031 
   37032 #ifdef __LITTLE_ENDIAN__
   37033 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
   37034   float32x4_t __ret;
   37035   __ret = (float32x4_t)(__p0);
   37036   return __ret;
   37037 }
   37038 #else
   37039 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
   37040   float32x4_t __ret;
   37041   __ret = (float32x4_t)(__p0);
   37042   return __ret;
   37043 }
   37044 #endif
   37045 
   37046 #ifdef __LITTLE_ENDIAN__
   37047 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
   37048   float16x8_t __ret;
   37049   __ret = (float16x8_t)(__p0);
   37050   return __ret;
   37051 }
   37052 #else
   37053 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
   37054   float16x8_t __ret;
   37055   __ret = (float16x8_t)(__p0);
   37056   return __ret;
   37057 }
   37058 #endif
   37059 
   37060 #ifdef __LITTLE_ENDIAN__
   37061 __ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
   37062   float16x8_t __ret;
   37063   __ret = (float16x8_t)(__p0);
   37064   return __ret;
   37065 }
   37066 #else
   37067 __ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
   37068   float16x8_t __ret;
   37069   __ret = (float16x8_t)(__p0);
   37070   return __ret;
   37071 }
   37072 #endif
   37073 
   37074 #ifdef __LITTLE_ENDIAN__
   37075 __ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
   37076   float16x8_t __ret;
   37077   __ret = (float16x8_t)(__p0);
   37078   return __ret;
   37079 }
   37080 #else
   37081 __ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
   37082   float16x8_t __ret;
   37083   __ret = (float16x8_t)(__p0);
   37084   return __ret;
   37085 }
   37086 #endif
   37087 
   37088 #ifdef __LITTLE_ENDIAN__
   37089 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
   37090   float16x8_t __ret;
   37091   __ret = (float16x8_t)(__p0);
   37092   return __ret;
   37093 }
   37094 #else
   37095 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
   37096   float16x8_t __ret;
   37097   __ret = (float16x8_t)(__p0);
   37098   return __ret;
   37099 }
   37100 #endif
   37101 
   37102 #ifdef __LITTLE_ENDIAN__
   37103 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
   37104   float16x8_t __ret;
   37105   __ret = (float16x8_t)(__p0);
   37106   return __ret;
   37107 }
   37108 #else
   37109 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
   37110   float16x8_t __ret;
   37111   __ret = (float16x8_t)(__p0);
   37112   return __ret;
   37113 }
   37114 #endif
   37115 
   37116 #ifdef __LITTLE_ENDIAN__
   37117 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
   37118   float16x8_t __ret;
   37119   __ret = (float16x8_t)(__p0);
   37120   return __ret;
   37121 }
   37122 #else
   37123 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
   37124   float16x8_t __ret;
   37125   __ret = (float16x8_t)(__p0);
   37126   return __ret;
   37127 }
   37128 #endif
   37129 
   37130 #ifdef __LITTLE_ENDIAN__
   37131 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
   37132   float16x8_t __ret;
   37133   __ret = (float16x8_t)(__p0);
   37134   return __ret;
   37135 }
   37136 #else
   37137 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
   37138   float16x8_t __ret;
   37139   __ret = (float16x8_t)(__p0);
   37140   return __ret;
   37141 }
   37142 #endif
   37143 
   37144 #ifdef __LITTLE_ENDIAN__
   37145 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
   37146   float16x8_t __ret;
   37147   __ret = (float16x8_t)(__p0);
   37148   return __ret;
   37149 }
   37150 #else
   37151 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
   37152   float16x8_t __ret;
   37153   __ret = (float16x8_t)(__p0);
   37154   return __ret;
   37155 }
   37156 #endif
   37157 
   37158 #ifdef __LITTLE_ENDIAN__
   37159 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
   37160   float16x8_t __ret;
   37161   __ret = (float16x8_t)(__p0);
   37162   return __ret;
   37163 }
   37164 #else
   37165 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
   37166   float16x8_t __ret;
   37167   __ret = (float16x8_t)(__p0);
   37168   return __ret;
   37169 }
   37170 #endif
   37171 
   37172 #ifdef __LITTLE_ENDIAN__
   37173 __ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
   37174   float16x8_t __ret;
   37175   __ret = (float16x8_t)(__p0);
   37176   return __ret;
   37177 }
   37178 #else
   37179 __ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
   37180   float16x8_t __ret;
   37181   __ret = (float16x8_t)(__p0);
   37182   return __ret;
   37183 }
   37184 #endif
   37185 
   37186 #ifdef __LITTLE_ENDIAN__
   37187 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
   37188   float16x8_t __ret;
   37189   __ret = (float16x8_t)(__p0);
   37190   return __ret;
   37191 }
   37192 #else
   37193 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
   37194   float16x8_t __ret;
   37195   __ret = (float16x8_t)(__p0);
   37196   return __ret;
   37197 }
   37198 #endif
   37199 
   37200 #ifdef __LITTLE_ENDIAN__
   37201 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
   37202   float16x8_t __ret;
   37203   __ret = (float16x8_t)(__p0);
   37204   return __ret;
   37205 }
   37206 #else
   37207 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
   37208   float16x8_t __ret;
   37209   __ret = (float16x8_t)(__p0);
   37210   return __ret;
   37211 }
   37212 #endif
   37213 
   37214 #ifdef __LITTLE_ENDIAN__
   37215 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
   37216   float16x8_t __ret;
   37217   __ret = (float16x8_t)(__p0);
   37218   return __ret;
   37219 }
   37220 #else
   37221 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
   37222   float16x8_t __ret;
   37223   __ret = (float16x8_t)(__p0);
   37224   return __ret;
   37225 }
   37226 #endif
   37227 
   37228 #ifdef __LITTLE_ENDIAN__
   37229 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
   37230   float16x8_t __ret;
   37231   __ret = (float16x8_t)(__p0);
   37232   return __ret;
   37233 }
   37234 #else
   37235 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
   37236   float16x8_t __ret;
   37237   __ret = (float16x8_t)(__p0);
   37238   return __ret;
   37239 }
   37240 #endif
   37241 
   37242 #ifdef __LITTLE_ENDIAN__
   37243 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
   37244   int32x4_t __ret;
   37245   __ret = (int32x4_t)(__p0);
   37246   return __ret;
   37247 }
   37248 #else
   37249 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
   37250   int32x4_t __ret;
   37251   __ret = (int32x4_t)(__p0);
   37252   return __ret;
   37253 }
   37254 #endif
   37255 
   37256 #ifdef __LITTLE_ENDIAN__
   37257 __ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
   37258   int32x4_t __ret;
   37259   __ret = (int32x4_t)(__p0);
   37260   return __ret;
   37261 }
   37262 #else
   37263 __ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
   37264   int32x4_t __ret;
   37265   __ret = (int32x4_t)(__p0);
   37266   return __ret;
   37267 }
   37268 #endif
   37269 
   37270 #ifdef __LITTLE_ENDIAN__
   37271 __ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
   37272   int32x4_t __ret;
   37273   __ret = (int32x4_t)(__p0);
   37274   return __ret;
   37275 }
   37276 #else
   37277 __ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
   37278   int32x4_t __ret;
   37279   __ret = (int32x4_t)(__p0);
   37280   return __ret;
   37281 }
   37282 #endif
   37283 
   37284 #ifdef __LITTLE_ENDIAN__
   37285 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
   37286   int32x4_t __ret;
   37287   __ret = (int32x4_t)(__p0);
   37288   return __ret;
   37289 }
   37290 #else
   37291 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
   37292   int32x4_t __ret;
   37293   __ret = (int32x4_t)(__p0);
   37294   return __ret;
   37295 }
   37296 #endif
   37297 
   37298 #ifdef __LITTLE_ENDIAN__
   37299 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
   37300   int32x4_t __ret;
   37301   __ret = (int32x4_t)(__p0);
   37302   return __ret;
   37303 }
   37304 #else
   37305 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
   37306   int32x4_t __ret;
   37307   __ret = (int32x4_t)(__p0);
   37308   return __ret;
   37309 }
   37310 #endif
   37311 
   37312 #ifdef __LITTLE_ENDIAN__
   37313 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
   37314   int32x4_t __ret;
   37315   __ret = (int32x4_t)(__p0);
   37316   return __ret;
   37317 }
   37318 #else
   37319 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
   37320   int32x4_t __ret;
   37321   __ret = (int32x4_t)(__p0);
   37322   return __ret;
   37323 }
   37324 #endif
   37325 
   37326 #ifdef __LITTLE_ENDIAN__
   37327 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
   37328   int32x4_t __ret;
   37329   __ret = (int32x4_t)(__p0);
   37330   return __ret;
   37331 }
   37332 #else
   37333 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
   37334   int32x4_t __ret;
   37335   __ret = (int32x4_t)(__p0);
   37336   return __ret;
   37337 }
   37338 #endif
   37339 
   37340 #ifdef __LITTLE_ENDIAN__
   37341 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
   37342   int32x4_t __ret;
   37343   __ret = (int32x4_t)(__p0);
   37344   return __ret;
   37345 }
   37346 #else
   37347 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
   37348   int32x4_t __ret;
   37349   __ret = (int32x4_t)(__p0);
   37350   return __ret;
   37351 }
   37352 #endif
   37353 
   37354 #ifdef __LITTLE_ENDIAN__
   37355 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
   37356   int32x4_t __ret;
   37357   __ret = (int32x4_t)(__p0);
   37358   return __ret;
   37359 }
   37360 #else
   37361 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
   37362   int32x4_t __ret;
   37363   __ret = (int32x4_t)(__p0);
   37364   return __ret;
   37365 }
   37366 #endif
   37367 
   37368 #ifdef __LITTLE_ENDIAN__
   37369 __ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
   37370   int32x4_t __ret;
   37371   __ret = (int32x4_t)(__p0);
   37372   return __ret;
   37373 }
   37374 #else
   37375 __ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
   37376   int32x4_t __ret;
   37377   __ret = (int32x4_t)(__p0);
   37378   return __ret;
   37379 }
   37380 #endif
   37381 
   37382 #ifdef __LITTLE_ENDIAN__
   37383 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
   37384   int32x4_t __ret;
   37385   __ret = (int32x4_t)(__p0);
   37386   return __ret;
   37387 }
   37388 #else
   37389 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
   37390   int32x4_t __ret;
   37391   __ret = (int32x4_t)(__p0);
   37392   return __ret;
   37393 }
   37394 #endif
   37395 
   37396 #ifdef __LITTLE_ENDIAN__
   37397 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
   37398   int32x4_t __ret;
   37399   __ret = (int32x4_t)(__p0);
   37400   return __ret;
   37401 }
   37402 #else
   37403 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
   37404   int32x4_t __ret;
   37405   __ret = (int32x4_t)(__p0);
   37406   return __ret;
   37407 }
   37408 #endif
   37409 
   37410 #ifdef __LITTLE_ENDIAN__
   37411 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
   37412   int32x4_t __ret;
   37413   __ret = (int32x4_t)(__p0);
   37414   return __ret;
   37415 }
   37416 #else
   37417 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
   37418   int32x4_t __ret;
   37419   __ret = (int32x4_t)(__p0);
   37420   return __ret;
   37421 }
   37422 #endif
   37423 
   37424 #ifdef __LITTLE_ENDIAN__
   37425 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
   37426   int32x4_t __ret;
   37427   __ret = (int32x4_t)(__p0);
   37428   return __ret;
   37429 }
   37430 #else
   37431 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
   37432   int32x4_t __ret;
   37433   __ret = (int32x4_t)(__p0);
   37434   return __ret;
   37435 }
   37436 #endif
   37437 
   37438 #ifdef __LITTLE_ENDIAN__
   37439 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
   37440   int64x2_t __ret;
   37441   __ret = (int64x2_t)(__p0);
   37442   return __ret;
   37443 }
   37444 #else
   37445 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
   37446   int64x2_t __ret;
   37447   __ret = (int64x2_t)(__p0);
   37448   return __ret;
   37449 }
   37450 #endif
   37451 
   37452 #ifdef __LITTLE_ENDIAN__
   37453 __ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
   37454   int64x2_t __ret;
   37455   __ret = (int64x2_t)(__p0);
   37456   return __ret;
   37457 }
   37458 #else
   37459 __ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
   37460   int64x2_t __ret;
   37461   __ret = (int64x2_t)(__p0);
   37462   return __ret;
   37463 }
   37464 #endif
   37465 
   37466 #ifdef __LITTLE_ENDIAN__
   37467 __ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
   37468   int64x2_t __ret;
   37469   __ret = (int64x2_t)(__p0);
   37470   return __ret;
   37471 }
   37472 #else
   37473 __ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
   37474   int64x2_t __ret;
   37475   __ret = (int64x2_t)(__p0);
   37476   return __ret;
   37477 }
   37478 #endif
   37479 
   37480 #ifdef __LITTLE_ENDIAN__
   37481 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
   37482   int64x2_t __ret;
   37483   __ret = (int64x2_t)(__p0);
   37484   return __ret;
   37485 }
   37486 #else
   37487 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
   37488   int64x2_t __ret;
   37489   __ret = (int64x2_t)(__p0);
   37490   return __ret;
   37491 }
   37492 #endif
   37493 
   37494 #ifdef __LITTLE_ENDIAN__
   37495 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
   37496   int64x2_t __ret;
   37497   __ret = (int64x2_t)(__p0);
   37498   return __ret;
   37499 }
   37500 #else
   37501 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
   37502   int64x2_t __ret;
   37503   __ret = (int64x2_t)(__p0);
   37504   return __ret;
   37505 }
   37506 #endif
   37507 
   37508 #ifdef __LITTLE_ENDIAN__
   37509 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
   37510   int64x2_t __ret;
   37511   __ret = (int64x2_t)(__p0);
   37512   return __ret;
   37513 }
   37514 #else
   37515 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
   37516   int64x2_t __ret;
   37517   __ret = (int64x2_t)(__p0);
   37518   return __ret;
   37519 }
   37520 #endif
   37521 
   37522 #ifdef __LITTLE_ENDIAN__
   37523 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
   37524   int64x2_t __ret;
   37525   __ret = (int64x2_t)(__p0);
   37526   return __ret;
   37527 }
   37528 #else
   37529 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
   37530   int64x2_t __ret;
   37531   __ret = (int64x2_t)(__p0);
   37532   return __ret;
   37533 }
   37534 #endif
   37535 
   37536 #ifdef __LITTLE_ENDIAN__
   37537 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
   37538   int64x2_t __ret;
   37539   __ret = (int64x2_t)(__p0);
   37540   return __ret;
   37541 }
   37542 #else
   37543 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
   37544   int64x2_t __ret;
   37545   __ret = (int64x2_t)(__p0);
   37546   return __ret;
   37547 }
   37548 #endif
   37549 
   37550 #ifdef __LITTLE_ENDIAN__
   37551 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
   37552   int64x2_t __ret;
   37553   __ret = (int64x2_t)(__p0);
   37554   return __ret;
   37555 }
   37556 #else
   37557 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
   37558   int64x2_t __ret;
   37559   __ret = (int64x2_t)(__p0);
   37560   return __ret;
   37561 }
   37562 #endif
   37563 
   37564 #ifdef __LITTLE_ENDIAN__
   37565 __ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
   37566   int64x2_t __ret;
   37567   __ret = (int64x2_t)(__p0);
   37568   return __ret;
   37569 }
   37570 #else
   37571 __ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
   37572   int64x2_t __ret;
   37573   __ret = (int64x2_t)(__p0);
   37574   return __ret;
   37575 }
   37576 #endif
   37577 
   37578 #ifdef __LITTLE_ENDIAN__
   37579 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
   37580   int64x2_t __ret;
   37581   __ret = (int64x2_t)(__p0);
   37582   return __ret;
   37583 }
   37584 #else
   37585 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
   37586   int64x2_t __ret;
   37587   __ret = (int64x2_t)(__p0);
   37588   return __ret;
   37589 }
   37590 #endif
   37591 
   37592 #ifdef __LITTLE_ENDIAN__
   37593 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
   37594   int64x2_t __ret;
   37595   __ret = (int64x2_t)(__p0);
   37596   return __ret;
   37597 }
   37598 #else
   37599 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
   37600   int64x2_t __ret;
   37601   __ret = (int64x2_t)(__p0);
   37602   return __ret;
   37603 }
   37604 #endif
   37605 
   37606 #ifdef __LITTLE_ENDIAN__
   37607 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
   37608   int64x2_t __ret;
   37609   __ret = (int64x2_t)(__p0);
   37610   return __ret;
   37611 }
   37612 #else
   37613 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
   37614   int64x2_t __ret;
   37615   __ret = (int64x2_t)(__p0);
   37616   return __ret;
   37617 }
   37618 #endif
   37619 
   37620 #ifdef __LITTLE_ENDIAN__
   37621 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
   37622   int64x2_t __ret;
   37623   __ret = (int64x2_t)(__p0);
   37624   return __ret;
   37625 }
   37626 #else
   37627 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
   37628   int64x2_t __ret;
   37629   __ret = (int64x2_t)(__p0);
   37630   return __ret;
   37631 }
   37632 #endif
   37633 
   37634 #ifdef __LITTLE_ENDIAN__
   37635 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
   37636   int16x8_t __ret;
   37637   __ret = (int16x8_t)(__p0);
   37638   return __ret;
   37639 }
   37640 #else
   37641 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
   37642   int16x8_t __ret;
   37643   __ret = (int16x8_t)(__p0);
   37644   return __ret;
   37645 }
   37646 #endif
   37647 
   37648 #ifdef __LITTLE_ENDIAN__
   37649 __ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
   37650   int16x8_t __ret;
   37651   __ret = (int16x8_t)(__p0);
   37652   return __ret;
   37653 }
   37654 #else
   37655 __ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
   37656   int16x8_t __ret;
   37657   __ret = (int16x8_t)(__p0);
   37658   return __ret;
   37659 }
   37660 #endif
   37661 
   37662 #ifdef __LITTLE_ENDIAN__
   37663 __ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
   37664   int16x8_t __ret;
   37665   __ret = (int16x8_t)(__p0);
   37666   return __ret;
   37667 }
   37668 #else
   37669 __ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
   37670   int16x8_t __ret;
   37671   __ret = (int16x8_t)(__p0);
   37672   return __ret;
   37673 }
   37674 #endif
   37675 
   37676 #ifdef __LITTLE_ENDIAN__
   37677 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
   37678   int16x8_t __ret;
   37679   __ret = (int16x8_t)(__p0);
   37680   return __ret;
   37681 }
   37682 #else
   37683 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
   37684   int16x8_t __ret;
   37685   __ret = (int16x8_t)(__p0);
   37686   return __ret;
   37687 }
   37688 #endif
   37689 
   37690 #ifdef __LITTLE_ENDIAN__
   37691 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
   37692   int16x8_t __ret;
   37693   __ret = (int16x8_t)(__p0);
   37694   return __ret;
   37695 }
   37696 #else
   37697 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
   37698   int16x8_t __ret;
   37699   __ret = (int16x8_t)(__p0);
   37700   return __ret;
   37701 }
   37702 #endif
   37703 
   37704 #ifdef __LITTLE_ENDIAN__
   37705 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
   37706   int16x8_t __ret;
   37707   __ret = (int16x8_t)(__p0);
   37708   return __ret;
   37709 }
   37710 #else
   37711 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
   37712   int16x8_t __ret;
   37713   __ret = (int16x8_t)(__p0);
   37714   return __ret;
   37715 }
   37716 #endif
   37717 
   37718 #ifdef __LITTLE_ENDIAN__
   37719 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
   37720   int16x8_t __ret;
   37721   __ret = (int16x8_t)(__p0);
   37722   return __ret;
   37723 }
   37724 #else
   37725 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
   37726   int16x8_t __ret;
   37727   __ret = (int16x8_t)(__p0);
   37728   return __ret;
   37729 }
   37730 #endif
   37731 
   37732 #ifdef __LITTLE_ENDIAN__
   37733 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
   37734   int16x8_t __ret;
   37735   __ret = (int16x8_t)(__p0);
   37736   return __ret;
   37737 }
   37738 #else
   37739 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
   37740   int16x8_t __ret;
   37741   __ret = (int16x8_t)(__p0);
   37742   return __ret;
   37743 }
   37744 #endif
   37745 
   37746 #ifdef __LITTLE_ENDIAN__
   37747 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
   37748   int16x8_t __ret;
   37749   __ret = (int16x8_t)(__p0);
   37750   return __ret;
   37751 }
   37752 #else
   37753 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
   37754   int16x8_t __ret;
   37755   __ret = (int16x8_t)(__p0);
   37756   return __ret;
   37757 }
   37758 #endif
   37759 
   37760 #ifdef __LITTLE_ENDIAN__
   37761 __ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
   37762   int16x8_t __ret;
   37763   __ret = (int16x8_t)(__p0);
   37764   return __ret;
   37765 }
   37766 #else
   37767 __ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
   37768   int16x8_t __ret;
   37769   __ret = (int16x8_t)(__p0);
   37770   return __ret;
   37771 }
   37772 #endif
   37773 
   37774 #ifdef __LITTLE_ENDIAN__
   37775 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
   37776   int16x8_t __ret;
   37777   __ret = (int16x8_t)(__p0);
   37778   return __ret;
   37779 }
   37780 #else
   37781 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
   37782   int16x8_t __ret;
   37783   __ret = (int16x8_t)(__p0);
   37784   return __ret;
   37785 }
   37786 #endif
   37787 
   37788 #ifdef __LITTLE_ENDIAN__
   37789 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
   37790   int16x8_t __ret;
   37791   __ret = (int16x8_t)(__p0);
   37792   return __ret;
   37793 }
   37794 #else
   37795 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
   37796   int16x8_t __ret;
   37797   __ret = (int16x8_t)(__p0);
   37798   return __ret;
   37799 }
   37800 #endif
   37801 
   37802 #ifdef __LITTLE_ENDIAN__
   37803 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
   37804   int16x8_t __ret;
   37805   __ret = (int16x8_t)(__p0);
   37806   return __ret;
   37807 }
   37808 #else
   37809 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
   37810   int16x8_t __ret;
   37811   __ret = (int16x8_t)(__p0);
   37812   return __ret;
   37813 }
   37814 #endif
   37815 
   37816 #ifdef __LITTLE_ENDIAN__
   37817 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
   37818   int16x8_t __ret;
   37819   __ret = (int16x8_t)(__p0);
   37820   return __ret;
   37821 }
   37822 #else
   37823 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
   37824   int16x8_t __ret;
   37825   __ret = (int16x8_t)(__p0);
   37826   return __ret;
   37827 }
   37828 #endif
   37829 
   37830 #ifdef __LITTLE_ENDIAN__
   37831 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
   37832   uint8x8_t __ret;
   37833   __ret = (uint8x8_t)(__p0);
   37834   return __ret;
   37835 }
   37836 #else
   37837 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
   37838   uint8x8_t __ret;
   37839   __ret = (uint8x8_t)(__p0);
   37840   return __ret;
   37841 }
   37842 #endif
   37843 
   37844 #ifdef __LITTLE_ENDIAN__
   37845 __ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
   37846   uint8x8_t __ret;
   37847   __ret = (uint8x8_t)(__p0);
   37848   return __ret;
   37849 }
   37850 #else
   37851 __ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
   37852   uint8x8_t __ret;
   37853   __ret = (uint8x8_t)(__p0);
   37854   return __ret;
   37855 }
   37856 #endif
   37857 
   37858 #ifdef __LITTLE_ENDIAN__
   37859 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
   37860   uint8x8_t __ret;
   37861   __ret = (uint8x8_t)(__p0);
   37862   return __ret;
   37863 }
   37864 #else
   37865 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
   37866   uint8x8_t __ret;
   37867   __ret = (uint8x8_t)(__p0);
   37868   return __ret;
   37869 }
   37870 #endif
   37871 
   37872 #ifdef __LITTLE_ENDIAN__
   37873 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
   37874   uint8x8_t __ret;
   37875   __ret = (uint8x8_t)(__p0);
   37876   return __ret;
   37877 }
   37878 #else
   37879 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
   37880   uint8x8_t __ret;
   37881   __ret = (uint8x8_t)(__p0);
   37882   return __ret;
   37883 }
   37884 #endif
   37885 
   37886 #ifdef __LITTLE_ENDIAN__
   37887 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
   37888   uint8x8_t __ret;
   37889   __ret = (uint8x8_t)(__p0);
   37890   return __ret;
   37891 }
   37892 #else
   37893 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
   37894   uint8x8_t __ret;
   37895   __ret = (uint8x8_t)(__p0);
   37896   return __ret;
   37897 }
   37898 #endif
   37899 
   37900 #ifdef __LITTLE_ENDIAN__
   37901 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
   37902   uint8x8_t __ret;
   37903   __ret = (uint8x8_t)(__p0);
   37904   return __ret;
   37905 }
   37906 #else
   37907 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
   37908   uint8x8_t __ret;
   37909   __ret = (uint8x8_t)(__p0);
   37910   return __ret;
   37911 }
   37912 #endif
   37913 
   37914 #ifdef __LITTLE_ENDIAN__
   37915 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
   37916   uint8x8_t __ret;
   37917   __ret = (uint8x8_t)(__p0);
   37918   return __ret;
   37919 }
   37920 #else
   37921 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
   37922   uint8x8_t __ret;
   37923   __ret = (uint8x8_t)(__p0);
   37924   return __ret;
   37925 }
   37926 #endif
   37927 
   37928 #ifdef __LITTLE_ENDIAN__
   37929 __ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
   37930   uint8x8_t __ret;
   37931   __ret = (uint8x8_t)(__p0);
   37932   return __ret;
   37933 }
   37934 #else
   37935 __ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
   37936   uint8x8_t __ret;
   37937   __ret = (uint8x8_t)(__p0);
   37938   return __ret;
   37939 }
   37940 #endif
   37941 
   37942 #ifdef __LITTLE_ENDIAN__
   37943 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
   37944   uint8x8_t __ret;
   37945   __ret = (uint8x8_t)(__p0);
   37946   return __ret;
   37947 }
   37948 #else
   37949 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
   37950   uint8x8_t __ret;
   37951   __ret = (uint8x8_t)(__p0);
   37952   return __ret;
   37953 }
   37954 #endif
   37955 
   37956 #ifdef __LITTLE_ENDIAN__
   37957 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
   37958   uint8x8_t __ret;
   37959   __ret = (uint8x8_t)(__p0);
   37960   return __ret;
   37961 }
   37962 #else
   37963 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
   37964   uint8x8_t __ret;
   37965   __ret = (uint8x8_t)(__p0);
   37966   return __ret;
   37967 }
   37968 #endif
   37969 
   37970 #ifdef __LITTLE_ENDIAN__
   37971 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
   37972   uint8x8_t __ret;
   37973   __ret = (uint8x8_t)(__p0);
   37974   return __ret;
   37975 }
   37976 #else
   37977 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
   37978   uint8x8_t __ret;
   37979   __ret = (uint8x8_t)(__p0);
   37980   return __ret;
   37981 }
   37982 #endif
   37983 
   37984 #ifdef __LITTLE_ENDIAN__
   37985 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
   37986   uint8x8_t __ret;
   37987   __ret = (uint8x8_t)(__p0);
   37988   return __ret;
   37989 }
   37990 #else
   37991 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
   37992   uint8x8_t __ret;
   37993   __ret = (uint8x8_t)(__p0);
   37994   return __ret;
   37995 }
   37996 #endif
   37997 
   37998 #ifdef __LITTLE_ENDIAN__
   37999 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
   38000   uint8x8_t __ret;
   38001   __ret = (uint8x8_t)(__p0);
   38002   return __ret;
   38003 }
   38004 #else
   38005 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
   38006   uint8x8_t __ret;
   38007   __ret = (uint8x8_t)(__p0);
   38008   return __ret;
   38009 }
   38010 #endif
   38011 
   38012 #ifdef __LITTLE_ENDIAN__
   38013 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
   38014   uint32x2_t __ret;
   38015   __ret = (uint32x2_t)(__p0);
   38016   return __ret;
   38017 }
   38018 #else
   38019 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
   38020   uint32x2_t __ret;
   38021   __ret = (uint32x2_t)(__p0);
   38022   return __ret;
   38023 }
   38024 #endif
   38025 
   38026 #ifdef __LITTLE_ENDIAN__
   38027 __ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
   38028   uint32x2_t __ret;
   38029   __ret = (uint32x2_t)(__p0);
   38030   return __ret;
   38031 }
   38032 #else
   38033 __ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
   38034   uint32x2_t __ret;
   38035   __ret = (uint32x2_t)(__p0);
   38036   return __ret;
   38037 }
   38038 #endif
   38039 
   38040 #ifdef __LITTLE_ENDIAN__
   38041 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
   38042   uint32x2_t __ret;
   38043   __ret = (uint32x2_t)(__p0);
   38044   return __ret;
   38045 }
   38046 #else
   38047 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
   38048   uint32x2_t __ret;
   38049   __ret = (uint32x2_t)(__p0);
   38050   return __ret;
   38051 }
   38052 #endif
   38053 
   38054 #ifdef __LITTLE_ENDIAN__
   38055 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
   38056   uint32x2_t __ret;
   38057   __ret = (uint32x2_t)(__p0);
   38058   return __ret;
   38059 }
   38060 #else
   38061 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
   38062   uint32x2_t __ret;
   38063   __ret = (uint32x2_t)(__p0);
   38064   return __ret;
   38065 }
   38066 #endif
   38067 
   38068 #ifdef __LITTLE_ENDIAN__
   38069 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
   38070   uint32x2_t __ret;
   38071   __ret = (uint32x2_t)(__p0);
   38072   return __ret;
   38073 }
   38074 #else
   38075 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
   38076   uint32x2_t __ret;
   38077   __ret = (uint32x2_t)(__p0);
   38078   return __ret;
   38079 }
   38080 #endif
   38081 
   38082 #ifdef __LITTLE_ENDIAN__
   38083 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
   38084   uint32x2_t __ret;
   38085   __ret = (uint32x2_t)(__p0);
   38086   return __ret;
   38087 }
   38088 #else
   38089 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
   38090   uint32x2_t __ret;
   38091   __ret = (uint32x2_t)(__p0);
   38092   return __ret;
   38093 }
   38094 #endif
   38095 
   38096 #ifdef __LITTLE_ENDIAN__
   38097 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
   38098   uint32x2_t __ret;
   38099   __ret = (uint32x2_t)(__p0);
   38100   return __ret;
   38101 }
   38102 #else
   38103 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
   38104   uint32x2_t __ret;
   38105   __ret = (uint32x2_t)(__p0);
   38106   return __ret;
   38107 }
   38108 #endif
   38109 
   38110 #ifdef __LITTLE_ENDIAN__
   38111 __ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
   38112   uint32x2_t __ret;
   38113   __ret = (uint32x2_t)(__p0);
   38114   return __ret;
   38115 }
   38116 #else
   38117 __ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
   38118   uint32x2_t __ret;
   38119   __ret = (uint32x2_t)(__p0);
   38120   return __ret;
   38121 }
   38122 #endif
   38123 
   38124 #ifdef __LITTLE_ENDIAN__
   38125 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
   38126   uint32x2_t __ret;
   38127   __ret = (uint32x2_t)(__p0);
   38128   return __ret;
   38129 }
   38130 #else
   38131 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
   38132   uint32x2_t __ret;
   38133   __ret = (uint32x2_t)(__p0);
   38134   return __ret;
   38135 }
   38136 #endif
   38137 
   38138 #ifdef __LITTLE_ENDIAN__
   38139 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
   38140   uint32x2_t __ret;
   38141   __ret = (uint32x2_t)(__p0);
   38142   return __ret;
   38143 }
   38144 #else
   38145 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
   38146   uint32x2_t __ret;
   38147   __ret = (uint32x2_t)(__p0);
   38148   return __ret;
   38149 }
   38150 #endif
   38151 
   38152 #ifdef __LITTLE_ENDIAN__
   38153 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
   38154   uint32x2_t __ret;
   38155   __ret = (uint32x2_t)(__p0);
   38156   return __ret;
   38157 }
   38158 #else
   38159 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
   38160   uint32x2_t __ret;
   38161   __ret = (uint32x2_t)(__p0);
   38162   return __ret;
   38163 }
   38164 #endif
   38165 
   38166 #ifdef __LITTLE_ENDIAN__
   38167 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
   38168   uint32x2_t __ret;
   38169   __ret = (uint32x2_t)(__p0);
   38170   return __ret;
   38171 }
   38172 #else
   38173 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
   38174   uint32x2_t __ret;
   38175   __ret = (uint32x2_t)(__p0);
   38176   return __ret;
   38177 }
   38178 #endif
   38179 
   38180 #ifdef __LITTLE_ENDIAN__
   38181 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
   38182   uint32x2_t __ret;
   38183   __ret = (uint32x2_t)(__p0);
   38184   return __ret;
   38185 }
   38186 #else
   38187 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
   38188   uint32x2_t __ret;
   38189   __ret = (uint32x2_t)(__p0);
   38190   return __ret;
   38191 }
   38192 #endif
   38193 
   38194 #ifdef __LITTLE_ENDIAN__
   38195 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
   38196   uint64x1_t __ret;
   38197   __ret = (uint64x1_t)(__p0);
   38198   return __ret;
   38199 }
   38200 #else
   38201 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
   38202   uint64x1_t __ret;
   38203   __ret = (uint64x1_t)(__p0);
   38204   return __ret;
   38205 }
   38206 #endif
   38207 
   38208 #ifdef __LITTLE_ENDIAN__
   38209 __ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
   38210   uint64x1_t __ret;
   38211   __ret = (uint64x1_t)(__p0);
   38212   return __ret;
   38213 }
   38214 #else
   38215 __ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
   38216   uint64x1_t __ret;
   38217   __ret = (uint64x1_t)(__p0);
   38218   return __ret;
   38219 }
   38220 #endif
   38221 
   38222 #ifdef __LITTLE_ENDIAN__
   38223 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
   38224   uint64x1_t __ret;
   38225   __ret = (uint64x1_t)(__p0);
   38226   return __ret;
   38227 }
   38228 #else
   38229 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
   38230   uint64x1_t __ret;
   38231   __ret = (uint64x1_t)(__p0);
   38232   return __ret;
   38233 }
   38234 #endif
   38235 
   38236 #ifdef __LITTLE_ENDIAN__
   38237 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
   38238   uint64x1_t __ret;
   38239   __ret = (uint64x1_t)(__p0);
   38240   return __ret;
   38241 }
   38242 #else
   38243 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
   38244   uint64x1_t __ret;
   38245   __ret = (uint64x1_t)(__p0);
   38246   return __ret;
   38247 }
   38248 #endif
   38249 
   38250 #ifdef __LITTLE_ENDIAN__
   38251 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
   38252   uint64x1_t __ret;
   38253   __ret = (uint64x1_t)(__p0);
   38254   return __ret;
   38255 }
   38256 #else
   38257 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
   38258   uint64x1_t __ret;
   38259   __ret = (uint64x1_t)(__p0);
   38260   return __ret;
   38261 }
   38262 #endif
   38263 
   38264 #ifdef __LITTLE_ENDIAN__
   38265 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
   38266   uint64x1_t __ret;
   38267   __ret = (uint64x1_t)(__p0);
   38268   return __ret;
   38269 }
   38270 #else
   38271 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
   38272   uint64x1_t __ret;
   38273   __ret = (uint64x1_t)(__p0);
   38274   return __ret;
   38275 }
   38276 #endif
   38277 
   38278 #ifdef __LITTLE_ENDIAN__
   38279 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
   38280   uint64x1_t __ret;
   38281   __ret = (uint64x1_t)(__p0);
   38282   return __ret;
   38283 }
   38284 #else
   38285 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
   38286   uint64x1_t __ret;
   38287   __ret = (uint64x1_t)(__p0);
   38288   return __ret;
   38289 }
   38290 #endif
   38291 
   38292 #ifdef __LITTLE_ENDIAN__
   38293 __ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
   38294   uint64x1_t __ret;
   38295   __ret = (uint64x1_t)(__p0);
   38296   return __ret;
   38297 }
   38298 #else
   38299 __ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
   38300   uint64x1_t __ret;
   38301   __ret = (uint64x1_t)(__p0);
   38302   return __ret;
   38303 }
   38304 #endif
   38305 
   38306 #ifdef __LITTLE_ENDIAN__
   38307 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
   38308   uint64x1_t __ret;
   38309   __ret = (uint64x1_t)(__p0);
   38310   return __ret;
   38311 }
   38312 #else
   38313 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
   38314   uint64x1_t __ret;
   38315   __ret = (uint64x1_t)(__p0);
   38316   return __ret;
   38317 }
   38318 #endif
   38319 
   38320 #ifdef __LITTLE_ENDIAN__
   38321 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
   38322   uint64x1_t __ret;
   38323   __ret = (uint64x1_t)(__p0);
   38324   return __ret;
   38325 }
   38326 #else
   38327 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
   38328   uint64x1_t __ret;
   38329   __ret = (uint64x1_t)(__p0);
   38330   return __ret;
   38331 }
   38332 #endif
   38333 
   38334 #ifdef __LITTLE_ENDIAN__
   38335 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
   38336   uint64x1_t __ret;
   38337   __ret = (uint64x1_t)(__p0);
   38338   return __ret;
   38339 }
   38340 #else
   38341 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
   38342   uint64x1_t __ret;
   38343   __ret = (uint64x1_t)(__p0);
   38344   return __ret;
   38345 }
   38346 #endif
   38347 
   38348 #ifdef __LITTLE_ENDIAN__
   38349 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
   38350   uint64x1_t __ret;
   38351   __ret = (uint64x1_t)(__p0);
   38352   return __ret;
   38353 }
   38354 #else
   38355 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
   38356   uint64x1_t __ret;
   38357   __ret = (uint64x1_t)(__p0);
   38358   return __ret;
   38359 }
   38360 #endif
   38361 
   38362 #ifdef __LITTLE_ENDIAN__
   38363 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
   38364   uint64x1_t __ret;
   38365   __ret = (uint64x1_t)(__p0);
   38366   return __ret;
   38367 }
   38368 #else
   38369 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
   38370   uint64x1_t __ret;
   38371   __ret = (uint64x1_t)(__p0);
   38372   return __ret;
   38373 }
   38374 #endif
   38375 
   38376 #ifdef __LITTLE_ENDIAN__
   38377 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
   38378   uint16x4_t __ret;
   38379   __ret = (uint16x4_t)(__p0);
   38380   return __ret;
   38381 }
   38382 #else
   38383 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
   38384   uint16x4_t __ret;
   38385   __ret = (uint16x4_t)(__p0);
   38386   return __ret;
   38387 }
   38388 #endif
   38389 
   38390 #ifdef __LITTLE_ENDIAN__
   38391 __ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
   38392   uint16x4_t __ret;
   38393   __ret = (uint16x4_t)(__p0);
   38394   return __ret;
   38395 }
   38396 #else
   38397 __ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
   38398   uint16x4_t __ret;
   38399   __ret = (uint16x4_t)(__p0);
   38400   return __ret;
   38401 }
   38402 #endif
   38403 
   38404 #ifdef __LITTLE_ENDIAN__
   38405 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
   38406   uint16x4_t __ret;
   38407   __ret = (uint16x4_t)(__p0);
   38408   return __ret;
   38409 }
   38410 #else
   38411 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
   38412   uint16x4_t __ret;
   38413   __ret = (uint16x4_t)(__p0);
   38414   return __ret;
   38415 }
   38416 #endif
   38417 
   38418 #ifdef __LITTLE_ENDIAN__
   38419 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
   38420   uint16x4_t __ret;
   38421   __ret = (uint16x4_t)(__p0);
   38422   return __ret;
   38423 }
   38424 #else
   38425 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
   38426   uint16x4_t __ret;
   38427   __ret = (uint16x4_t)(__p0);
   38428   return __ret;
   38429 }
   38430 #endif
   38431 
   38432 #ifdef __LITTLE_ENDIAN__
   38433 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
   38434   uint16x4_t __ret;
   38435   __ret = (uint16x4_t)(__p0);
   38436   return __ret;
   38437 }
   38438 #else
   38439 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
   38440   uint16x4_t __ret;
   38441   __ret = (uint16x4_t)(__p0);
   38442   return __ret;
   38443 }
   38444 #endif
   38445 
   38446 #ifdef __LITTLE_ENDIAN__
   38447 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
   38448   uint16x4_t __ret;
   38449   __ret = (uint16x4_t)(__p0);
   38450   return __ret;
   38451 }
   38452 #else
   38453 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
   38454   uint16x4_t __ret;
   38455   __ret = (uint16x4_t)(__p0);
   38456   return __ret;
   38457 }
   38458 #endif
   38459 
   38460 #ifdef __LITTLE_ENDIAN__
   38461 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
   38462   uint16x4_t __ret;
   38463   __ret = (uint16x4_t)(__p0);
   38464   return __ret;
   38465 }
   38466 #else
   38467 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
   38468   uint16x4_t __ret;
   38469   __ret = (uint16x4_t)(__p0);
   38470   return __ret;
   38471 }
   38472 #endif
   38473 
   38474 #ifdef __LITTLE_ENDIAN__
   38475 __ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
   38476   uint16x4_t __ret;
   38477   __ret = (uint16x4_t)(__p0);
   38478   return __ret;
   38479 }
   38480 #else
   38481 __ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
   38482   uint16x4_t __ret;
   38483   __ret = (uint16x4_t)(__p0);
   38484   return __ret;
   38485 }
   38486 #endif
   38487 
   38488 #ifdef __LITTLE_ENDIAN__
   38489 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
   38490   uint16x4_t __ret;
   38491   __ret = (uint16x4_t)(__p0);
   38492   return __ret;
   38493 }
   38494 #else
   38495 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
   38496   uint16x4_t __ret;
   38497   __ret = (uint16x4_t)(__p0);
   38498   return __ret;
   38499 }
   38500 #endif
   38501 
   38502 #ifdef __LITTLE_ENDIAN__
   38503 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
   38504   uint16x4_t __ret;
   38505   __ret = (uint16x4_t)(__p0);
   38506   return __ret;
   38507 }
   38508 #else
   38509 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
   38510   uint16x4_t __ret;
   38511   __ret = (uint16x4_t)(__p0);
   38512   return __ret;
   38513 }
   38514 #endif
   38515 
   38516 #ifdef __LITTLE_ENDIAN__
   38517 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
   38518   uint16x4_t __ret;
   38519   __ret = (uint16x4_t)(__p0);
   38520   return __ret;
   38521 }
   38522 #else
   38523 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
   38524   uint16x4_t __ret;
   38525   __ret = (uint16x4_t)(__p0);
   38526   return __ret;
   38527 }
   38528 #endif
   38529 
   38530 #ifdef __LITTLE_ENDIAN__
   38531 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
   38532   uint16x4_t __ret;
   38533   __ret = (uint16x4_t)(__p0);
   38534   return __ret;
   38535 }
   38536 #else
   38537 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
   38538   uint16x4_t __ret;
   38539   __ret = (uint16x4_t)(__p0);
   38540   return __ret;
   38541 }
   38542 #endif
   38543 
   38544 #ifdef __LITTLE_ENDIAN__
   38545 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
   38546   uint16x4_t __ret;
   38547   __ret = (uint16x4_t)(__p0);
   38548   return __ret;
   38549 }
   38550 #else
   38551 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
   38552   uint16x4_t __ret;
   38553   __ret = (uint16x4_t)(__p0);
   38554   return __ret;
   38555 }
   38556 #endif
   38557 
   38558 #ifdef __LITTLE_ENDIAN__
   38559 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
   38560   int8x8_t __ret;
   38561   __ret = (int8x8_t)(__p0);
   38562   return __ret;
   38563 }
   38564 #else
   38565 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
   38566   int8x8_t __ret;
   38567   __ret = (int8x8_t)(__p0);
   38568   return __ret;
   38569 }
   38570 #endif
   38571 
   38572 #ifdef __LITTLE_ENDIAN__
   38573 __ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
   38574   int8x8_t __ret;
   38575   __ret = (int8x8_t)(__p0);
   38576   return __ret;
   38577 }
   38578 #else
   38579 __ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
   38580   int8x8_t __ret;
   38581   __ret = (int8x8_t)(__p0);
   38582   return __ret;
   38583 }
   38584 #endif
   38585 
   38586 #ifdef __LITTLE_ENDIAN__
   38587 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
   38588   int8x8_t __ret;
   38589   __ret = (int8x8_t)(__p0);
   38590   return __ret;
   38591 }
   38592 #else
   38593 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
   38594   int8x8_t __ret;
   38595   __ret = (int8x8_t)(__p0);
   38596   return __ret;
   38597 }
   38598 #endif
   38599 
   38600 #ifdef __LITTLE_ENDIAN__
   38601 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
   38602   int8x8_t __ret;
   38603   __ret = (int8x8_t)(__p0);
   38604   return __ret;
   38605 }
   38606 #else
   38607 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
   38608   int8x8_t __ret;
   38609   __ret = (int8x8_t)(__p0);
   38610   return __ret;
   38611 }
   38612 #endif
   38613 
   38614 #ifdef __LITTLE_ENDIAN__
   38615 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
   38616   int8x8_t __ret;
   38617   __ret = (int8x8_t)(__p0);
   38618   return __ret;
   38619 }
   38620 #else
   38621 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
   38622   int8x8_t __ret;
   38623   __ret = (int8x8_t)(__p0);
   38624   return __ret;
   38625 }
   38626 #endif
   38627 
   38628 #ifdef __LITTLE_ENDIAN__
   38629 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
   38630   int8x8_t __ret;
   38631   __ret = (int8x8_t)(__p0);
   38632   return __ret;
   38633 }
   38634 #else
   38635 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
   38636   int8x8_t __ret;
   38637   __ret = (int8x8_t)(__p0);
   38638   return __ret;
   38639 }
   38640 #endif
   38641 
   38642 #ifdef __LITTLE_ENDIAN__
   38643 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
   38644   int8x8_t __ret;
   38645   __ret = (int8x8_t)(__p0);
   38646   return __ret;
   38647 }
   38648 #else
   38649 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
   38650   int8x8_t __ret;
   38651   __ret = (int8x8_t)(__p0);
   38652   return __ret;
   38653 }
   38654 #endif
   38655 
   38656 #ifdef __LITTLE_ENDIAN__
   38657 __ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
   38658   int8x8_t __ret;
   38659   __ret = (int8x8_t)(__p0);
   38660   return __ret;
   38661 }
   38662 #else
   38663 __ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
   38664   int8x8_t __ret;
   38665   __ret = (int8x8_t)(__p0);
   38666   return __ret;
   38667 }
   38668 #endif
   38669 
   38670 #ifdef __LITTLE_ENDIAN__
   38671 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
   38672   int8x8_t __ret;
   38673   __ret = (int8x8_t)(__p0);
   38674   return __ret;
   38675 }
   38676 #else
   38677 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
   38678   int8x8_t __ret;
   38679   __ret = (int8x8_t)(__p0);
   38680   return __ret;
   38681 }
   38682 #endif
   38683 
   38684 #ifdef __LITTLE_ENDIAN__
   38685 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
   38686   int8x8_t __ret;
   38687   __ret = (int8x8_t)(__p0);
   38688   return __ret;
   38689 }
   38690 #else
   38691 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
   38692   int8x8_t __ret;
   38693   __ret = (int8x8_t)(__p0);
   38694   return __ret;
   38695 }
   38696 #endif
   38697 
   38698 #ifdef __LITTLE_ENDIAN__
   38699 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
   38700   int8x8_t __ret;
   38701   __ret = (int8x8_t)(__p0);
   38702   return __ret;
   38703 }
   38704 #else
   38705 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
   38706   int8x8_t __ret;
   38707   __ret = (int8x8_t)(__p0);
   38708   return __ret;
   38709 }
   38710 #endif
   38711 
   38712 #ifdef __LITTLE_ENDIAN__
   38713 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
   38714   int8x8_t __ret;
   38715   __ret = (int8x8_t)(__p0);
   38716   return __ret;
   38717 }
   38718 #else
   38719 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
   38720   int8x8_t __ret;
   38721   __ret = (int8x8_t)(__p0);
   38722   return __ret;
   38723 }
   38724 #endif
   38725 
   38726 #ifdef __LITTLE_ENDIAN__
   38727 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
   38728   int8x8_t __ret;
   38729   __ret = (int8x8_t)(__p0);
   38730   return __ret;
   38731 }
   38732 #else
   38733 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
   38734   int8x8_t __ret;
   38735   __ret = (int8x8_t)(__p0);
   38736   return __ret;
   38737 }
   38738 #endif
   38739 
   38740 #ifdef __LITTLE_ENDIAN__
   38741 __ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
   38742   float64x1_t __ret;
   38743   __ret = (float64x1_t)(__p0);
   38744   return __ret;
   38745 }
   38746 #else
   38747 __ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
   38748   float64x1_t __ret;
   38749   __ret = (float64x1_t)(__p0);
   38750   return __ret;
   38751 }
   38752 #endif
   38753 
   38754 #ifdef __LITTLE_ENDIAN__
   38755 __ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
   38756   float64x1_t __ret;
   38757   __ret = (float64x1_t)(__p0);
   38758   return __ret;
   38759 }
   38760 #else
   38761 __ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
   38762   float64x1_t __ret;
   38763   __ret = (float64x1_t)(__p0);
   38764   return __ret;
   38765 }
   38766 #endif
   38767 
   38768 #ifdef __LITTLE_ENDIAN__
   38769 __ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
   38770   float64x1_t __ret;
   38771   __ret = (float64x1_t)(__p0);
   38772   return __ret;
   38773 }
   38774 #else
   38775 __ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
   38776   float64x1_t __ret;
   38777   __ret = (float64x1_t)(__p0);
   38778   return __ret;
   38779 }
   38780 #endif
   38781 
   38782 #ifdef __LITTLE_ENDIAN__
   38783 __ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
   38784   float64x1_t __ret;
   38785   __ret = (float64x1_t)(__p0);
   38786   return __ret;
   38787 }
   38788 #else
   38789 __ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
   38790   float64x1_t __ret;
   38791   __ret = (float64x1_t)(__p0);
   38792   return __ret;
   38793 }
   38794 #endif
   38795 
   38796 #ifdef __LITTLE_ENDIAN__
   38797 __ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
   38798   float64x1_t __ret;
   38799   __ret = (float64x1_t)(__p0);
   38800   return __ret;
   38801 }
   38802 #else
   38803 __ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
   38804   float64x1_t __ret;
   38805   __ret = (float64x1_t)(__p0);
   38806   return __ret;
   38807 }
   38808 #endif
   38809 
   38810 #ifdef __LITTLE_ENDIAN__
   38811 __ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
   38812   float64x1_t __ret;
   38813   __ret = (float64x1_t)(__p0);
   38814   return __ret;
   38815 }
   38816 #else
   38817 __ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
   38818   float64x1_t __ret;
   38819   __ret = (float64x1_t)(__p0);
   38820   return __ret;
   38821 }
   38822 #endif
   38823 
   38824 #ifdef __LITTLE_ENDIAN__
   38825 __ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
   38826   float64x1_t __ret;
   38827   __ret = (float64x1_t)(__p0);
   38828   return __ret;
   38829 }
   38830 #else
   38831 __ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
   38832   float64x1_t __ret;
   38833   __ret = (float64x1_t)(__p0);
   38834   return __ret;
   38835 }
   38836 #endif
   38837 
   38838 #ifdef __LITTLE_ENDIAN__
   38839 __ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
   38840   float64x1_t __ret;
   38841   __ret = (float64x1_t)(__p0);
   38842   return __ret;
   38843 }
   38844 #else
   38845 __ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
   38846   float64x1_t __ret;
   38847   __ret = (float64x1_t)(__p0);
   38848   return __ret;
   38849 }
   38850 #endif
   38851 
   38852 #ifdef __LITTLE_ENDIAN__
   38853 __ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
   38854   float64x1_t __ret;
   38855   __ret = (float64x1_t)(__p0);
   38856   return __ret;
   38857 }
   38858 #else
   38859 __ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
   38860   float64x1_t __ret;
   38861   __ret = (float64x1_t)(__p0);
   38862   return __ret;
   38863 }
   38864 #endif
   38865 
   38866 #ifdef __LITTLE_ENDIAN__
   38867 __ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
   38868   float64x1_t __ret;
   38869   __ret = (float64x1_t)(__p0);
   38870   return __ret;
   38871 }
   38872 #else
   38873 __ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
   38874   float64x1_t __ret;
   38875   __ret = (float64x1_t)(__p0);
   38876   return __ret;
   38877 }
   38878 #endif
   38879 
   38880 #ifdef __LITTLE_ENDIAN__
   38881 __ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
   38882   float64x1_t __ret;
   38883   __ret = (float64x1_t)(__p0);
   38884   return __ret;
   38885 }
   38886 #else
   38887 __ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
   38888   float64x1_t __ret;
   38889   __ret = (float64x1_t)(__p0);
   38890   return __ret;
   38891 }
   38892 #endif
   38893 
   38894 #ifdef __LITTLE_ENDIAN__
   38895 __ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
   38896   float64x1_t __ret;
   38897   __ret = (float64x1_t)(__p0);
   38898   return __ret;
   38899 }
   38900 #else
   38901 __ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
   38902   float64x1_t __ret;
   38903   __ret = (float64x1_t)(__p0);
   38904   return __ret;
   38905 }
   38906 #endif
   38907 
   38908 #ifdef __LITTLE_ENDIAN__
   38909 __ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
   38910   float64x1_t __ret;
   38911   __ret = (float64x1_t)(__p0);
   38912   return __ret;
   38913 }
   38914 #else
   38915 __ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
   38916   float64x1_t __ret;
   38917   __ret = (float64x1_t)(__p0);
   38918   return __ret;
   38919 }
   38920 #endif
   38921 
   38922 #ifdef __LITTLE_ENDIAN__
   38923 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
   38924   float32x2_t __ret;
   38925   __ret = (float32x2_t)(__p0);
   38926   return __ret;
   38927 }
   38928 #else
   38929 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
   38930   float32x2_t __ret;
   38931   __ret = (float32x2_t)(__p0);
   38932   return __ret;
   38933 }
   38934 #endif
   38935 
   38936 #ifdef __LITTLE_ENDIAN__
   38937 __ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
   38938   float32x2_t __ret;
   38939   __ret = (float32x2_t)(__p0);
   38940   return __ret;
   38941 }
   38942 #else
   38943 __ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
   38944   float32x2_t __ret;
   38945   __ret = (float32x2_t)(__p0);
   38946   return __ret;
   38947 }
   38948 #endif
   38949 
   38950 #ifdef __LITTLE_ENDIAN__
   38951 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
   38952   float32x2_t __ret;
   38953   __ret = (float32x2_t)(__p0);
   38954   return __ret;
   38955 }
   38956 #else
   38957 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
   38958   float32x2_t __ret;
   38959   __ret = (float32x2_t)(__p0);
   38960   return __ret;
   38961 }
   38962 #endif
   38963 
   38964 #ifdef __LITTLE_ENDIAN__
   38965 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
   38966   float32x2_t __ret;
   38967   __ret = (float32x2_t)(__p0);
   38968   return __ret;
   38969 }
   38970 #else
   38971 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
   38972   float32x2_t __ret;
   38973   __ret = (float32x2_t)(__p0);
   38974   return __ret;
   38975 }
   38976 #endif
   38977 
   38978 #ifdef __LITTLE_ENDIAN__
   38979 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
   38980   float32x2_t __ret;
   38981   __ret = (float32x2_t)(__p0);
   38982   return __ret;
   38983 }
   38984 #else
   38985 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
   38986   float32x2_t __ret;
   38987   __ret = (float32x2_t)(__p0);
   38988   return __ret;
   38989 }
   38990 #endif
   38991 
   38992 #ifdef __LITTLE_ENDIAN__
   38993 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
   38994   float32x2_t __ret;
   38995   __ret = (float32x2_t)(__p0);
   38996   return __ret;
   38997 }
   38998 #else
   38999 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
   39000   float32x2_t __ret;
   39001   __ret = (float32x2_t)(__p0);
   39002   return __ret;
   39003 }
   39004 #endif
   39005 
   39006 #ifdef __LITTLE_ENDIAN__
   39007 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
   39008   float32x2_t __ret;
   39009   __ret = (float32x2_t)(__p0);
   39010   return __ret;
   39011 }
   39012 #else
   39013 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
   39014   float32x2_t __ret;
   39015   __ret = (float32x2_t)(__p0);
   39016   return __ret;
   39017 }
   39018 #endif
   39019 
   39020 #ifdef __LITTLE_ENDIAN__
   39021 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
   39022   float32x2_t __ret;
   39023   __ret = (float32x2_t)(__p0);
   39024   return __ret;
   39025 }
   39026 #else
   39027 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
   39028   float32x2_t __ret;
   39029   __ret = (float32x2_t)(__p0);
   39030   return __ret;
   39031 }
   39032 #endif
   39033 
   39034 #ifdef __LITTLE_ENDIAN__
   39035 __ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
   39036   float32x2_t __ret;
   39037   __ret = (float32x2_t)(__p0);
   39038   return __ret;
   39039 }
   39040 #else
   39041 __ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
   39042   float32x2_t __ret;
   39043   __ret = (float32x2_t)(__p0);
   39044   return __ret;
   39045 }
   39046 #endif
   39047 
   39048 #ifdef __LITTLE_ENDIAN__
   39049 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
   39050   float32x2_t __ret;
   39051   __ret = (float32x2_t)(__p0);
   39052   return __ret;
   39053 }
   39054 #else
   39055 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
   39056   float32x2_t __ret;
   39057   __ret = (float32x2_t)(__p0);
   39058   return __ret;
   39059 }
   39060 #endif
   39061 
   39062 #ifdef __LITTLE_ENDIAN__
   39063 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
   39064   float32x2_t __ret;
   39065   __ret = (float32x2_t)(__p0);
   39066   return __ret;
   39067 }
   39068 #else
   39069 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
   39070   float32x2_t __ret;
   39071   __ret = (float32x2_t)(__p0);
   39072   return __ret;
   39073 }
   39074 #endif
   39075 
   39076 #ifdef __LITTLE_ENDIAN__
   39077 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
   39078   float32x2_t __ret;
   39079   __ret = (float32x2_t)(__p0);
   39080   return __ret;
   39081 }
   39082 #else
   39083 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
   39084   float32x2_t __ret;
   39085   __ret = (float32x2_t)(__p0);
   39086   return __ret;
   39087 }
   39088 #endif
   39089 
   39090 #ifdef __LITTLE_ENDIAN__
   39091 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
   39092   float32x2_t __ret;
   39093   __ret = (float32x2_t)(__p0);
   39094   return __ret;
   39095 }
   39096 #else
   39097 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
   39098   float32x2_t __ret;
   39099   __ret = (float32x2_t)(__p0);
   39100   return __ret;
   39101 }
   39102 #endif
   39103 
   39104 #ifdef __LITTLE_ENDIAN__
   39105 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
   39106   float16x4_t __ret;
   39107   __ret = (float16x4_t)(__p0);
   39108   return __ret;
   39109 }
   39110 #else
   39111 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
   39112   float16x4_t __ret;
   39113   __ret = (float16x4_t)(__p0);
   39114   return __ret;
   39115 }
   39116 #endif
   39117 
   39118 #ifdef __LITTLE_ENDIAN__
   39119 __ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
   39120   float16x4_t __ret;
   39121   __ret = (float16x4_t)(__p0);
   39122   return __ret;
   39123 }
   39124 #else
   39125 __ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
   39126   float16x4_t __ret;
   39127   __ret = (float16x4_t)(__p0);
   39128   return __ret;
   39129 }
   39130 #endif
   39131 
   39132 #ifdef __LITTLE_ENDIAN__
   39133 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
   39134   float16x4_t __ret;
   39135   __ret = (float16x4_t)(__p0);
   39136   return __ret;
   39137 }
   39138 #else
   39139 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
   39140   float16x4_t __ret;
   39141   __ret = (float16x4_t)(__p0);
   39142   return __ret;
   39143 }
   39144 #endif
   39145 
   39146 #ifdef __LITTLE_ENDIAN__
   39147 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
   39148   float16x4_t __ret;
   39149   __ret = (float16x4_t)(__p0);
   39150   return __ret;
   39151 }
   39152 #else
   39153 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
   39154   float16x4_t __ret;
   39155   __ret = (float16x4_t)(__p0);
   39156   return __ret;
   39157 }
   39158 #endif
   39159 
   39160 #ifdef __LITTLE_ENDIAN__
   39161 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
   39162   float16x4_t __ret;
   39163   __ret = (float16x4_t)(__p0);
   39164   return __ret;
   39165 }
   39166 #else
   39167 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
   39168   float16x4_t __ret;
   39169   __ret = (float16x4_t)(__p0);
   39170   return __ret;
   39171 }
   39172 #endif
   39173 
   39174 #ifdef __LITTLE_ENDIAN__
   39175 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
   39176   float16x4_t __ret;
   39177   __ret = (float16x4_t)(__p0);
   39178   return __ret;
   39179 }
   39180 #else
   39181 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
   39182   float16x4_t __ret;
   39183   __ret = (float16x4_t)(__p0);
   39184   return __ret;
   39185 }
   39186 #endif
   39187 
   39188 #ifdef __LITTLE_ENDIAN__
   39189 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
   39190   float16x4_t __ret;
   39191   __ret = (float16x4_t)(__p0);
   39192   return __ret;
   39193 }
   39194 #else
   39195 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
   39196   float16x4_t __ret;
   39197   __ret = (float16x4_t)(__p0);
   39198   return __ret;
   39199 }
   39200 #endif
   39201 
   39202 #ifdef __LITTLE_ENDIAN__
   39203 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
   39204   float16x4_t __ret;
   39205   __ret = (float16x4_t)(__p0);
   39206   return __ret;
   39207 }
   39208 #else
   39209 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
   39210   float16x4_t __ret;
   39211   __ret = (float16x4_t)(__p0);
   39212   return __ret;
   39213 }
   39214 #endif
   39215 
   39216 #ifdef __LITTLE_ENDIAN__
   39217 __ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
   39218   float16x4_t __ret;
   39219   __ret = (float16x4_t)(__p0);
   39220   return __ret;
   39221 }
   39222 #else
   39223 __ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
   39224   float16x4_t __ret;
   39225   __ret = (float16x4_t)(__p0);
   39226   return __ret;
   39227 }
   39228 #endif
   39229 
   39230 #ifdef __LITTLE_ENDIAN__
   39231 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
   39232   float16x4_t __ret;
   39233   __ret = (float16x4_t)(__p0);
   39234   return __ret;
   39235 }
   39236 #else
   39237 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
   39238   float16x4_t __ret;
   39239   __ret = (float16x4_t)(__p0);
   39240   return __ret;
   39241 }
   39242 #endif
   39243 
   39244 #ifdef __LITTLE_ENDIAN__
   39245 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
   39246   float16x4_t __ret;
   39247   __ret = (float16x4_t)(__p0);
   39248   return __ret;
   39249 }
   39250 #else
   39251 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
   39252   float16x4_t __ret;
   39253   __ret = (float16x4_t)(__p0);
   39254   return __ret;
   39255 }
   39256 #endif
   39257 
   39258 #ifdef __LITTLE_ENDIAN__
   39259 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
   39260   float16x4_t __ret;
   39261   __ret = (float16x4_t)(__p0);
   39262   return __ret;
   39263 }
   39264 #else
   39265 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
   39266   float16x4_t __ret;
   39267   __ret = (float16x4_t)(__p0);
   39268   return __ret;
   39269 }
   39270 #endif
   39271 
   39272 #ifdef __LITTLE_ENDIAN__
   39273 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
   39274   float16x4_t __ret;
   39275   __ret = (float16x4_t)(__p0);
   39276   return __ret;
   39277 }
   39278 #else
   39279 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
   39280   float16x4_t __ret;
   39281   __ret = (float16x4_t)(__p0);
   39282   return __ret;
   39283 }
   39284 #endif
   39285 
   39286 #ifdef __LITTLE_ENDIAN__
   39287 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
   39288   int32x2_t __ret;
   39289   __ret = (int32x2_t)(__p0);
   39290   return __ret;
   39291 }
   39292 #else
   39293 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
   39294   int32x2_t __ret;
   39295   __ret = (int32x2_t)(__p0);
   39296   return __ret;
   39297 }
   39298 #endif
   39299 
   39300 #ifdef __LITTLE_ENDIAN__
   39301 __ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
   39302   int32x2_t __ret;
   39303   __ret = (int32x2_t)(__p0);
   39304   return __ret;
   39305 }
   39306 #else
   39307 __ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
   39308   int32x2_t __ret;
   39309   __ret = (int32x2_t)(__p0);
   39310   return __ret;
   39311 }
   39312 #endif
   39313 
   39314 #ifdef __LITTLE_ENDIAN__
   39315 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
   39316   int32x2_t __ret;
   39317   __ret = (int32x2_t)(__p0);
   39318   return __ret;
   39319 }
   39320 #else
   39321 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
   39322   int32x2_t __ret;
   39323   __ret = (int32x2_t)(__p0);
   39324   return __ret;
   39325 }
   39326 #endif
   39327 
   39328 #ifdef __LITTLE_ENDIAN__
   39329 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
   39330   int32x2_t __ret;
   39331   __ret = (int32x2_t)(__p0);
   39332   return __ret;
   39333 }
   39334 #else
   39335 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
   39336   int32x2_t __ret;
   39337   __ret = (int32x2_t)(__p0);
   39338   return __ret;
   39339 }
   39340 #endif
   39341 
   39342 #ifdef __LITTLE_ENDIAN__
   39343 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
   39344   int32x2_t __ret;
   39345   __ret = (int32x2_t)(__p0);
   39346   return __ret;
   39347 }
   39348 #else
   39349 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
   39350   int32x2_t __ret;
   39351   __ret = (int32x2_t)(__p0);
   39352   return __ret;
   39353 }
   39354 #endif
   39355 
   39356 #ifdef __LITTLE_ENDIAN__
   39357 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
   39358   int32x2_t __ret;
   39359   __ret = (int32x2_t)(__p0);
   39360   return __ret;
   39361 }
   39362 #else
   39363 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
   39364   int32x2_t __ret;
   39365   __ret = (int32x2_t)(__p0);
   39366   return __ret;
   39367 }
   39368 #endif
   39369 
   39370 #ifdef __LITTLE_ENDIAN__
   39371 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
   39372   int32x2_t __ret;
   39373   __ret = (int32x2_t)(__p0);
   39374   return __ret;
   39375 }
   39376 #else
   39377 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
   39378   int32x2_t __ret;
   39379   __ret = (int32x2_t)(__p0);
   39380   return __ret;
   39381 }
   39382 #endif
   39383 
   39384 #ifdef __LITTLE_ENDIAN__
   39385 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
   39386   int32x2_t __ret;
   39387   __ret = (int32x2_t)(__p0);
   39388   return __ret;
   39389 }
   39390 #else
   39391 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
   39392   int32x2_t __ret;
   39393   __ret = (int32x2_t)(__p0);
   39394   return __ret;
   39395 }
   39396 #endif
   39397 
   39398 #ifdef __LITTLE_ENDIAN__
   39399 __ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
   39400   int32x2_t __ret;
   39401   __ret = (int32x2_t)(__p0);
   39402   return __ret;
   39403 }
   39404 #else
   39405 __ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
   39406   int32x2_t __ret;
   39407   __ret = (int32x2_t)(__p0);
   39408   return __ret;
   39409 }
   39410 #endif
   39411 
   39412 #ifdef __LITTLE_ENDIAN__
   39413 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
   39414   int32x2_t __ret;
   39415   __ret = (int32x2_t)(__p0);
   39416   return __ret;
   39417 }
   39418 #else
   39419 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
   39420   int32x2_t __ret;
   39421   __ret = (int32x2_t)(__p0);
   39422   return __ret;
   39423 }
   39424 #endif
   39425 
   39426 #ifdef __LITTLE_ENDIAN__
   39427 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
   39428   int32x2_t __ret;
   39429   __ret = (int32x2_t)(__p0);
   39430   return __ret;
   39431 }
   39432 #else
   39433 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
   39434   int32x2_t __ret;
   39435   __ret = (int32x2_t)(__p0);
   39436   return __ret;
   39437 }
   39438 #endif
   39439 
   39440 #ifdef __LITTLE_ENDIAN__
   39441 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
   39442   int32x2_t __ret;
   39443   __ret = (int32x2_t)(__p0);
   39444   return __ret;
   39445 }
   39446 #else
   39447 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
   39448   int32x2_t __ret;
   39449   __ret = (int32x2_t)(__p0);
   39450   return __ret;
   39451 }
   39452 #endif
   39453 
   39454 #ifdef __LITTLE_ENDIAN__
   39455 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
   39456   int32x2_t __ret;
   39457   __ret = (int32x2_t)(__p0);
   39458   return __ret;
   39459 }
   39460 #else
   39461 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
   39462   int32x2_t __ret;
   39463   __ret = (int32x2_t)(__p0);
   39464   return __ret;
   39465 }
   39466 #endif
   39467 
   39468 #ifdef __LITTLE_ENDIAN__
   39469 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
   39470   int64x1_t __ret;
   39471   __ret = (int64x1_t)(__p0);
   39472   return __ret;
   39473 }
   39474 #else
   39475 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
   39476   int64x1_t __ret;
   39477   __ret = (int64x1_t)(__p0);
   39478   return __ret;
   39479 }
   39480 #endif
   39481 
   39482 #ifdef __LITTLE_ENDIAN__
   39483 __ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
   39484   int64x1_t __ret;
   39485   __ret = (int64x1_t)(__p0);
   39486   return __ret;
   39487 }
   39488 #else
   39489 __ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
   39490   int64x1_t __ret;
   39491   __ret = (int64x1_t)(__p0);
   39492   return __ret;
   39493 }
   39494 #endif
   39495 
   39496 #ifdef __LITTLE_ENDIAN__
   39497 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
   39498   int64x1_t __ret;
   39499   __ret = (int64x1_t)(__p0);
   39500   return __ret;
   39501 }
   39502 #else
   39503 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
   39504   int64x1_t __ret;
   39505   __ret = (int64x1_t)(__p0);
   39506   return __ret;
   39507 }
   39508 #endif
   39509 
   39510 #ifdef __LITTLE_ENDIAN__
   39511 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
   39512   int64x1_t __ret;
   39513   __ret = (int64x1_t)(__p0);
   39514   return __ret;
   39515 }
   39516 #else
   39517 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
   39518   int64x1_t __ret;
   39519   __ret = (int64x1_t)(__p0);
   39520   return __ret;
   39521 }
   39522 #endif
   39523 
   39524 #ifdef __LITTLE_ENDIAN__
   39525 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
   39526   int64x1_t __ret;
   39527   __ret = (int64x1_t)(__p0);
   39528   return __ret;
   39529 }
   39530 #else
   39531 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
   39532   int64x1_t __ret;
   39533   __ret = (int64x1_t)(__p0);
   39534   return __ret;
   39535 }
   39536 #endif
   39537 
   39538 #ifdef __LITTLE_ENDIAN__
   39539 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
   39540   int64x1_t __ret;
   39541   __ret = (int64x1_t)(__p0);
   39542   return __ret;
   39543 }
   39544 #else
   39545 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
   39546   int64x1_t __ret;
   39547   __ret = (int64x1_t)(__p0);
   39548   return __ret;
   39549 }
   39550 #endif
   39551 
   39552 #ifdef __LITTLE_ENDIAN__
   39553 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
   39554   int64x1_t __ret;
   39555   __ret = (int64x1_t)(__p0);
   39556   return __ret;
   39557 }
   39558 #else
   39559 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
   39560   int64x1_t __ret;
   39561   __ret = (int64x1_t)(__p0);
   39562   return __ret;
   39563 }
   39564 #endif
   39565 
   39566 #ifdef __LITTLE_ENDIAN__
   39567 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
   39568   int64x1_t __ret;
   39569   __ret = (int64x1_t)(__p0);
   39570   return __ret;
   39571 }
   39572 #else
   39573 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
   39574   int64x1_t __ret;
   39575   __ret = (int64x1_t)(__p0);
   39576   return __ret;
   39577 }
   39578 #endif
   39579 
   39580 #ifdef __LITTLE_ENDIAN__
   39581 __ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
   39582   int64x1_t __ret;
   39583   __ret = (int64x1_t)(__p0);
   39584   return __ret;
   39585 }
   39586 #else
   39587 __ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
   39588   int64x1_t __ret;
   39589   __ret = (int64x1_t)(__p0);
   39590   return __ret;
   39591 }
   39592 #endif
   39593 
   39594 #ifdef __LITTLE_ENDIAN__
   39595 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
   39596   int64x1_t __ret;
   39597   __ret = (int64x1_t)(__p0);
   39598   return __ret;
   39599 }
   39600 #else
   39601 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
   39602   int64x1_t __ret;
   39603   __ret = (int64x1_t)(__p0);
   39604   return __ret;
   39605 }
   39606 #endif
   39607 
   39608 #ifdef __LITTLE_ENDIAN__
   39609 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
   39610   int64x1_t __ret;
   39611   __ret = (int64x1_t)(__p0);
   39612   return __ret;
   39613 }
   39614 #else
   39615 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
   39616   int64x1_t __ret;
   39617   __ret = (int64x1_t)(__p0);
   39618   return __ret;
   39619 }
   39620 #endif
   39621 
   39622 #ifdef __LITTLE_ENDIAN__
   39623 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
   39624   int64x1_t __ret;
   39625   __ret = (int64x1_t)(__p0);
   39626   return __ret;
   39627 }
   39628 #else
   39629 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
   39630   int64x1_t __ret;
   39631   __ret = (int64x1_t)(__p0);
   39632   return __ret;
   39633 }
   39634 #endif
   39635 
   39636 #ifdef __LITTLE_ENDIAN__
   39637 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
   39638   int64x1_t __ret;
   39639   __ret = (int64x1_t)(__p0);
   39640   return __ret;
   39641 }
   39642 #else
   39643 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
   39644   int64x1_t __ret;
   39645   __ret = (int64x1_t)(__p0);
   39646   return __ret;
   39647 }
   39648 #endif
   39649 
   39650 #ifdef __LITTLE_ENDIAN__
   39651 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
   39652   int16x4_t __ret;
   39653   __ret = (int16x4_t)(__p0);
   39654   return __ret;
   39655 }
   39656 #else
   39657 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
   39658   int16x4_t __ret;
   39659   __ret = (int16x4_t)(__p0);
   39660   return __ret;
   39661 }
   39662 #endif
   39663 
   39664 #ifdef __LITTLE_ENDIAN__
   39665 __ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
   39666   int16x4_t __ret;
   39667   __ret = (int16x4_t)(__p0);
   39668   return __ret;
   39669 }
   39670 #else
   39671 __ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
   39672   int16x4_t __ret;
   39673   __ret = (int16x4_t)(__p0);
   39674   return __ret;
   39675 }
   39676 #endif
   39677 
   39678 #ifdef __LITTLE_ENDIAN__
   39679 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
   39680   int16x4_t __ret;
   39681   __ret = (int16x4_t)(__p0);
   39682   return __ret;
   39683 }
   39684 #else
   39685 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
   39686   int16x4_t __ret;
   39687   __ret = (int16x4_t)(__p0);
   39688   return __ret;
   39689 }
   39690 #endif
   39691 
   39692 #ifdef __LITTLE_ENDIAN__
   39693 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
   39694   int16x4_t __ret;
   39695   __ret = (int16x4_t)(__p0);
   39696   return __ret;
   39697 }
   39698 #else
   39699 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
   39700   int16x4_t __ret;
   39701   __ret = (int16x4_t)(__p0);
   39702   return __ret;
   39703 }
   39704 #endif
   39705 
   39706 #ifdef __LITTLE_ENDIAN__
   39707 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
   39708   int16x4_t __ret;
   39709   __ret = (int16x4_t)(__p0);
   39710   return __ret;
   39711 }
   39712 #else
   39713 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
   39714   int16x4_t __ret;
   39715   __ret = (int16x4_t)(__p0);
   39716   return __ret;
   39717 }
   39718 #endif
   39719 
   39720 #ifdef __LITTLE_ENDIAN__
   39721 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
   39722   int16x4_t __ret;
   39723   __ret = (int16x4_t)(__p0);
   39724   return __ret;
   39725 }
   39726 #else
   39727 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
   39728   int16x4_t __ret;
   39729   __ret = (int16x4_t)(__p0);
   39730   return __ret;
   39731 }
   39732 #endif
   39733 
   39734 #ifdef __LITTLE_ENDIAN__
   39735 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
   39736   int16x4_t __ret;
   39737   __ret = (int16x4_t)(__p0);
   39738   return __ret;
   39739 }
   39740 #else
   39741 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
   39742   int16x4_t __ret;
   39743   __ret = (int16x4_t)(__p0);
   39744   return __ret;
   39745 }
   39746 #endif
   39747 
   39748 #ifdef __LITTLE_ENDIAN__
   39749 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
   39750   int16x4_t __ret;
   39751   __ret = (int16x4_t)(__p0);
   39752   return __ret;
   39753 }
   39754 #else
   39755 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
   39756   int16x4_t __ret;
   39757   __ret = (int16x4_t)(__p0);
   39758   return __ret;
   39759 }
   39760 #endif
   39761 
   39762 #ifdef __LITTLE_ENDIAN__
   39763 __ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
   39764   int16x4_t __ret;
   39765   __ret = (int16x4_t)(__p0);
   39766   return __ret;
   39767 }
   39768 #else
   39769 __ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
   39770   int16x4_t __ret;
   39771   __ret = (int16x4_t)(__p0);
   39772   return __ret;
   39773 }
   39774 #endif
   39775 
   39776 #ifdef __LITTLE_ENDIAN__
   39777 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
   39778   int16x4_t __ret;
   39779   __ret = (int16x4_t)(__p0);
   39780   return __ret;
   39781 }
   39782 #else
   39783 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
   39784   int16x4_t __ret;
   39785   __ret = (int16x4_t)(__p0);
   39786   return __ret;
   39787 }
   39788 #endif
   39789 
   39790 #ifdef __LITTLE_ENDIAN__
   39791 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
   39792   int16x4_t __ret;
   39793   __ret = (int16x4_t)(__p0);
   39794   return __ret;
   39795 }
   39796 #else
   39797 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
   39798   int16x4_t __ret;
   39799   __ret = (int16x4_t)(__p0);
   39800   return __ret;
   39801 }
   39802 #endif
   39803 
   39804 #ifdef __LITTLE_ENDIAN__
   39805 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
   39806   int16x4_t __ret;
   39807   __ret = (int16x4_t)(__p0);
   39808   return __ret;
   39809 }
   39810 #else
   39811 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
   39812   int16x4_t __ret;
   39813   __ret = (int16x4_t)(__p0);
   39814   return __ret;
   39815 }
   39816 #endif
   39817 
   39818 #ifdef __LITTLE_ENDIAN__
   39819 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
   39820   int16x4_t __ret;
   39821   __ret = (int16x4_t)(__p0);
   39822   return __ret;
   39823 }
   39824 #else
   39825 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
   39826   int16x4_t __ret;
   39827   __ret = (int16x4_t)(__p0);
   39828   return __ret;
   39829 }
   39830 #endif
   39831 
   39832 #endif
   39833 #if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
   39834 #ifdef __LITTLE_ENDIAN__
   39835 __ai float64x2_t vrndq_f64(float64x2_t __p0) {
   39836   float64x2_t __ret;
   39837   __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 42);
   39838   return __ret;
   39839 }
   39840 #else
   39841 __ai float64x2_t vrndq_f64(float64x2_t __p0) {
   39842   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   39843   float64x2_t __ret;
   39844   __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 42);
   39845   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   39846   return __ret;
   39847 }
   39848 #endif
   39849 
   39850 #ifdef __LITTLE_ENDIAN__
   39851 __ai float64x1_t vrnd_f64(float64x1_t __p0) {
   39852   float64x1_t __ret;
   39853   __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
   39854   return __ret;
   39855 }
   39856 #else
   39857 __ai float64x1_t vrnd_f64(float64x1_t __p0) {
   39858   float64x1_t __ret;
   39859   __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
   39860   return __ret;
   39861 }
   39862 #endif
   39863 
   39864 #ifdef __LITTLE_ENDIAN__
   39865 __ai float64x2_t vrndaq_f64(float64x2_t __p0) {
   39866   float64x2_t __ret;
   39867   __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 42);
   39868   return __ret;
   39869 }
   39870 #else
   39871 __ai float64x2_t vrndaq_f64(float64x2_t __p0) {
   39872   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   39873   float64x2_t __ret;
   39874   __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 42);
   39875   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   39876   return __ret;
   39877 }
   39878 #endif
   39879 
   39880 #ifdef __LITTLE_ENDIAN__
   39881 __ai float64x1_t vrnda_f64(float64x1_t __p0) {
   39882   float64x1_t __ret;
   39883   __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
   39884   return __ret;
   39885 }
   39886 #else
   39887 __ai float64x1_t vrnda_f64(float64x1_t __p0) {
   39888   float64x1_t __ret;
   39889   __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
   39890   return __ret;
   39891 }
   39892 #endif
   39893 
   39894 #ifdef __LITTLE_ENDIAN__
   39895 __ai float64x2_t vrndiq_f64(float64x2_t __p0) {
   39896   float64x2_t __ret;
   39897   __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 42);
   39898   return __ret;
   39899 }
   39900 #else
   39901 __ai float64x2_t vrndiq_f64(float64x2_t __p0) {
   39902   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   39903   float64x2_t __ret;
   39904   __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 42);
   39905   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   39906   return __ret;
   39907 }
   39908 #endif
   39909 
   39910 #ifdef __LITTLE_ENDIAN__
   39911 __ai float32x4_t vrndiq_f32(float32x4_t __p0) {
   39912   float32x4_t __ret;
   39913   __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 41);
   39914   return __ret;
   39915 }
   39916 #else
   39917 __ai float32x4_t vrndiq_f32(float32x4_t __p0) {
   39918   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   39919   float32x4_t __ret;
   39920   __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 41);
   39921   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   39922   return __ret;
   39923 }
   39924 #endif
   39925 
   39926 #ifdef __LITTLE_ENDIAN__
   39927 __ai float64x1_t vrndi_f64(float64x1_t __p0) {
   39928   float64x1_t __ret;
   39929   __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
   39930   return __ret;
   39931 }
   39932 #else
   39933 __ai float64x1_t vrndi_f64(float64x1_t __p0) {
   39934   float64x1_t __ret;
   39935   __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
   39936   return __ret;
   39937 }
   39938 #endif
   39939 
   39940 #ifdef __LITTLE_ENDIAN__
   39941 __ai float32x2_t vrndi_f32(float32x2_t __p0) {
   39942   float32x2_t __ret;
   39943   __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 9);
   39944   return __ret;
   39945 }
   39946 #else
   39947 __ai float32x2_t vrndi_f32(float32x2_t __p0) {
   39948   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   39949   float32x2_t __ret;
   39950   __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__rev0, 9);
   39951   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   39952   return __ret;
   39953 }
   39954 #endif
   39955 
   39956 #ifdef __LITTLE_ENDIAN__
   39957 __ai float64x2_t vrndmq_f64(float64x2_t __p0) {
   39958   float64x2_t __ret;
   39959   __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 42);
   39960   return __ret;
   39961 }
   39962 #else
   39963 __ai float64x2_t vrndmq_f64(float64x2_t __p0) {
   39964   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   39965   float64x2_t __ret;
   39966   __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 42);
   39967   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   39968   return __ret;
   39969 }
   39970 #endif
   39971 
   39972 #ifdef __LITTLE_ENDIAN__
   39973 __ai float64x1_t vrndm_f64(float64x1_t __p0) {
   39974   float64x1_t __ret;
   39975   __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
   39976   return __ret;
   39977 }
   39978 #else
   39979 __ai float64x1_t vrndm_f64(float64x1_t __p0) {
   39980   float64x1_t __ret;
   39981   __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
   39982   return __ret;
   39983 }
   39984 #endif
   39985 
   39986 #ifdef __LITTLE_ENDIAN__
   39987 __ai float64x2_t vrndnq_f64(float64x2_t __p0) {
   39988   float64x2_t __ret;
   39989   __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 42);
   39990   return __ret;
   39991 }
   39992 #else
   39993 __ai float64x2_t vrndnq_f64(float64x2_t __p0) {
   39994   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   39995   float64x2_t __ret;
   39996   __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 42);
   39997   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   39998   return __ret;
   39999 }
   40000 #endif
   40001 
   40002 #ifdef __LITTLE_ENDIAN__
   40003 __ai float64x1_t vrndn_f64(float64x1_t __p0) {
   40004   float64x1_t __ret;
   40005   __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
   40006   return __ret;
   40007 }
   40008 #else
   40009 __ai float64x1_t vrndn_f64(float64x1_t __p0) {
   40010   float64x1_t __ret;
   40011   __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
   40012   return __ret;
   40013 }
   40014 #endif
   40015 
   40016 #ifdef __LITTLE_ENDIAN__
   40017 __ai float64x2_t vrndpq_f64(float64x2_t __p0) {
   40018   float64x2_t __ret;
   40019   __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 42);
   40020   return __ret;
   40021 }
   40022 #else
   40023 __ai float64x2_t vrndpq_f64(float64x2_t __p0) {
   40024   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   40025   float64x2_t __ret;
   40026   __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 42);
   40027   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   40028   return __ret;
   40029 }
   40030 #endif
   40031 
   40032 #ifdef __LITTLE_ENDIAN__
   40033 __ai float64x1_t vrndp_f64(float64x1_t __p0) {
   40034   float64x1_t __ret;
   40035   __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
   40036   return __ret;
   40037 }
   40038 #else
   40039 __ai float64x1_t vrndp_f64(float64x1_t __p0) {
   40040   float64x1_t __ret;
   40041   __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
   40042   return __ret;
   40043 }
   40044 #endif
   40045 
   40046 #ifdef __LITTLE_ENDIAN__
   40047 __ai float64x2_t vrndxq_f64(float64x2_t __p0) {
   40048   float64x2_t __ret;
   40049   __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 42);
   40050   return __ret;
   40051 }
   40052 #else
   40053 __ai float64x2_t vrndxq_f64(float64x2_t __p0) {
   40054   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   40055   float64x2_t __ret;
   40056   __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 42);
   40057   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   40058   return __ret;
   40059 }
   40060 #endif
   40061 
   40062 #ifdef __LITTLE_ENDIAN__
   40063 __ai float64x1_t vrndx_f64(float64x1_t __p0) {
   40064   float64x1_t __ret;
   40065   __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
   40066   return __ret;
   40067 }
   40068 #else
   40069 __ai float64x1_t vrndx_f64(float64x1_t __p0) {
   40070   float64x1_t __ret;
   40071   __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
   40072   return __ret;
   40073 }
   40074 #endif
   40075 
   40076 #endif
   40077 #if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
   40078 #ifdef __LITTLE_ENDIAN__
   40079 __ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
   40080   float64x2_t __ret;
   40081   __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   40082   return __ret;
   40083 }
   40084 #else
   40085 __ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
   40086   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   40087   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   40088   float64x2_t __ret;
   40089   __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   40090   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   40091   return __ret;
   40092 }
   40093 #endif
   40094 
   40095 #ifdef __LITTLE_ENDIAN__
   40096 __ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
   40097   float64x1_t __ret;
   40098   __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   40099   return __ret;
   40100 }
   40101 #else
   40102 __ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
   40103   float64x1_t __ret;
   40104   __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   40105   return __ret;
   40106 }
   40107 #endif
   40108 
   40109 #ifdef __LITTLE_ENDIAN__
   40110 __ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
   40111   float64x2_t __ret;
   40112   __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   40113   return __ret;
   40114 }
   40115 #else
   40116 __ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
   40117   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   40118   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   40119   float64x2_t __ret;
   40120   __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   40121   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   40122   return __ret;
   40123 }
   40124 #endif
   40125 
   40126 #ifdef __LITTLE_ENDIAN__
   40127 __ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
   40128   float64x1_t __ret;
   40129   __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   40130   return __ret;
   40131 }
   40132 #else
   40133 __ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
   40134   float64x1_t __ret;
   40135   __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   40136   return __ret;
   40137 }
   40138 #endif
   40139 
   40140 #endif
   40141 #if __ARM_FEATURE_CRYPTO
   40142 #ifdef __LITTLE_ENDIAN__
   40143 __ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   40144   uint8x16_t __ret;
   40145   __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   40146   return __ret;
   40147 }
   40148 #else
   40149 __ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   40150   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40151   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40152   uint8x16_t __ret;
   40153   __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   40154   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40155   return __ret;
   40156 }
   40157 #endif
   40158 
   40159 #ifdef __LITTLE_ENDIAN__
   40160 __ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   40161   uint8x16_t __ret;
   40162   __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   40163   return __ret;
   40164 }
   40165 #else
   40166 __ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   40167   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40168   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40169   uint8x16_t __ret;
   40170   __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   40171   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40172   return __ret;
   40173 }
   40174 #endif
   40175 
   40176 #ifdef __LITTLE_ENDIAN__
   40177 __ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
   40178   uint8x16_t __ret;
   40179   __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__p0, 48);
   40180   return __ret;
   40181 }
   40182 #else
   40183 __ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
   40184   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40185   uint8x16_t __ret;
   40186   __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__rev0, 48);
   40187   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40188   return __ret;
   40189 }
   40190 #endif
   40191 
   40192 #ifdef __LITTLE_ENDIAN__
   40193 __ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
   40194   uint8x16_t __ret;
   40195   __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__p0, 48);
   40196   return __ret;
   40197 }
   40198 #else
   40199 __ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
   40200   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40201   uint8x16_t __ret;
   40202   __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__rev0, 48);
   40203   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   40204   return __ret;
   40205 }
   40206 #endif
   40207 
   40208 #ifdef __LITTLE_ENDIAN__
   40209 __ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
   40210   uint32x4_t __ret;
   40211   __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
   40212   return __ret;
   40213 }
   40214 #else
   40215 __ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
   40216   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40217   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40218   uint32x4_t __ret;
   40219   __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
   40220   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40221   return __ret;
   40222 }
   40223 #endif
   40224 
   40225 #ifdef __LITTLE_ENDIAN__
   40226 __ai uint32_t vsha1h_u32(uint32_t __p0) {
   40227   uint32_t __ret;
   40228   __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
   40229   return __ret;
   40230 }
   40231 #else
   40232 __ai uint32_t vsha1h_u32(uint32_t __p0) {
   40233   uint32_t __ret;
   40234   __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
   40235   return __ret;
   40236 }
   40237 #endif
   40238 
   40239 #ifdef __LITTLE_ENDIAN__
   40240 __ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
   40241   uint32x4_t __ret;
   40242   __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
   40243   return __ret;
   40244 }
   40245 #else
   40246 __ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
   40247   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40248   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40249   uint32x4_t __ret;
   40250   __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
   40251   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40252   return __ret;
   40253 }
   40254 #endif
   40255 
   40256 #ifdef __LITTLE_ENDIAN__
   40257 __ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
   40258   uint32x4_t __ret;
   40259   __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
   40260   return __ret;
   40261 }
   40262 #else
   40263 __ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
   40264   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40265   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40266   uint32x4_t __ret;
   40267   __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
   40268   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40269   return __ret;
   40270 }
   40271 #endif
   40272 
   40273 #ifdef __LITTLE_ENDIAN__
   40274 __ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   40275   uint32x4_t __ret;
   40276   __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
   40277   return __ret;
   40278 }
   40279 #else
   40280 __ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   40281   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40282   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40283   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40284   uint32x4_t __ret;
   40285   __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
   40286   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40287   return __ret;
   40288 }
   40289 #endif
   40290 
   40291 #ifdef __LITTLE_ENDIAN__
   40292 __ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   40293   uint32x4_t __ret;
   40294   __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   40295   return __ret;
   40296 }
   40297 #else
   40298 __ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   40299   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40300   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40301   uint32x4_t __ret;
   40302   __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   40303   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40304   return __ret;
   40305 }
   40306 #endif
   40307 
   40308 #ifdef __LITTLE_ENDIAN__
   40309 __ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   40310   uint32x4_t __ret;
   40311   __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
   40312   return __ret;
   40313 }
   40314 #else
   40315 __ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   40316   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40317   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40318   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40319   uint32x4_t __ret;
   40320   __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
   40321   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40322   return __ret;
   40323 }
   40324 #endif
   40325 
   40326 #ifdef __LITTLE_ENDIAN__
   40327 __ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   40328   uint32x4_t __ret;
   40329   __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
   40330   return __ret;
   40331 }
   40332 #else
   40333 __ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   40334   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40335   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40336   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40337   uint32x4_t __ret;
   40338   __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
   40339   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40340   return __ret;
   40341 }
   40342 #endif
   40343 
   40344 #ifdef __LITTLE_ENDIAN__
   40345 __ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   40346   uint32x4_t __ret;
   40347   __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   40348   return __ret;
   40349 }
   40350 #else
   40351 __ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   40352   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40353   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40354   uint32x4_t __ret;
   40355   __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   40356   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40357   return __ret;
   40358 }
   40359 #endif
   40360 
   40361 #ifdef __LITTLE_ENDIAN__
   40362 __ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   40363   uint32x4_t __ret;
   40364   __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
   40365   return __ret;
   40366 }
   40367 #else
   40368 __ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   40369   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40370   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40371   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40372   uint32x4_t __ret;
   40373   __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
   40374   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40375   return __ret;
   40376 }
   40377 #endif
   40378 
   40379 #endif
   40380 #if defined(__ARM_FEATURE_FMA)
   40381 #ifdef __LITTLE_ENDIAN__
   40382 __ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   40383   float32x4_t __ret;
   40384   __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
   40385   return __ret;
   40386 }
   40387 #else
   40388 __ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   40389   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40390   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40391   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40392   float32x4_t __ret;
   40393   __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
   40394   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40395   return __ret;
   40396 }
   40397 __ai float32x4_t __noswap_vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   40398   float32x4_t __ret;
   40399   __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
   40400   return __ret;
   40401 }
   40402 #endif
   40403 
   40404 #ifdef __LITTLE_ENDIAN__
   40405 __ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   40406   float32x2_t __ret;
   40407   __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
   40408   return __ret;
   40409 }
   40410 #else
   40411 __ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   40412   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   40413   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   40414   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   40415   float32x2_t __ret;
   40416   __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
   40417   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   40418   return __ret;
   40419 }
   40420 __ai float32x2_t __noswap_vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   40421   float32x2_t __ret;
   40422   __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
   40423   return __ret;
   40424 }
   40425 #endif
   40426 
   40427 #ifdef __LITTLE_ENDIAN__
   40428 __ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   40429   float32x4_t __ret;
   40430   __ret = vfmaq_f32(__p0, -__p1, __p2);
   40431   return __ret;
   40432 }
   40433 #else
   40434 __ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
   40435   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40436   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40437   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40438   float32x4_t __ret;
   40439   __ret = __noswap_vfmaq_f32(__rev0, -__rev1, __rev2);
   40440   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40441   return __ret;
   40442 }
   40443 #endif
   40444 
   40445 #ifdef __LITTLE_ENDIAN__
   40446 __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   40447   float32x2_t __ret;
   40448   __ret = vfma_f32(__p0, -__p1, __p2);
   40449   return __ret;
   40450 }
   40451 #else
   40452 __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
   40453   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   40454   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   40455   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   40456   float32x2_t __ret;
   40457   __ret = __noswap_vfma_f32(__rev0, -__rev1, __rev2);
   40458   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   40459   return __ret;
   40460 }
   40461 #endif
   40462 
   40463 #endif
   40464 #if defined(__ARM_FEATURE_QRDMX)
   40465 #ifdef __LITTLE_ENDIAN__
   40466 __ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   40467   int32x4_t __ret;
   40468   __ret = vqaddq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
   40469   return __ret;
   40470 }
   40471 #else
   40472 __ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   40473   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40474   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40475   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40476   int32x4_t __ret;
   40477   __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
   40478   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40479   return __ret;
   40480 }
   40481 #endif
   40482 
   40483 #ifdef __LITTLE_ENDIAN__
   40484 __ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   40485   int16x8_t __ret;
   40486   __ret = vqaddq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
   40487   return __ret;
   40488 }
   40489 #else
   40490 __ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   40491   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   40492   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   40493   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   40494   int16x8_t __ret;
   40495   __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
   40496   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   40497   return __ret;
   40498 }
   40499 #endif
   40500 
   40501 #ifdef __LITTLE_ENDIAN__
   40502 __ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   40503   int32x2_t __ret;
   40504   __ret = vqadd_s32(__p0, vqrdmulh_s32(__p1, __p2));
   40505   return __ret;
   40506 }
   40507 #else
   40508 __ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   40509   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   40510   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   40511   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   40512   int32x2_t __ret;
   40513   __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
   40514   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   40515   return __ret;
   40516 }
   40517 #endif
   40518 
   40519 #ifdef __LITTLE_ENDIAN__
   40520 __ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   40521   int16x4_t __ret;
   40522   __ret = vqadd_s16(__p0, vqrdmulh_s16(__p1, __p2));
   40523   return __ret;
   40524 }
   40525 #else
   40526 __ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   40527   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40528   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40529   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40530   int16x4_t __ret;
   40531   __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
   40532   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40533   return __ret;
   40534 }
   40535 #endif
   40536 
   40537 #ifdef __LITTLE_ENDIAN__
   40538 #define vqrdmlahq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40539   int32x4_t __s0 = __p0; \
   40540   int32x4_t __s1 = __p1; \
   40541   int32x2_t __s2 = __p2; \
   40542   int32x4_t __ret; \
   40543   __ret = vqaddq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
   40544   __ret; \
   40545 })
   40546 #else
   40547 #define vqrdmlahq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40548   int32x4_t __s0 = __p0; \
   40549   int32x4_t __s1 = __p1; \
   40550   int32x2_t __s2 = __p2; \
   40551   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   40552   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   40553   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   40554   int32x4_t __ret; \
   40555   __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
   40556   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   40557   __ret; \
   40558 })
   40559 #endif
   40560 
   40561 #ifdef __LITTLE_ENDIAN__
   40562 #define vqrdmlahq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40563   int16x8_t __s0 = __p0; \
   40564   int16x8_t __s1 = __p1; \
   40565   int16x4_t __s2 = __p2; \
   40566   int16x8_t __ret; \
   40567   __ret = vqaddq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
   40568   __ret; \
   40569 })
   40570 #else
   40571 #define vqrdmlahq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40572   int16x8_t __s0 = __p0; \
   40573   int16x8_t __s1 = __p1; \
   40574   int16x4_t __s2 = __p2; \
   40575   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   40576   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   40577   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   40578   int16x8_t __ret; \
   40579   __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
   40580   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   40581   __ret; \
   40582 })
   40583 #endif
   40584 
   40585 #ifdef __LITTLE_ENDIAN__
   40586 #define vqrdmlah_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40587   int32x2_t __s0 = __p0; \
   40588   int32x2_t __s1 = __p1; \
   40589   int32x2_t __s2 = __p2; \
   40590   int32x2_t __ret; \
   40591   __ret = vqadd_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
   40592   __ret; \
   40593 })
   40594 #else
   40595 #define vqrdmlah_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40596   int32x2_t __s0 = __p0; \
   40597   int32x2_t __s1 = __p1; \
   40598   int32x2_t __s2 = __p2; \
   40599   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   40600   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   40601   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   40602   int32x2_t __ret; \
   40603   __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
   40604   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   40605   __ret; \
   40606 })
   40607 #endif
   40608 
   40609 #ifdef __LITTLE_ENDIAN__
   40610 #define vqrdmlah_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40611   int16x4_t __s0 = __p0; \
   40612   int16x4_t __s1 = __p1; \
   40613   int16x4_t __s2 = __p2; \
   40614   int16x4_t __ret; \
   40615   __ret = vqadd_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
   40616   __ret; \
   40617 })
   40618 #else
   40619 #define vqrdmlah_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40620   int16x4_t __s0 = __p0; \
   40621   int16x4_t __s1 = __p1; \
   40622   int16x4_t __s2 = __p2; \
   40623   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   40624   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   40625   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   40626   int16x4_t __ret; \
   40627   __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
   40628   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   40629   __ret; \
   40630 })
   40631 #endif
   40632 
   40633 #ifdef __LITTLE_ENDIAN__
   40634 __ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   40635   int32x4_t __ret;
   40636   __ret = vqsubq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
   40637   return __ret;
   40638 }
   40639 #else
   40640 __ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   40641   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40642   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40643   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40644   int32x4_t __ret;
   40645   __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
   40646   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40647   return __ret;
   40648 }
   40649 #endif
   40650 
   40651 #ifdef __LITTLE_ENDIAN__
   40652 __ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   40653   int16x8_t __ret;
   40654   __ret = vqsubq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
   40655   return __ret;
   40656 }
   40657 #else
   40658 __ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   40659   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   40660   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   40661   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   40662   int16x8_t __ret;
   40663   __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
   40664   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   40665   return __ret;
   40666 }
   40667 #endif
   40668 
   40669 #ifdef __LITTLE_ENDIAN__
   40670 __ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   40671   int32x2_t __ret;
   40672   __ret = vqsub_s32(__p0, vqrdmulh_s32(__p1, __p2));
   40673   return __ret;
   40674 }
   40675 #else
   40676 __ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   40677   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   40678   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   40679   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   40680   int32x2_t __ret;
   40681   __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
   40682   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   40683   return __ret;
   40684 }
   40685 #endif
   40686 
   40687 #ifdef __LITTLE_ENDIAN__
   40688 __ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   40689   int16x4_t __ret;
   40690   __ret = vqsub_s16(__p0, vqrdmulh_s16(__p1, __p2));
   40691   return __ret;
   40692 }
   40693 #else
   40694 __ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   40695   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   40696   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   40697   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   40698   int16x4_t __ret;
   40699   __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
   40700   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   40701   return __ret;
   40702 }
   40703 #endif
   40704 
   40705 #ifdef __LITTLE_ENDIAN__
   40706 #define vqrdmlshq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40707   int32x4_t __s0 = __p0; \
   40708   int32x4_t __s1 = __p1; \
   40709   int32x2_t __s2 = __p2; \
   40710   int32x4_t __ret; \
   40711   __ret = vqsubq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
   40712   __ret; \
   40713 })
   40714 #else
   40715 #define vqrdmlshq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40716   int32x4_t __s0 = __p0; \
   40717   int32x4_t __s1 = __p1; \
   40718   int32x2_t __s2 = __p2; \
   40719   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   40720   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   40721   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   40722   int32x4_t __ret; \
   40723   __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
   40724   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   40725   __ret; \
   40726 })
   40727 #endif
   40728 
   40729 #ifdef __LITTLE_ENDIAN__
   40730 #define vqrdmlshq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40731   int16x8_t __s0 = __p0; \
   40732   int16x8_t __s1 = __p1; \
   40733   int16x4_t __s2 = __p2; \
   40734   int16x8_t __ret; \
   40735   __ret = vqsubq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
   40736   __ret; \
   40737 })
   40738 #else
   40739 #define vqrdmlshq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40740   int16x8_t __s0 = __p0; \
   40741   int16x8_t __s1 = __p1; \
   40742   int16x4_t __s2 = __p2; \
   40743   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   40744   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   40745   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   40746   int16x8_t __ret; \
   40747   __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
   40748   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   40749   __ret; \
   40750 })
   40751 #endif
   40752 
   40753 #ifdef __LITTLE_ENDIAN__
   40754 #define vqrdmlsh_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40755   int32x2_t __s0 = __p0; \
   40756   int32x2_t __s1 = __p1; \
   40757   int32x2_t __s2 = __p2; \
   40758   int32x2_t __ret; \
   40759   __ret = vqsub_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
   40760   __ret; \
   40761 })
   40762 #else
   40763 #define vqrdmlsh_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40764   int32x2_t __s0 = __p0; \
   40765   int32x2_t __s1 = __p1; \
   40766   int32x2_t __s2 = __p2; \
   40767   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   40768   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   40769   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   40770   int32x2_t __ret; \
   40771   __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
   40772   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   40773   __ret; \
   40774 })
   40775 #endif
   40776 
   40777 #ifdef __LITTLE_ENDIAN__
   40778 #define vqrdmlsh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40779   int16x4_t __s0 = __p0; \
   40780   int16x4_t __s1 = __p1; \
   40781   int16x4_t __s2 = __p2; \
   40782   int16x4_t __ret; \
   40783   __ret = vqsub_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
   40784   __ret; \
   40785 })
   40786 #else
   40787 #define vqrdmlsh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40788   int16x4_t __s0 = __p0; \
   40789   int16x4_t __s1 = __p1; \
   40790   int16x4_t __s2 = __p2; \
   40791   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   40792   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   40793   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   40794   int16x4_t __ret; \
   40795   __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
   40796   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   40797   __ret; \
   40798 })
   40799 #endif
   40800 
   40801 #endif
   40802 #if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
   40803 #ifdef __LITTLE_ENDIAN__
   40804 #define vqrdmlahq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40805   int32x4_t __s0 = __p0; \
   40806   int32x4_t __s1 = __p1; \
   40807   int32x4_t __s2 = __p2; \
   40808   int32x4_t __ret; \
   40809   __ret = vqaddq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
   40810   __ret; \
   40811 })
   40812 #else
   40813 #define vqrdmlahq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40814   int32x4_t __s0 = __p0; \
   40815   int32x4_t __s1 = __p1; \
   40816   int32x4_t __s2 = __p2; \
   40817   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   40818   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   40819   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   40820   int32x4_t __ret; \
   40821   __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
   40822   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   40823   __ret; \
   40824 })
   40825 #endif
   40826 
   40827 #ifdef __LITTLE_ENDIAN__
   40828 #define vqrdmlahq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40829   int16x8_t __s0 = __p0; \
   40830   int16x8_t __s1 = __p1; \
   40831   int16x8_t __s2 = __p2; \
   40832   int16x8_t __ret; \
   40833   __ret = vqaddq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
   40834   __ret; \
   40835 })
   40836 #else
   40837 #define vqrdmlahq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40838   int16x8_t __s0 = __p0; \
   40839   int16x8_t __s1 = __p1; \
   40840   int16x8_t __s2 = __p2; \
   40841   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   40842   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   40843   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   40844   int16x8_t __ret; \
   40845   __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
   40846   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   40847   __ret; \
   40848 })
   40849 #endif
   40850 
   40851 #ifdef __LITTLE_ENDIAN__
   40852 #define vqrdmlah_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40853   int32x2_t __s0 = __p0; \
   40854   int32x2_t __s1 = __p1; \
   40855   int32x4_t __s2 = __p2; \
   40856   int32x2_t __ret; \
   40857   __ret = vqadd_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
   40858   __ret; \
   40859 })
   40860 #else
   40861 #define vqrdmlah_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40862   int32x2_t __s0 = __p0; \
   40863   int32x2_t __s1 = __p1; \
   40864   int32x4_t __s2 = __p2; \
   40865   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   40866   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   40867   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   40868   int32x2_t __ret; \
   40869   __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
   40870   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   40871   __ret; \
   40872 })
   40873 #endif
   40874 
   40875 #ifdef __LITTLE_ENDIAN__
   40876 #define vqrdmlah_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40877   int16x4_t __s0 = __p0; \
   40878   int16x4_t __s1 = __p1; \
   40879   int16x8_t __s2 = __p2; \
   40880   int16x4_t __ret; \
   40881   __ret = vqadd_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
   40882   __ret; \
   40883 })
   40884 #else
   40885 #define vqrdmlah_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40886   int16x4_t __s0 = __p0; \
   40887   int16x4_t __s1 = __p1; \
   40888   int16x8_t __s2 = __p2; \
   40889   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   40890   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   40891   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   40892   int16x4_t __ret; \
   40893   __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
   40894   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   40895   __ret; \
   40896 })
   40897 #endif
   40898 
   40899 #ifdef __LITTLE_ENDIAN__
   40900 #define vqrdmlshq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40901   int32x4_t __s0 = __p0; \
   40902   int32x4_t __s1 = __p1; \
   40903   int32x4_t __s2 = __p2; \
   40904   int32x4_t __ret; \
   40905   __ret = vqsubq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
   40906   __ret; \
   40907 })
   40908 #else
   40909 #define vqrdmlshq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40910   int32x4_t __s0 = __p0; \
   40911   int32x4_t __s1 = __p1; \
   40912   int32x4_t __s2 = __p2; \
   40913   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   40914   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   40915   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   40916   int32x4_t __ret; \
   40917   __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
   40918   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   40919   __ret; \
   40920 })
   40921 #endif
   40922 
   40923 #ifdef __LITTLE_ENDIAN__
   40924 #define vqrdmlshq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40925   int16x8_t __s0 = __p0; \
   40926   int16x8_t __s1 = __p1; \
   40927   int16x8_t __s2 = __p2; \
   40928   int16x8_t __ret; \
   40929   __ret = vqsubq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
   40930   __ret; \
   40931 })
   40932 #else
   40933 #define vqrdmlshq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40934   int16x8_t __s0 = __p0; \
   40935   int16x8_t __s1 = __p1; \
   40936   int16x8_t __s2 = __p2; \
   40937   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   40938   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   40939   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   40940   int16x8_t __ret; \
   40941   __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
   40942   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   40943   __ret; \
   40944 })
   40945 #endif
   40946 
   40947 #ifdef __LITTLE_ENDIAN__
   40948 #define vqrdmlsh_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40949   int32x2_t __s0 = __p0; \
   40950   int32x2_t __s1 = __p1; \
   40951   int32x4_t __s2 = __p2; \
   40952   int32x2_t __ret; \
   40953   __ret = vqsub_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
   40954   __ret; \
   40955 })
   40956 #else
   40957 #define vqrdmlsh_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   40958   int32x2_t __s0 = __p0; \
   40959   int32x2_t __s1 = __p1; \
   40960   int32x4_t __s2 = __p2; \
   40961   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   40962   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   40963   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   40964   int32x2_t __ret; \
   40965   __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
   40966   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   40967   __ret; \
   40968 })
   40969 #endif
   40970 
   40971 #ifdef __LITTLE_ENDIAN__
   40972 #define vqrdmlsh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40973   int16x4_t __s0 = __p0; \
   40974   int16x4_t __s1 = __p1; \
   40975   int16x8_t __s2 = __p2; \
   40976   int16x4_t __ret; \
   40977   __ret = vqsub_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
   40978   __ret; \
   40979 })
   40980 #else
   40981 #define vqrdmlsh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   40982   int16x4_t __s0 = __p0; \
   40983   int16x4_t __s1 = __p1; \
   40984   int16x8_t __s2 = __p2; \
   40985   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   40986   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   40987   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   40988   int16x4_t __ret; \
   40989   __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
   40990   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   40991   __ret; \
   40992 })
   40993 #endif
   40994 
   40995 #endif
   40996 #if defined(__aarch64__)
   40997 #ifdef __LITTLE_ENDIAN__
   40998 __ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
   40999   float64x2_t __ret;
   41000   __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   41001   return __ret;
   41002 }
   41003 #else
   41004 __ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
   41005   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41006   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41007   float64x2_t __ret;
   41008   __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   41009   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41010   return __ret;
   41011 }
   41012 #endif
   41013 
   41014 #ifdef __LITTLE_ENDIAN__
   41015 __ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
   41016   float64x1_t __ret;
   41017   __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   41018   return __ret;
   41019 }
   41020 #else
   41021 __ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
   41022   float64x1_t __ret;
   41023   __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   41024   return __ret;
   41025 }
   41026 #endif
   41027 
   41028 #ifdef __LITTLE_ENDIAN__
   41029 __ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
   41030   float64_t __ret;
   41031   __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
   41032   return __ret;
   41033 }
   41034 #else
   41035 __ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
   41036   float64_t __ret;
   41037   __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
   41038   return __ret;
   41039 }
   41040 #endif
   41041 
   41042 #ifdef __LITTLE_ENDIAN__
   41043 __ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
   41044   float32_t __ret;
   41045   __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
   41046   return __ret;
   41047 }
   41048 #else
   41049 __ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
   41050   float32_t __ret;
   41051   __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
   41052   return __ret;
   41053 }
   41054 #endif
   41055 
   41056 #ifdef __LITTLE_ENDIAN__
   41057 __ai float64x2_t vabsq_f64(float64x2_t __p0) {
   41058   float64x2_t __ret;
   41059   __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 42);
   41060   return __ret;
   41061 }
   41062 #else
   41063 __ai float64x2_t vabsq_f64(float64x2_t __p0) {
   41064   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41065   float64x2_t __ret;
   41066   __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 42);
   41067   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41068   return __ret;
   41069 }
   41070 #endif
   41071 
   41072 #ifdef __LITTLE_ENDIAN__
   41073 __ai int64x2_t vabsq_s64(int64x2_t __p0) {
   41074   int64x2_t __ret;
   41075   __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 35);
   41076   return __ret;
   41077 }
   41078 #else
   41079 __ai int64x2_t vabsq_s64(int64x2_t __p0) {
   41080   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41081   int64x2_t __ret;
   41082   __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 35);
   41083   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41084   return __ret;
   41085 }
   41086 #endif
   41087 
   41088 #ifdef __LITTLE_ENDIAN__
   41089 __ai float64x1_t vabs_f64(float64x1_t __p0) {
   41090   float64x1_t __ret;
   41091   __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
   41092   return __ret;
   41093 }
   41094 #else
   41095 __ai float64x1_t vabs_f64(float64x1_t __p0) {
   41096   float64x1_t __ret;
   41097   __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
   41098   return __ret;
   41099 }
   41100 #endif
   41101 
   41102 #ifdef __LITTLE_ENDIAN__
   41103 __ai int64x1_t vabs_s64(int64x1_t __p0) {
   41104   int64x1_t __ret;
   41105   __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
   41106   return __ret;
   41107 }
   41108 #else
   41109 __ai int64x1_t vabs_s64(int64x1_t __p0) {
   41110   int64x1_t __ret;
   41111   __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
   41112   return __ret;
   41113 }
   41114 #endif
   41115 
   41116 #ifdef __LITTLE_ENDIAN__
   41117 __ai int64_t vabsd_s64(int64_t __p0) {
   41118   int64_t __ret;
   41119   __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
   41120   return __ret;
   41121 }
   41122 #else
   41123 __ai int64_t vabsd_s64(int64_t __p0) {
   41124   int64_t __ret;
   41125   __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
   41126   return __ret;
   41127 }
   41128 #endif
   41129 
   41130 #ifdef __LITTLE_ENDIAN__
   41131 __ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
   41132   float64x2_t __ret;
   41133   __ret = __p0 + __p1;
   41134   return __ret;
   41135 }
   41136 #else
   41137 __ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
   41138   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41139   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41140   float64x2_t __ret;
   41141   __ret = __rev0 + __rev1;
   41142   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41143   return __ret;
   41144 }
   41145 #endif
   41146 
   41147 #ifdef __LITTLE_ENDIAN__
   41148 __ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
   41149   float64x1_t __ret;
   41150   __ret = __p0 + __p1;
   41151   return __ret;
   41152 }
   41153 #else
   41154 __ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
   41155   float64x1_t __ret;
   41156   __ret = __p0 + __p1;
   41157   return __ret;
   41158 }
   41159 #endif
   41160 
   41161 #ifdef __LITTLE_ENDIAN__
   41162 __ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
   41163   uint64_t __ret;
   41164   __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
   41165   return __ret;
   41166 }
   41167 #else
   41168 __ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
   41169   uint64_t __ret;
   41170   __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
   41171   return __ret;
   41172 }
   41173 #endif
   41174 
   41175 #ifdef __LITTLE_ENDIAN__
   41176 __ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
   41177   int64_t __ret;
   41178   __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
   41179   return __ret;
   41180 }
   41181 #else
   41182 __ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
   41183   int64_t __ret;
   41184   __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
   41185   return __ret;
   41186 }
   41187 #endif
   41188 
   41189 #ifdef __LITTLE_ENDIAN__
   41190 __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   41191   uint16x8_t __ret;
   41192   __ret = vcombine_u16(__p0, vaddhn_u32(__p1, __p2));
   41193   return __ret;
   41194 }
   41195 #else
   41196 __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   41197   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41198   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   41199   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   41200   uint16x8_t __ret;
   41201   __ret = __noswap_vcombine_u16(__rev0, __noswap_vaddhn_u32(__rev1, __rev2));
   41202   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   41203   return __ret;
   41204 }
   41205 #endif
   41206 
   41207 #ifdef __LITTLE_ENDIAN__
   41208 __ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   41209   uint32x4_t __ret;
   41210   __ret = vcombine_u32(__p0, vaddhn_u64(__p1, __p2));
   41211   return __ret;
   41212 }
   41213 #else
   41214 __ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   41215   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41216   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41217   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   41218   uint32x4_t __ret;
   41219   __ret = __noswap_vcombine_u32(__rev0, __noswap_vaddhn_u64(__rev1, __rev2));
   41220   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   41221   return __ret;
   41222 }
   41223 #endif
   41224 
   41225 #ifdef __LITTLE_ENDIAN__
   41226 __ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   41227   uint8x16_t __ret;
   41228   __ret = vcombine_u8(__p0, vaddhn_u16(__p1, __p2));
   41229   return __ret;
   41230 }
   41231 #else
   41232 __ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   41233   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41234   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   41235   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   41236   uint8x16_t __ret;
   41237   __ret = __noswap_vcombine_u8(__rev0, __noswap_vaddhn_u16(__rev1, __rev2));
   41238   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   41239   return __ret;
   41240 }
   41241 #endif
   41242 
   41243 #ifdef __LITTLE_ENDIAN__
   41244 __ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   41245   int16x8_t __ret;
   41246   __ret = vcombine_s16(__p0, vaddhn_s32(__p1, __p2));
   41247   return __ret;
   41248 }
   41249 #else
   41250 __ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   41251   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41252   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   41253   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   41254   int16x8_t __ret;
   41255   __ret = __noswap_vcombine_s16(__rev0, __noswap_vaddhn_s32(__rev1, __rev2));
   41256   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   41257   return __ret;
   41258 }
   41259 #endif
   41260 
   41261 #ifdef __LITTLE_ENDIAN__
   41262 __ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   41263   int32x4_t __ret;
   41264   __ret = vcombine_s32(__p0, vaddhn_s64(__p1, __p2));
   41265   return __ret;
   41266 }
   41267 #else
   41268 __ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   41269   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41270   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41271   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   41272   int32x4_t __ret;
   41273   __ret = __noswap_vcombine_s32(__rev0, __noswap_vaddhn_s64(__rev1, __rev2));
   41274   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   41275   return __ret;
   41276 }
   41277 #endif
   41278 
   41279 #ifdef __LITTLE_ENDIAN__
   41280 __ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   41281   int8x16_t __ret;
   41282   __ret = vcombine_s8(__p0, vaddhn_s16(__p1, __p2));
   41283   return __ret;
   41284 }
   41285 #else
   41286 __ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   41287   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41288   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   41289   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   41290   int8x16_t __ret;
   41291   __ret = __noswap_vcombine_s8(__rev0, __noswap_vaddhn_s16(__rev1, __rev2));
   41292   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   41293   return __ret;
   41294 }
   41295 #endif
   41296 
   41297 #ifdef __LITTLE_ENDIAN__
   41298 __ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
   41299   uint16_t __ret;
   41300   __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__p0);
   41301   return __ret;
   41302 }
   41303 #else
   41304 __ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
   41305   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   41306   uint16_t __ret;
   41307   __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__rev0);
   41308   return __ret;
   41309 }
   41310 #endif
   41311 
   41312 #ifdef __LITTLE_ENDIAN__
   41313 __ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
   41314   uint64_t __ret;
   41315   __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__p0);
   41316   return __ret;
   41317 }
   41318 #else
   41319 __ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
   41320   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41321   uint64_t __ret;
   41322   __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__rev0);
   41323   return __ret;
   41324 }
   41325 #endif
   41326 
   41327 #ifdef __LITTLE_ENDIAN__
   41328 __ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
   41329   uint32_t __ret;
   41330   __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__p0);
   41331   return __ret;
   41332 }
   41333 #else
   41334 __ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
   41335   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41336   uint32_t __ret;
   41337   __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__rev0);
   41338   return __ret;
   41339 }
   41340 #endif
   41341 
   41342 #ifdef __LITTLE_ENDIAN__
   41343 __ai int16_t vaddlvq_s8(int8x16_t __p0) {
   41344   int16_t __ret;
   41345   __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__p0);
   41346   return __ret;
   41347 }
   41348 #else
   41349 __ai int16_t vaddlvq_s8(int8x16_t __p0) {
   41350   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   41351   int16_t __ret;
   41352   __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__rev0);
   41353   return __ret;
   41354 }
   41355 #endif
   41356 
   41357 #ifdef __LITTLE_ENDIAN__
   41358 __ai int64_t vaddlvq_s32(int32x4_t __p0) {
   41359   int64_t __ret;
   41360   __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__p0);
   41361   return __ret;
   41362 }
   41363 #else
   41364 __ai int64_t vaddlvq_s32(int32x4_t __p0) {
   41365   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41366   int64_t __ret;
   41367   __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__rev0);
   41368   return __ret;
   41369 }
   41370 #endif
   41371 
   41372 #ifdef __LITTLE_ENDIAN__
   41373 __ai int32_t vaddlvq_s16(int16x8_t __p0) {
   41374   int32_t __ret;
   41375   __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__p0);
   41376   return __ret;
   41377 }
   41378 #else
   41379 __ai int32_t vaddlvq_s16(int16x8_t __p0) {
   41380   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41381   int32_t __ret;
   41382   __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__rev0);
   41383   return __ret;
   41384 }
   41385 #endif
   41386 
   41387 #ifdef __LITTLE_ENDIAN__
   41388 __ai uint16_t vaddlv_u8(uint8x8_t __p0) {
   41389   uint16_t __ret;
   41390   __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__p0);
   41391   return __ret;
   41392 }
   41393 #else
   41394 __ai uint16_t vaddlv_u8(uint8x8_t __p0) {
   41395   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41396   uint16_t __ret;
   41397   __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__rev0);
   41398   return __ret;
   41399 }
   41400 #endif
   41401 
   41402 #ifdef __LITTLE_ENDIAN__
   41403 __ai uint64_t vaddlv_u32(uint32x2_t __p0) {
   41404   uint64_t __ret;
   41405   __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__p0);
   41406   return __ret;
   41407 }
   41408 #else
   41409 __ai uint64_t vaddlv_u32(uint32x2_t __p0) {
   41410   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41411   uint64_t __ret;
   41412   __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__rev0);
   41413   return __ret;
   41414 }
   41415 #endif
   41416 
   41417 #ifdef __LITTLE_ENDIAN__
   41418 __ai uint32_t vaddlv_u16(uint16x4_t __p0) {
   41419   uint32_t __ret;
   41420   __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__p0);
   41421   return __ret;
   41422 }
   41423 #else
   41424 __ai uint32_t vaddlv_u16(uint16x4_t __p0) {
   41425   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41426   uint32_t __ret;
   41427   __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__rev0);
   41428   return __ret;
   41429 }
   41430 #endif
   41431 
   41432 #ifdef __LITTLE_ENDIAN__
   41433 __ai int16_t vaddlv_s8(int8x8_t __p0) {
   41434   int16_t __ret;
   41435   __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__p0);
   41436   return __ret;
   41437 }
   41438 #else
   41439 __ai int16_t vaddlv_s8(int8x8_t __p0) {
   41440   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41441   int16_t __ret;
   41442   __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__rev0);
   41443   return __ret;
   41444 }
   41445 #endif
   41446 
   41447 #ifdef __LITTLE_ENDIAN__
   41448 __ai int64_t vaddlv_s32(int32x2_t __p0) {
   41449   int64_t __ret;
   41450   __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__p0);
   41451   return __ret;
   41452 }
   41453 #else
   41454 __ai int64_t vaddlv_s32(int32x2_t __p0) {
   41455   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41456   int64_t __ret;
   41457   __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__rev0);
   41458   return __ret;
   41459 }
   41460 #endif
   41461 
   41462 #ifdef __LITTLE_ENDIAN__
   41463 __ai int32_t vaddlv_s16(int16x4_t __p0) {
   41464   int32_t __ret;
   41465   __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__p0);
   41466   return __ret;
   41467 }
   41468 #else
   41469 __ai int32_t vaddlv_s16(int16x4_t __p0) {
   41470   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41471   int32_t __ret;
   41472   __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__rev0);
   41473   return __ret;
   41474 }
   41475 #endif
   41476 
   41477 #ifdef __LITTLE_ENDIAN__
   41478 __ai uint8_t vaddvq_u8(uint8x16_t __p0) {
   41479   uint8_t __ret;
   41480   __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__p0);
   41481   return __ret;
   41482 }
   41483 #else
   41484 __ai uint8_t vaddvq_u8(uint8x16_t __p0) {
   41485   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   41486   uint8_t __ret;
   41487   __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__rev0);
   41488   return __ret;
   41489 }
   41490 #endif
   41491 
   41492 #ifdef __LITTLE_ENDIAN__
   41493 __ai uint32_t vaddvq_u32(uint32x4_t __p0) {
   41494   uint32_t __ret;
   41495   __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__p0);
   41496   return __ret;
   41497 }
   41498 #else
   41499 __ai uint32_t vaddvq_u32(uint32x4_t __p0) {
   41500   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41501   uint32_t __ret;
   41502   __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__rev0);
   41503   return __ret;
   41504 }
   41505 #endif
   41506 
   41507 #ifdef __LITTLE_ENDIAN__
   41508 __ai uint64_t vaddvq_u64(uint64x2_t __p0) {
   41509   uint64_t __ret;
   41510   __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__p0);
   41511   return __ret;
   41512 }
   41513 #else
   41514 __ai uint64_t vaddvq_u64(uint64x2_t __p0) {
   41515   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41516   uint64_t __ret;
   41517   __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__rev0);
   41518   return __ret;
   41519 }
   41520 #endif
   41521 
   41522 #ifdef __LITTLE_ENDIAN__
   41523 __ai uint16_t vaddvq_u16(uint16x8_t __p0) {
   41524   uint16_t __ret;
   41525   __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__p0);
   41526   return __ret;
   41527 }
   41528 #else
   41529 __ai uint16_t vaddvq_u16(uint16x8_t __p0) {
   41530   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41531   uint16_t __ret;
   41532   __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__rev0);
   41533   return __ret;
   41534 }
   41535 #endif
   41536 
   41537 #ifdef __LITTLE_ENDIAN__
   41538 __ai int8_t vaddvq_s8(int8x16_t __p0) {
   41539   int8_t __ret;
   41540   __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__p0);
   41541   return __ret;
   41542 }
   41543 #else
   41544 __ai int8_t vaddvq_s8(int8x16_t __p0) {
   41545   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   41546   int8_t __ret;
   41547   __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__rev0);
   41548   return __ret;
   41549 }
   41550 #endif
   41551 
   41552 #ifdef __LITTLE_ENDIAN__
   41553 __ai float64_t vaddvq_f64(float64x2_t __p0) {
   41554   float64_t __ret;
   41555   __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__p0);
   41556   return __ret;
   41557 }
   41558 #else
   41559 __ai float64_t vaddvq_f64(float64x2_t __p0) {
   41560   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41561   float64_t __ret;
   41562   __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__rev0);
   41563   return __ret;
   41564 }
   41565 #endif
   41566 
   41567 #ifdef __LITTLE_ENDIAN__
   41568 __ai float32_t vaddvq_f32(float32x4_t __p0) {
   41569   float32_t __ret;
   41570   __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__p0);
   41571   return __ret;
   41572 }
   41573 #else
   41574 __ai float32_t vaddvq_f32(float32x4_t __p0) {
   41575   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41576   float32_t __ret;
   41577   __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__rev0);
   41578   return __ret;
   41579 }
   41580 #endif
   41581 
   41582 #ifdef __LITTLE_ENDIAN__
   41583 __ai int32_t vaddvq_s32(int32x4_t __p0) {
   41584   int32_t __ret;
   41585   __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__p0);
   41586   return __ret;
   41587 }
   41588 #else
   41589 __ai int32_t vaddvq_s32(int32x4_t __p0) {
   41590   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41591   int32_t __ret;
   41592   __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__rev0);
   41593   return __ret;
   41594 }
   41595 #endif
   41596 
   41597 #ifdef __LITTLE_ENDIAN__
   41598 __ai int64_t vaddvq_s64(int64x2_t __p0) {
   41599   int64_t __ret;
   41600   __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__p0);
   41601   return __ret;
   41602 }
   41603 #else
   41604 __ai int64_t vaddvq_s64(int64x2_t __p0) {
   41605   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41606   int64_t __ret;
   41607   __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__rev0);
   41608   return __ret;
   41609 }
   41610 #endif
   41611 
   41612 #ifdef __LITTLE_ENDIAN__
   41613 __ai int16_t vaddvq_s16(int16x8_t __p0) {
   41614   int16_t __ret;
   41615   __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__p0);
   41616   return __ret;
   41617 }
   41618 #else
   41619 __ai int16_t vaddvq_s16(int16x8_t __p0) {
   41620   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41621   int16_t __ret;
   41622   __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__rev0);
   41623   return __ret;
   41624 }
   41625 #endif
   41626 
   41627 #ifdef __LITTLE_ENDIAN__
   41628 __ai uint8_t vaddv_u8(uint8x8_t __p0) {
   41629   uint8_t __ret;
   41630   __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__p0);
   41631   return __ret;
   41632 }
   41633 #else
   41634 __ai uint8_t vaddv_u8(uint8x8_t __p0) {
   41635   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41636   uint8_t __ret;
   41637   __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__rev0);
   41638   return __ret;
   41639 }
   41640 #endif
   41641 
   41642 #ifdef __LITTLE_ENDIAN__
   41643 __ai uint32_t vaddv_u32(uint32x2_t __p0) {
   41644   uint32_t __ret;
   41645   __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__p0);
   41646   return __ret;
   41647 }
   41648 #else
   41649 __ai uint32_t vaddv_u32(uint32x2_t __p0) {
   41650   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41651   uint32_t __ret;
   41652   __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__rev0);
   41653   return __ret;
   41654 }
   41655 #endif
   41656 
   41657 #ifdef __LITTLE_ENDIAN__
   41658 __ai uint16_t vaddv_u16(uint16x4_t __p0) {
   41659   uint16_t __ret;
   41660   __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__p0);
   41661   return __ret;
   41662 }
   41663 #else
   41664 __ai uint16_t vaddv_u16(uint16x4_t __p0) {
   41665   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41666   uint16_t __ret;
   41667   __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__rev0);
   41668   return __ret;
   41669 }
   41670 #endif
   41671 
   41672 #ifdef __LITTLE_ENDIAN__
   41673 __ai int8_t vaddv_s8(int8x8_t __p0) {
   41674   int8_t __ret;
   41675   __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__p0);
   41676   return __ret;
   41677 }
   41678 #else
   41679 __ai int8_t vaddv_s8(int8x8_t __p0) {
   41680   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   41681   int8_t __ret;
   41682   __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__rev0);
   41683   return __ret;
   41684 }
   41685 #endif
   41686 
   41687 #ifdef __LITTLE_ENDIAN__
   41688 __ai float32_t vaddv_f32(float32x2_t __p0) {
   41689   float32_t __ret;
   41690   __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__p0);
   41691   return __ret;
   41692 }
   41693 #else
   41694 __ai float32_t vaddv_f32(float32x2_t __p0) {
   41695   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41696   float32_t __ret;
   41697   __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__rev0);
   41698   return __ret;
   41699 }
   41700 #endif
   41701 
   41702 #ifdef __LITTLE_ENDIAN__
   41703 __ai int32_t vaddv_s32(int32x2_t __p0) {
   41704   int32_t __ret;
   41705   __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__p0);
   41706   return __ret;
   41707 }
   41708 #else
   41709 __ai int32_t vaddv_s32(int32x2_t __p0) {
   41710   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41711   int32_t __ret;
   41712   __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__rev0);
   41713   return __ret;
   41714 }
   41715 #endif
   41716 
   41717 #ifdef __LITTLE_ENDIAN__
   41718 __ai int16_t vaddv_s16(int16x4_t __p0) {
   41719   int16_t __ret;
   41720   __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__p0);
   41721   return __ret;
   41722 }
   41723 #else
   41724 __ai int16_t vaddv_s16(int16x4_t __p0) {
   41725   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   41726   int16_t __ret;
   41727   __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__rev0);
   41728   return __ret;
   41729 }
   41730 #endif
   41731 
   41732 #ifdef __LITTLE_ENDIAN__
   41733 __ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
   41734   poly64x1_t __ret;
   41735   __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
   41736   return __ret;
   41737 }
   41738 #else
   41739 __ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
   41740   poly64x1_t __ret;
   41741   __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
   41742   return __ret;
   41743 }
   41744 #endif
   41745 
   41746 #ifdef __LITTLE_ENDIAN__
   41747 __ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
   41748   poly64x2_t __ret;
   41749   __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 38);
   41750   return __ret;
   41751 }
   41752 #else
   41753 __ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
   41754   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41755   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41756   poly64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   41757   poly64x2_t __ret;
   41758   __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 38);
   41759   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41760   return __ret;
   41761 }
   41762 #endif
   41763 
   41764 #ifdef __LITTLE_ENDIAN__
   41765 __ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   41766   float64x2_t __ret;
   41767   __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
   41768   return __ret;
   41769 }
   41770 #else
   41771 __ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   41772   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41773   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41774   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   41775   float64x2_t __ret;
   41776   __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
   41777   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41778   return __ret;
   41779 }
   41780 #endif
   41781 
   41782 #ifdef __LITTLE_ENDIAN__
   41783 __ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   41784   float64x1_t __ret;
   41785   __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
   41786   return __ret;
   41787 }
   41788 #else
   41789 __ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   41790   float64x1_t __ret;
   41791   __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
   41792   return __ret;
   41793 }
   41794 #endif
   41795 
   41796 #ifdef __LITTLE_ENDIAN__
   41797 __ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
   41798   uint64x2_t __ret;
   41799   __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   41800   return __ret;
   41801 }
   41802 #else
   41803 __ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
   41804   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41805   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41806   uint64x2_t __ret;
   41807   __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   41808   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41809   return __ret;
   41810 }
   41811 #endif
   41812 
   41813 #ifdef __LITTLE_ENDIAN__
   41814 __ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
   41815   uint64x1_t __ret;
   41816   __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   41817   return __ret;
   41818 }
   41819 #else
   41820 __ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
   41821   uint64x1_t __ret;
   41822   __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   41823   return __ret;
   41824 }
   41825 #endif
   41826 
   41827 #ifdef __LITTLE_ENDIAN__
   41828 __ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
   41829   uint64_t __ret;
   41830   __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
   41831   return __ret;
   41832 }
   41833 #else
   41834 __ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
   41835   uint64_t __ret;
   41836   __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
   41837   return __ret;
   41838 }
   41839 #endif
   41840 
   41841 #ifdef __LITTLE_ENDIAN__
   41842 __ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
   41843   uint32_t __ret;
   41844   __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
   41845   return __ret;
   41846 }
   41847 #else
   41848 __ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
   41849   uint32_t __ret;
   41850   __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
   41851   return __ret;
   41852 }
   41853 #endif
   41854 
   41855 #ifdef __LITTLE_ENDIAN__
   41856 __ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
   41857   uint64x2_t __ret;
   41858   __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   41859   return __ret;
   41860 }
   41861 #else
   41862 __ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
   41863   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41864   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41865   uint64x2_t __ret;
   41866   __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   41867   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41868   return __ret;
   41869 }
   41870 #endif
   41871 
   41872 #ifdef __LITTLE_ENDIAN__
   41873 __ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
   41874   uint64x1_t __ret;
   41875   __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   41876   return __ret;
   41877 }
   41878 #else
   41879 __ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
   41880   uint64x1_t __ret;
   41881   __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   41882   return __ret;
   41883 }
   41884 #endif
   41885 
   41886 #ifdef __LITTLE_ENDIAN__
   41887 __ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
   41888   uint64_t __ret;
   41889   __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
   41890   return __ret;
   41891 }
   41892 #else
   41893 __ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
   41894   uint64_t __ret;
   41895   __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
   41896   return __ret;
   41897 }
   41898 #endif
   41899 
   41900 #ifdef __LITTLE_ENDIAN__
   41901 __ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
   41902   uint32_t __ret;
   41903   __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
   41904   return __ret;
   41905 }
   41906 #else
   41907 __ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
   41908   uint32_t __ret;
   41909   __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
   41910   return __ret;
   41911 }
   41912 #endif
   41913 
   41914 #ifdef __LITTLE_ENDIAN__
   41915 __ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
   41916   uint64x2_t __ret;
   41917   __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   41918   return __ret;
   41919 }
   41920 #else
   41921 __ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
   41922   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41923   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41924   uint64x2_t __ret;
   41925   __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   41926   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41927   return __ret;
   41928 }
   41929 #endif
   41930 
   41931 #ifdef __LITTLE_ENDIAN__
   41932 __ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
   41933   uint64x1_t __ret;
   41934   __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   41935   return __ret;
   41936 }
   41937 #else
   41938 __ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
   41939   uint64x1_t __ret;
   41940   __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   41941   return __ret;
   41942 }
   41943 #endif
   41944 
   41945 #ifdef __LITTLE_ENDIAN__
   41946 __ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
   41947   uint64_t __ret;
   41948   __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
   41949   return __ret;
   41950 }
   41951 #else
   41952 __ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
   41953   uint64_t __ret;
   41954   __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
   41955   return __ret;
   41956 }
   41957 #endif
   41958 
   41959 #ifdef __LITTLE_ENDIAN__
   41960 __ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
   41961   uint32_t __ret;
   41962   __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
   41963   return __ret;
   41964 }
   41965 #else
   41966 __ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
   41967   uint32_t __ret;
   41968   __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
   41969   return __ret;
   41970 }
   41971 #endif
   41972 
   41973 #ifdef __LITTLE_ENDIAN__
   41974 __ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
   41975   uint64x2_t __ret;
   41976   __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   41977   return __ret;
   41978 }
   41979 #else
   41980 __ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
   41981   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   41982   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   41983   uint64x2_t __ret;
   41984   __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   41985   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   41986   return __ret;
   41987 }
   41988 #endif
   41989 
   41990 #ifdef __LITTLE_ENDIAN__
   41991 __ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
   41992   uint64x1_t __ret;
   41993   __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   41994   return __ret;
   41995 }
   41996 #else
   41997 __ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
   41998   uint64x1_t __ret;
   41999   __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   42000   return __ret;
   42001 }
   42002 #endif
   42003 
   42004 #ifdef __LITTLE_ENDIAN__
   42005 __ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
   42006   uint64_t __ret;
   42007   __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
   42008   return __ret;
   42009 }
   42010 #else
   42011 __ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
   42012   uint64_t __ret;
   42013   __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
   42014   return __ret;
   42015 }
   42016 #endif
   42017 
   42018 #ifdef __LITTLE_ENDIAN__
   42019 __ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
   42020   uint32_t __ret;
   42021   __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
   42022   return __ret;
   42023 }
   42024 #else
   42025 __ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
   42026   uint32_t __ret;
   42027   __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
   42028   return __ret;
   42029 }
   42030 #endif
   42031 
   42032 #ifdef __LITTLE_ENDIAN__
   42033 __ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
   42034   uint64x1_t __ret;
   42035   __ret = (uint64x1_t)(__p0 == __p1);
   42036   return __ret;
   42037 }
   42038 #else
   42039 __ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
   42040   uint64x1_t __ret;
   42041   __ret = (uint64x1_t)(__p0 == __p1);
   42042   return __ret;
   42043 }
   42044 #endif
   42045 
   42046 #ifdef __LITTLE_ENDIAN__
   42047 __ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
   42048   uint64x2_t __ret;
   42049   __ret = (uint64x2_t)(__p0 == __p1);
   42050   return __ret;
   42051 }
   42052 #else
   42053 __ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
   42054   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42055   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   42056   uint64x2_t __ret;
   42057   __ret = (uint64x2_t)(__rev0 == __rev1);
   42058   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42059   return __ret;
   42060 }
   42061 #endif
   42062 
   42063 #ifdef __LITTLE_ENDIAN__
   42064 __ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   42065   uint64x2_t __ret;
   42066   __ret = (uint64x2_t)(__p0 == __p1);
   42067   return __ret;
   42068 }
   42069 #else
   42070 __ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   42071   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42072   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   42073   uint64x2_t __ret;
   42074   __ret = (uint64x2_t)(__rev0 == __rev1);
   42075   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42076   return __ret;
   42077 }
   42078 #endif
   42079 
   42080 #ifdef __LITTLE_ENDIAN__
   42081 __ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
   42082   uint64x2_t __ret;
   42083   __ret = (uint64x2_t)(__p0 == __p1);
   42084   return __ret;
   42085 }
   42086 #else
   42087 __ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
   42088   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42089   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   42090   uint64x2_t __ret;
   42091   __ret = (uint64x2_t)(__rev0 == __rev1);
   42092   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42093   return __ret;
   42094 }
   42095 #endif
   42096 
   42097 #ifdef __LITTLE_ENDIAN__
   42098 __ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
   42099   uint64x2_t __ret;
   42100   __ret = (uint64x2_t)(__p0 == __p1);
   42101   return __ret;
   42102 }
   42103 #else
   42104 __ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
   42105   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42106   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   42107   uint64x2_t __ret;
   42108   __ret = (uint64x2_t)(__rev0 == __rev1);
   42109   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42110   return __ret;
   42111 }
   42112 #endif
   42113 
   42114 #ifdef __LITTLE_ENDIAN__
   42115 __ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
   42116   uint64x1_t __ret;
   42117   __ret = (uint64x1_t)(__p0 == __p1);
   42118   return __ret;
   42119 }
   42120 #else
   42121 __ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
   42122   uint64x1_t __ret;
   42123   __ret = (uint64x1_t)(__p0 == __p1);
   42124   return __ret;
   42125 }
   42126 #endif
   42127 
   42128 #ifdef __LITTLE_ENDIAN__
   42129 __ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
   42130   uint64x1_t __ret;
   42131   __ret = (uint64x1_t)(__p0 == __p1);
   42132   return __ret;
   42133 }
   42134 #else
   42135 __ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
   42136   uint64x1_t __ret;
   42137   __ret = (uint64x1_t)(__p0 == __p1);
   42138   return __ret;
   42139 }
   42140 #endif
   42141 
   42142 #ifdef __LITTLE_ENDIAN__
   42143 __ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
   42144   uint64x1_t __ret;
   42145   __ret = (uint64x1_t)(__p0 == __p1);
   42146   return __ret;
   42147 }
   42148 #else
   42149 __ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
   42150   uint64x1_t __ret;
   42151   __ret = (uint64x1_t)(__p0 == __p1);
   42152   return __ret;
   42153 }
   42154 #endif
   42155 
   42156 #ifdef __LITTLE_ENDIAN__
   42157 __ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
   42158   uint64_t __ret;
   42159   __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
   42160   return __ret;
   42161 }
   42162 #else
   42163 __ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
   42164   uint64_t __ret;
   42165   __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
   42166   return __ret;
   42167 }
   42168 #endif
   42169 
   42170 #ifdef __LITTLE_ENDIAN__
   42171 __ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
   42172   int64_t __ret;
   42173   __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
   42174   return __ret;
   42175 }
   42176 #else
   42177 __ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
   42178   int64_t __ret;
   42179   __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
   42180   return __ret;
   42181 }
   42182 #endif
   42183 
   42184 #ifdef __LITTLE_ENDIAN__
   42185 __ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
   42186   uint64_t __ret;
   42187   __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
   42188   return __ret;
   42189 }
   42190 #else
   42191 __ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
   42192   uint64_t __ret;
   42193   __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
   42194   return __ret;
   42195 }
   42196 #endif
   42197 
   42198 #ifdef __LITTLE_ENDIAN__
   42199 __ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
   42200   uint32_t __ret;
   42201   __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
   42202   return __ret;
   42203 }
   42204 #else
   42205 __ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
   42206   uint32_t __ret;
   42207   __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
   42208   return __ret;
   42209 }
   42210 #endif
   42211 
   42212 #ifdef __LITTLE_ENDIAN__
   42213 __ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
   42214   uint8x8_t __ret;
   42215   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
   42216   return __ret;
   42217 }
   42218 #else
   42219 __ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
   42220   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   42221   uint8x8_t __ret;
   42222   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
   42223   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   42224   return __ret;
   42225 }
   42226 #endif
   42227 
   42228 #ifdef __LITTLE_ENDIAN__
   42229 __ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
   42230   uint64x1_t __ret;
   42231   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
   42232   return __ret;
   42233 }
   42234 #else
   42235 __ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
   42236   uint64x1_t __ret;
   42237   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
   42238   return __ret;
   42239 }
   42240 #endif
   42241 
   42242 #ifdef __LITTLE_ENDIAN__
   42243 __ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
   42244   uint16x4_t __ret;
   42245   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
   42246   return __ret;
   42247 }
   42248 #else
   42249 __ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
   42250   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   42251   uint16x4_t __ret;
   42252   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
   42253   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   42254   return __ret;
   42255 }
   42256 #endif
   42257 
   42258 #ifdef __LITTLE_ENDIAN__
   42259 __ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
   42260   uint8x16_t __ret;
   42261   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
   42262   return __ret;
   42263 }
   42264 #else
   42265 __ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
   42266   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   42267   uint8x16_t __ret;
   42268   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
   42269   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   42270   return __ret;
   42271 }
   42272 #endif
   42273 
   42274 #ifdef __LITTLE_ENDIAN__
   42275 __ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
   42276   uint64x2_t __ret;
   42277   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
   42278   return __ret;
   42279 }
   42280 #else
   42281 __ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
   42282   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42283   uint64x2_t __ret;
   42284   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
   42285   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42286   return __ret;
   42287 }
   42288 #endif
   42289 
   42290 #ifdef __LITTLE_ENDIAN__
   42291 __ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
   42292   uint16x8_t __ret;
   42293   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
   42294   return __ret;
   42295 }
   42296 #else
   42297 __ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
   42298   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   42299   uint16x8_t __ret;
   42300   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
   42301   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   42302   return __ret;
   42303 }
   42304 #endif
   42305 
   42306 #ifdef __LITTLE_ENDIAN__
   42307 __ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
   42308   uint8x16_t __ret;
   42309   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
   42310   return __ret;
   42311 }
   42312 #else
   42313 __ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
   42314   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   42315   uint8x16_t __ret;
   42316   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
   42317   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   42318   return __ret;
   42319 }
   42320 #endif
   42321 
   42322 #ifdef __LITTLE_ENDIAN__
   42323 __ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
   42324   uint32x4_t __ret;
   42325   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
   42326   return __ret;
   42327 }
   42328 #else
   42329 __ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
   42330   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   42331   uint32x4_t __ret;
   42332   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
   42333   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   42334   return __ret;
   42335 }
   42336 #endif
   42337 
   42338 #ifdef __LITTLE_ENDIAN__
   42339 __ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
   42340   uint64x2_t __ret;
   42341   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
   42342   return __ret;
   42343 }
   42344 #else
   42345 __ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
   42346   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42347   uint64x2_t __ret;
   42348   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
   42349   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42350   return __ret;
   42351 }
   42352 #endif
   42353 
   42354 #ifdef __LITTLE_ENDIAN__
   42355 __ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
   42356   uint16x8_t __ret;
   42357   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
   42358   return __ret;
   42359 }
   42360 #else
   42361 __ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
   42362   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   42363   uint16x8_t __ret;
   42364   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
   42365   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   42366   return __ret;
   42367 }
   42368 #endif
   42369 
   42370 #ifdef __LITTLE_ENDIAN__
   42371 __ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
   42372   uint8x16_t __ret;
   42373   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
   42374   return __ret;
   42375 }
   42376 #else
   42377 __ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
   42378   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   42379   uint8x16_t __ret;
   42380   __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
   42381   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   42382   return __ret;
   42383 }
   42384 #endif
   42385 
   42386 #ifdef __LITTLE_ENDIAN__
   42387 __ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
   42388   uint64x2_t __ret;
   42389   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
   42390   return __ret;
   42391 }
   42392 #else
   42393 __ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
   42394   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42395   uint64x2_t __ret;
   42396   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
   42397   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42398   return __ret;
   42399 }
   42400 #endif
   42401 
   42402 #ifdef __LITTLE_ENDIAN__
   42403 __ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
   42404   uint32x4_t __ret;
   42405   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
   42406   return __ret;
   42407 }
   42408 #else
   42409 __ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
   42410   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   42411   uint32x4_t __ret;
   42412   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
   42413   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   42414   return __ret;
   42415 }
   42416 #endif
   42417 
   42418 #ifdef __LITTLE_ENDIAN__
   42419 __ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
   42420   uint32x4_t __ret;
   42421   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
   42422   return __ret;
   42423 }
   42424 #else
   42425 __ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
   42426   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   42427   uint32x4_t __ret;
   42428   __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
   42429   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   42430   return __ret;
   42431 }
   42432 #endif
   42433 
   42434 #ifdef __LITTLE_ENDIAN__
   42435 __ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
   42436   uint64x2_t __ret;
   42437   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
   42438   return __ret;
   42439 }
   42440 #else
   42441 __ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
   42442   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42443   uint64x2_t __ret;
   42444   __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
   42445   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42446   return __ret;
   42447 }
   42448 #endif
   42449 
   42450 #ifdef __LITTLE_ENDIAN__
   42451 __ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
   42452   uint16x8_t __ret;
   42453   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
   42454   return __ret;
   42455 }
   42456 #else
   42457 __ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
   42458   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   42459   uint16x8_t __ret;
   42460   __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
   42461   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   42462   return __ret;
   42463 }
   42464 #endif
   42465 
   42466 #ifdef __LITTLE_ENDIAN__
   42467 __ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
   42468   uint8x8_t __ret;
   42469   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
   42470   return __ret;
   42471 }
   42472 #else
   42473 __ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
   42474   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   42475   uint8x8_t __ret;
   42476   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
   42477   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   42478   return __ret;
   42479 }
   42480 #endif
   42481 
   42482 #ifdef __LITTLE_ENDIAN__
   42483 __ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
   42484   uint32x2_t __ret;
   42485   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
   42486   return __ret;
   42487 }
   42488 #else
   42489 __ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
   42490   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42491   uint32x2_t __ret;
   42492   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
   42493   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42494   return __ret;
   42495 }
   42496 #endif
   42497 
   42498 #ifdef __LITTLE_ENDIAN__
   42499 __ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
   42500   uint64x1_t __ret;
   42501   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
   42502   return __ret;
   42503 }
   42504 #else
   42505 __ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
   42506   uint64x1_t __ret;
   42507   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
   42508   return __ret;
   42509 }
   42510 #endif
   42511 
   42512 #ifdef __LITTLE_ENDIAN__
   42513 __ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
   42514   uint16x4_t __ret;
   42515   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
   42516   return __ret;
   42517 }
   42518 #else
   42519 __ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
   42520   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   42521   uint16x4_t __ret;
   42522   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
   42523   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   42524   return __ret;
   42525 }
   42526 #endif
   42527 
   42528 #ifdef __LITTLE_ENDIAN__
   42529 __ai uint8x8_t vceqz_s8(int8x8_t __p0) {
   42530   uint8x8_t __ret;
   42531   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
   42532   return __ret;
   42533 }
   42534 #else
   42535 __ai uint8x8_t vceqz_s8(int8x8_t __p0) {
   42536   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   42537   uint8x8_t __ret;
   42538   __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
   42539   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   42540   return __ret;
   42541 }
   42542 #endif
   42543 
   42544 #ifdef __LITTLE_ENDIAN__
   42545 __ai uint64x1_t vceqz_f64(float64x1_t __p0) {
   42546   uint64x1_t __ret;
   42547   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
   42548   return __ret;
   42549 }
   42550 #else
   42551 __ai uint64x1_t vceqz_f64(float64x1_t __p0) {
   42552   uint64x1_t __ret;
   42553   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
   42554   return __ret;
   42555 }
   42556 #endif
   42557 
   42558 #ifdef __LITTLE_ENDIAN__
   42559 __ai uint32x2_t vceqz_f32(float32x2_t __p0) {
   42560   uint32x2_t __ret;
   42561   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
   42562   return __ret;
   42563 }
   42564 #else
   42565 __ai uint32x2_t vceqz_f32(float32x2_t __p0) {
   42566   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42567   uint32x2_t __ret;
   42568   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
   42569   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42570   return __ret;
   42571 }
   42572 #endif
   42573 
   42574 #ifdef __LITTLE_ENDIAN__
   42575 __ai uint32x2_t vceqz_s32(int32x2_t __p0) {
   42576   uint32x2_t __ret;
   42577   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
   42578   return __ret;
   42579 }
   42580 #else
   42581 __ai uint32x2_t vceqz_s32(int32x2_t __p0) {
   42582   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42583   uint32x2_t __ret;
   42584   __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
   42585   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42586   return __ret;
   42587 }
   42588 #endif
   42589 
   42590 #ifdef __LITTLE_ENDIAN__
   42591 __ai uint64x1_t vceqz_s64(int64x1_t __p0) {
   42592   uint64x1_t __ret;
   42593   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
   42594   return __ret;
   42595 }
   42596 #else
   42597 __ai uint64x1_t vceqz_s64(int64x1_t __p0) {
   42598   uint64x1_t __ret;
   42599   __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
   42600   return __ret;
   42601 }
   42602 #endif
   42603 
   42604 #ifdef __LITTLE_ENDIAN__
   42605 __ai uint16x4_t vceqz_s16(int16x4_t __p0) {
   42606   uint16x4_t __ret;
   42607   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
   42608   return __ret;
   42609 }
   42610 #else
   42611 __ai uint16x4_t vceqz_s16(int16x4_t __p0) {
   42612   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   42613   uint16x4_t __ret;
   42614   __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
   42615   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   42616   return __ret;
   42617 }
   42618 #endif
   42619 
   42620 #ifdef __LITTLE_ENDIAN__
   42621 __ai uint64_t vceqzd_u64(uint64_t __p0) {
   42622   uint64_t __ret;
   42623   __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
   42624   return __ret;
   42625 }
   42626 #else
   42627 __ai uint64_t vceqzd_u64(uint64_t __p0) {
   42628   uint64_t __ret;
   42629   __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
   42630   return __ret;
   42631 }
   42632 #endif
   42633 
   42634 #ifdef __LITTLE_ENDIAN__
   42635 __ai int64_t vceqzd_s64(int64_t __p0) {
   42636   int64_t __ret;
   42637   __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
   42638   return __ret;
   42639 }
   42640 #else
   42641 __ai int64_t vceqzd_s64(int64_t __p0) {
   42642   int64_t __ret;
   42643   __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
   42644   return __ret;
   42645 }
   42646 #endif
   42647 
   42648 #ifdef __LITTLE_ENDIAN__
   42649 __ai uint64_t vceqzd_f64(float64_t __p0) {
   42650   uint64_t __ret;
   42651   __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
   42652   return __ret;
   42653 }
   42654 #else
   42655 __ai uint64_t vceqzd_f64(float64_t __p0) {
   42656   uint64_t __ret;
   42657   __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
   42658   return __ret;
   42659 }
   42660 #endif
   42661 
   42662 #ifdef __LITTLE_ENDIAN__
   42663 __ai uint32_t vceqzs_f32(float32_t __p0) {
   42664   uint32_t __ret;
   42665   __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
   42666   return __ret;
   42667 }
   42668 #else
   42669 __ai uint32_t vceqzs_f32(float32_t __p0) {
   42670   uint32_t __ret;
   42671   __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
   42672   return __ret;
   42673 }
   42674 #endif
   42675 
   42676 #ifdef __LITTLE_ENDIAN__
   42677 __ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   42678   uint64x2_t __ret;
   42679   __ret = (uint64x2_t)(__p0 >= __p1);
   42680   return __ret;
   42681 }
   42682 #else
   42683 __ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   42684   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42685   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   42686   uint64x2_t __ret;
   42687   __ret = (uint64x2_t)(__rev0 >= __rev1);
   42688   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42689   return __ret;
   42690 }
   42691 #endif
   42692 
   42693 #ifdef __LITTLE_ENDIAN__
   42694 __ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
   42695   uint64x2_t __ret;
   42696   __ret = (uint64x2_t)(__p0 >= __p1);
   42697   return __ret;
   42698 }
   42699 #else
   42700 __ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
   42701   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42702   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   42703   uint64x2_t __ret;
   42704   __ret = (uint64x2_t)(__rev0 >= __rev1);
   42705   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42706   return __ret;
   42707 }
   42708 #endif
   42709 
   42710 #ifdef __LITTLE_ENDIAN__
   42711 __ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
   42712   uint64x2_t __ret;
   42713   __ret = (uint64x2_t)(__p0 >= __p1);
   42714   return __ret;
   42715 }
   42716 #else
   42717 __ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
   42718   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42719   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   42720   uint64x2_t __ret;
   42721   __ret = (uint64x2_t)(__rev0 >= __rev1);
   42722   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42723   return __ret;
   42724 }
   42725 #endif
   42726 
   42727 #ifdef __LITTLE_ENDIAN__
   42728 __ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
   42729   uint64x1_t __ret;
   42730   __ret = (uint64x1_t)(__p0 >= __p1);
   42731   return __ret;
   42732 }
   42733 #else
   42734 __ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
   42735   uint64x1_t __ret;
   42736   __ret = (uint64x1_t)(__p0 >= __p1);
   42737   return __ret;
   42738 }
   42739 #endif
   42740 
   42741 #ifdef __LITTLE_ENDIAN__
   42742 __ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
   42743   uint64x1_t __ret;
   42744   __ret = (uint64x1_t)(__p0 >= __p1);
   42745   return __ret;
   42746 }
   42747 #else
   42748 __ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
   42749   uint64x1_t __ret;
   42750   __ret = (uint64x1_t)(__p0 >= __p1);
   42751   return __ret;
   42752 }
   42753 #endif
   42754 
   42755 #ifdef __LITTLE_ENDIAN__
   42756 __ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
   42757   uint64x1_t __ret;
   42758   __ret = (uint64x1_t)(__p0 >= __p1);
   42759   return __ret;
   42760 }
   42761 #else
   42762 __ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
   42763   uint64x1_t __ret;
   42764   __ret = (uint64x1_t)(__p0 >= __p1);
   42765   return __ret;
   42766 }
   42767 #endif
   42768 
   42769 #ifdef __LITTLE_ENDIAN__
   42770 __ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
   42771   int64_t __ret;
   42772   __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
   42773   return __ret;
   42774 }
   42775 #else
   42776 __ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
   42777   int64_t __ret;
   42778   __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
   42779   return __ret;
   42780 }
   42781 #endif
   42782 
   42783 #ifdef __LITTLE_ENDIAN__
   42784 __ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
   42785   uint64_t __ret;
   42786   __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
   42787   return __ret;
   42788 }
   42789 #else
   42790 __ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
   42791   uint64_t __ret;
   42792   __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
   42793   return __ret;
   42794 }
   42795 #endif
   42796 
   42797 #ifdef __LITTLE_ENDIAN__
   42798 __ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
   42799   uint64_t __ret;
   42800   __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
   42801   return __ret;
   42802 }
   42803 #else
   42804 __ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
   42805   uint64_t __ret;
   42806   __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
   42807   return __ret;
   42808 }
   42809 #endif
   42810 
   42811 #ifdef __LITTLE_ENDIAN__
   42812 __ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
   42813   uint32_t __ret;
   42814   __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
   42815   return __ret;
   42816 }
   42817 #else
   42818 __ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
   42819   uint32_t __ret;
   42820   __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
   42821   return __ret;
   42822 }
   42823 #endif
   42824 
   42825 #ifdef __LITTLE_ENDIAN__
   42826 __ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
   42827   uint8x16_t __ret;
   42828   __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 48);
   42829   return __ret;
   42830 }
   42831 #else
   42832 __ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
   42833   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   42834   uint8x16_t __ret;
   42835   __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 48);
   42836   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   42837   return __ret;
   42838 }
   42839 #endif
   42840 
   42841 #ifdef __LITTLE_ENDIAN__
   42842 __ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
   42843   uint64x2_t __ret;
   42844   __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
   42845   return __ret;
   42846 }
   42847 #else
   42848 __ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
   42849   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42850   uint64x2_t __ret;
   42851   __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
   42852   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42853   return __ret;
   42854 }
   42855 #endif
   42856 
   42857 #ifdef __LITTLE_ENDIAN__
   42858 __ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
   42859   uint32x4_t __ret;
   42860   __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
   42861   return __ret;
   42862 }
   42863 #else
   42864 __ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
   42865   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   42866   uint32x4_t __ret;
   42867   __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
   42868   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   42869   return __ret;
   42870 }
   42871 #endif
   42872 
   42873 #ifdef __LITTLE_ENDIAN__
   42874 __ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
   42875   uint32x4_t __ret;
   42876   __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
   42877   return __ret;
   42878 }
   42879 #else
   42880 __ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
   42881   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   42882   uint32x4_t __ret;
   42883   __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
   42884   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   42885   return __ret;
   42886 }
   42887 #endif
   42888 
   42889 #ifdef __LITTLE_ENDIAN__
   42890 __ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
   42891   uint64x2_t __ret;
   42892   __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
   42893   return __ret;
   42894 }
   42895 #else
   42896 __ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
   42897   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42898   uint64x2_t __ret;
   42899   __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
   42900   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42901   return __ret;
   42902 }
   42903 #endif
   42904 
   42905 #ifdef __LITTLE_ENDIAN__
   42906 __ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
   42907   uint16x8_t __ret;
   42908   __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 49);
   42909   return __ret;
   42910 }
   42911 #else
   42912 __ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
   42913   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   42914   uint16x8_t __ret;
   42915   __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 49);
   42916   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   42917   return __ret;
   42918 }
   42919 #endif
   42920 
   42921 #ifdef __LITTLE_ENDIAN__
   42922 __ai uint8x8_t vcgez_s8(int8x8_t __p0) {
   42923   uint8x8_t __ret;
   42924   __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 16);
   42925   return __ret;
   42926 }
   42927 #else
   42928 __ai uint8x8_t vcgez_s8(int8x8_t __p0) {
   42929   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   42930   uint8x8_t __ret;
   42931   __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 16);
   42932   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   42933   return __ret;
   42934 }
   42935 #endif
   42936 
   42937 #ifdef __LITTLE_ENDIAN__
   42938 __ai uint64x1_t vcgez_f64(float64x1_t __p0) {
   42939   uint64x1_t __ret;
   42940   __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
   42941   return __ret;
   42942 }
   42943 #else
   42944 __ai uint64x1_t vcgez_f64(float64x1_t __p0) {
   42945   uint64x1_t __ret;
   42946   __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
   42947   return __ret;
   42948 }
   42949 #endif
   42950 
   42951 #ifdef __LITTLE_ENDIAN__
   42952 __ai uint32x2_t vcgez_f32(float32x2_t __p0) {
   42953   uint32x2_t __ret;
   42954   __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
   42955   return __ret;
   42956 }
   42957 #else
   42958 __ai uint32x2_t vcgez_f32(float32x2_t __p0) {
   42959   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42960   uint32x2_t __ret;
   42961   __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
   42962   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42963   return __ret;
   42964 }
   42965 #endif
   42966 
   42967 #ifdef __LITTLE_ENDIAN__
   42968 __ai uint32x2_t vcgez_s32(int32x2_t __p0) {
   42969   uint32x2_t __ret;
   42970   __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
   42971   return __ret;
   42972 }
   42973 #else
   42974 __ai uint32x2_t vcgez_s32(int32x2_t __p0) {
   42975   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   42976   uint32x2_t __ret;
   42977   __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
   42978   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   42979   return __ret;
   42980 }
   42981 #endif
   42982 
   42983 #ifdef __LITTLE_ENDIAN__
   42984 __ai uint64x1_t vcgez_s64(int64x1_t __p0) {
   42985   uint64x1_t __ret;
   42986   __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
   42987   return __ret;
   42988 }
   42989 #else
   42990 __ai uint64x1_t vcgez_s64(int64x1_t __p0) {
   42991   uint64x1_t __ret;
   42992   __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
   42993   return __ret;
   42994 }
   42995 #endif
   42996 
   42997 #ifdef __LITTLE_ENDIAN__
   42998 __ai uint16x4_t vcgez_s16(int16x4_t __p0) {
   42999   uint16x4_t __ret;
   43000   __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 17);
   43001   return __ret;
   43002 }
   43003 #else
   43004 __ai uint16x4_t vcgez_s16(int16x4_t __p0) {
   43005   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   43006   uint16x4_t __ret;
   43007   __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 17);
   43008   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   43009   return __ret;
   43010 }
   43011 #endif
   43012 
   43013 #ifdef __LITTLE_ENDIAN__
   43014 __ai int64_t vcgezd_s64(int64_t __p0) {
   43015   int64_t __ret;
   43016   __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
   43017   return __ret;
   43018 }
   43019 #else
   43020 __ai int64_t vcgezd_s64(int64_t __p0) {
   43021   int64_t __ret;
   43022   __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
   43023   return __ret;
   43024 }
   43025 #endif
   43026 
   43027 #ifdef __LITTLE_ENDIAN__
   43028 __ai uint64_t vcgezd_f64(float64_t __p0) {
   43029   uint64_t __ret;
   43030   __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
   43031   return __ret;
   43032 }
   43033 #else
   43034 __ai uint64_t vcgezd_f64(float64_t __p0) {
   43035   uint64_t __ret;
   43036   __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
   43037   return __ret;
   43038 }
   43039 #endif
   43040 
   43041 #ifdef __LITTLE_ENDIAN__
   43042 __ai uint32_t vcgezs_f32(float32_t __p0) {
   43043   uint32_t __ret;
   43044   __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
   43045   return __ret;
   43046 }
   43047 #else
   43048 __ai uint32_t vcgezs_f32(float32_t __p0) {
   43049   uint32_t __ret;
   43050   __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
   43051   return __ret;
   43052 }
   43053 #endif
   43054 
   43055 #ifdef __LITTLE_ENDIAN__
   43056 __ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   43057   uint64x2_t __ret;
   43058   __ret = (uint64x2_t)(__p0 > __p1);
   43059   return __ret;
   43060 }
   43061 #else
   43062 __ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   43063   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43064   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   43065   uint64x2_t __ret;
   43066   __ret = (uint64x2_t)(__rev0 > __rev1);
   43067   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43068   return __ret;
   43069 }
   43070 #endif
   43071 
   43072 #ifdef __LITTLE_ENDIAN__
   43073 __ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
   43074   uint64x2_t __ret;
   43075   __ret = (uint64x2_t)(__p0 > __p1);
   43076   return __ret;
   43077 }
   43078 #else
   43079 __ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
   43080   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43081   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   43082   uint64x2_t __ret;
   43083   __ret = (uint64x2_t)(__rev0 > __rev1);
   43084   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43085   return __ret;
   43086 }
   43087 #endif
   43088 
   43089 #ifdef __LITTLE_ENDIAN__
   43090 __ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
   43091   uint64x2_t __ret;
   43092   __ret = (uint64x2_t)(__p0 > __p1);
   43093   return __ret;
   43094 }
   43095 #else
   43096 __ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
   43097   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43098   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   43099   uint64x2_t __ret;
   43100   __ret = (uint64x2_t)(__rev0 > __rev1);
   43101   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43102   return __ret;
   43103 }
   43104 #endif
   43105 
   43106 #ifdef __LITTLE_ENDIAN__
   43107 __ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
   43108   uint64x1_t __ret;
   43109   __ret = (uint64x1_t)(__p0 > __p1);
   43110   return __ret;
   43111 }
   43112 #else
   43113 __ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
   43114   uint64x1_t __ret;
   43115   __ret = (uint64x1_t)(__p0 > __p1);
   43116   return __ret;
   43117 }
   43118 #endif
   43119 
   43120 #ifdef __LITTLE_ENDIAN__
   43121 __ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
   43122   uint64x1_t __ret;
   43123   __ret = (uint64x1_t)(__p0 > __p1);
   43124   return __ret;
   43125 }
   43126 #else
   43127 __ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
   43128   uint64x1_t __ret;
   43129   __ret = (uint64x1_t)(__p0 > __p1);
   43130   return __ret;
   43131 }
   43132 #endif
   43133 
   43134 #ifdef __LITTLE_ENDIAN__
   43135 __ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
   43136   uint64x1_t __ret;
   43137   __ret = (uint64x1_t)(__p0 > __p1);
   43138   return __ret;
   43139 }
   43140 #else
   43141 __ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
   43142   uint64x1_t __ret;
   43143   __ret = (uint64x1_t)(__p0 > __p1);
   43144   return __ret;
   43145 }
   43146 #endif
   43147 
   43148 #ifdef __LITTLE_ENDIAN__
   43149 __ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
   43150   int64_t __ret;
   43151   __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
   43152   return __ret;
   43153 }
   43154 #else
   43155 __ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
   43156   int64_t __ret;
   43157   __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
   43158   return __ret;
   43159 }
   43160 #endif
   43161 
   43162 #ifdef __LITTLE_ENDIAN__
   43163 __ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
   43164   uint64_t __ret;
   43165   __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
   43166   return __ret;
   43167 }
   43168 #else
   43169 __ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
   43170   uint64_t __ret;
   43171   __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
   43172   return __ret;
   43173 }
   43174 #endif
   43175 
   43176 #ifdef __LITTLE_ENDIAN__
   43177 __ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
   43178   uint64_t __ret;
   43179   __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
   43180   return __ret;
   43181 }
   43182 #else
   43183 __ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
   43184   uint64_t __ret;
   43185   __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
   43186   return __ret;
   43187 }
   43188 #endif
   43189 
   43190 #ifdef __LITTLE_ENDIAN__
   43191 __ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
   43192   uint32_t __ret;
   43193   __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
   43194   return __ret;
   43195 }
   43196 #else
   43197 __ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
   43198   uint32_t __ret;
   43199   __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
   43200   return __ret;
   43201 }
   43202 #endif
   43203 
   43204 #ifdef __LITTLE_ENDIAN__
   43205 __ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
   43206   uint8x16_t __ret;
   43207   __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 48);
   43208   return __ret;
   43209 }
   43210 #else
   43211 __ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
   43212   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   43213   uint8x16_t __ret;
   43214   __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 48);
   43215   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   43216   return __ret;
   43217 }
   43218 #endif
   43219 
   43220 #ifdef __LITTLE_ENDIAN__
   43221 __ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
   43222   uint64x2_t __ret;
   43223   __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
   43224   return __ret;
   43225 }
   43226 #else
   43227 __ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
   43228   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43229   uint64x2_t __ret;
   43230   __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
   43231   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43232   return __ret;
   43233 }
   43234 #endif
   43235 
   43236 #ifdef __LITTLE_ENDIAN__
   43237 __ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
   43238   uint32x4_t __ret;
   43239   __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
   43240   return __ret;
   43241 }
   43242 #else
   43243 __ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
   43244   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   43245   uint32x4_t __ret;
   43246   __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
   43247   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   43248   return __ret;
   43249 }
   43250 #endif
   43251 
   43252 #ifdef __LITTLE_ENDIAN__
   43253 __ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
   43254   uint32x4_t __ret;
   43255   __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
   43256   return __ret;
   43257 }
   43258 #else
   43259 __ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
   43260   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   43261   uint32x4_t __ret;
   43262   __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
   43263   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   43264   return __ret;
   43265 }
   43266 #endif
   43267 
   43268 #ifdef __LITTLE_ENDIAN__
   43269 __ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
   43270   uint64x2_t __ret;
   43271   __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
   43272   return __ret;
   43273 }
   43274 #else
   43275 __ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
   43276   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43277   uint64x2_t __ret;
   43278   __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
   43279   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43280   return __ret;
   43281 }
   43282 #endif
   43283 
   43284 #ifdef __LITTLE_ENDIAN__
   43285 __ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
   43286   uint16x8_t __ret;
   43287   __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 49);
   43288   return __ret;
   43289 }
   43290 #else
   43291 __ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
   43292   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   43293   uint16x8_t __ret;
   43294   __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 49);
   43295   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   43296   return __ret;
   43297 }
   43298 #endif
   43299 
   43300 #ifdef __LITTLE_ENDIAN__
   43301 __ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
   43302   uint8x8_t __ret;
   43303   __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 16);
   43304   return __ret;
   43305 }
   43306 #else
   43307 __ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
   43308   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   43309   uint8x8_t __ret;
   43310   __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 16);
   43311   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   43312   return __ret;
   43313 }
   43314 #endif
   43315 
   43316 #ifdef __LITTLE_ENDIAN__
   43317 __ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
   43318   uint64x1_t __ret;
   43319   __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
   43320   return __ret;
   43321 }
   43322 #else
   43323 __ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
   43324   uint64x1_t __ret;
   43325   __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
   43326   return __ret;
   43327 }
   43328 #endif
   43329 
   43330 #ifdef __LITTLE_ENDIAN__
   43331 __ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
   43332   uint32x2_t __ret;
   43333   __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
   43334   return __ret;
   43335 }
   43336 #else
   43337 __ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
   43338   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43339   uint32x2_t __ret;
   43340   __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
   43341   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43342   return __ret;
   43343 }
   43344 #endif
   43345 
   43346 #ifdef __LITTLE_ENDIAN__
   43347 __ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
   43348   uint32x2_t __ret;
   43349   __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
   43350   return __ret;
   43351 }
   43352 #else
   43353 __ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
   43354   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43355   uint32x2_t __ret;
   43356   __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
   43357   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43358   return __ret;
   43359 }
   43360 #endif
   43361 
   43362 #ifdef __LITTLE_ENDIAN__
   43363 __ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
   43364   uint64x1_t __ret;
   43365   __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
   43366   return __ret;
   43367 }
   43368 #else
   43369 __ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
   43370   uint64x1_t __ret;
   43371   __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
   43372   return __ret;
   43373 }
   43374 #endif
   43375 
   43376 #ifdef __LITTLE_ENDIAN__
   43377 __ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
   43378   uint16x4_t __ret;
   43379   __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 17);
   43380   return __ret;
   43381 }
   43382 #else
   43383 __ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
   43384   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   43385   uint16x4_t __ret;
   43386   __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 17);
   43387   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   43388   return __ret;
   43389 }
   43390 #endif
   43391 
   43392 #ifdef __LITTLE_ENDIAN__
   43393 __ai int64_t vcgtzd_s64(int64_t __p0) {
   43394   int64_t __ret;
   43395   __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
   43396   return __ret;
   43397 }
   43398 #else
   43399 __ai int64_t vcgtzd_s64(int64_t __p0) {
   43400   int64_t __ret;
   43401   __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
   43402   return __ret;
   43403 }
   43404 #endif
   43405 
   43406 #ifdef __LITTLE_ENDIAN__
   43407 __ai uint64_t vcgtzd_f64(float64_t __p0) {
   43408   uint64_t __ret;
   43409   __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
   43410   return __ret;
   43411 }
   43412 #else
   43413 __ai uint64_t vcgtzd_f64(float64_t __p0) {
   43414   uint64_t __ret;
   43415   __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
   43416   return __ret;
   43417 }
   43418 #endif
   43419 
   43420 #ifdef __LITTLE_ENDIAN__
   43421 __ai uint32_t vcgtzs_f32(float32_t __p0) {
   43422   uint32_t __ret;
   43423   __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
   43424   return __ret;
   43425 }
   43426 #else
   43427 __ai uint32_t vcgtzs_f32(float32_t __p0) {
   43428   uint32_t __ret;
   43429   __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
   43430   return __ret;
   43431 }
   43432 #endif
   43433 
   43434 #ifdef __LITTLE_ENDIAN__
   43435 __ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   43436   uint64x2_t __ret;
   43437   __ret = (uint64x2_t)(__p0 <= __p1);
   43438   return __ret;
   43439 }
   43440 #else
   43441 __ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   43442   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43443   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   43444   uint64x2_t __ret;
   43445   __ret = (uint64x2_t)(__rev0 <= __rev1);
   43446   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43447   return __ret;
   43448 }
   43449 #endif
   43450 
   43451 #ifdef __LITTLE_ENDIAN__
   43452 __ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
   43453   uint64x2_t __ret;
   43454   __ret = (uint64x2_t)(__p0 <= __p1);
   43455   return __ret;
   43456 }
   43457 #else
   43458 __ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
   43459   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43460   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   43461   uint64x2_t __ret;
   43462   __ret = (uint64x2_t)(__rev0 <= __rev1);
   43463   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43464   return __ret;
   43465 }
   43466 #endif
   43467 
   43468 #ifdef __LITTLE_ENDIAN__
   43469 __ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
   43470   uint64x2_t __ret;
   43471   __ret = (uint64x2_t)(__p0 <= __p1);
   43472   return __ret;
   43473 }
   43474 #else
   43475 __ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
   43476   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43477   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   43478   uint64x2_t __ret;
   43479   __ret = (uint64x2_t)(__rev0 <= __rev1);
   43480   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43481   return __ret;
   43482 }
   43483 #endif
   43484 
   43485 #ifdef __LITTLE_ENDIAN__
   43486 __ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
   43487   uint64x1_t __ret;
   43488   __ret = (uint64x1_t)(__p0 <= __p1);
   43489   return __ret;
   43490 }
   43491 #else
   43492 __ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
   43493   uint64x1_t __ret;
   43494   __ret = (uint64x1_t)(__p0 <= __p1);
   43495   return __ret;
   43496 }
   43497 #endif
   43498 
   43499 #ifdef __LITTLE_ENDIAN__
   43500 __ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
   43501   uint64x1_t __ret;
   43502   __ret = (uint64x1_t)(__p0 <= __p1);
   43503   return __ret;
   43504 }
   43505 #else
   43506 __ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
   43507   uint64x1_t __ret;
   43508   __ret = (uint64x1_t)(__p0 <= __p1);
   43509   return __ret;
   43510 }
   43511 #endif
   43512 
   43513 #ifdef __LITTLE_ENDIAN__
   43514 __ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
   43515   uint64x1_t __ret;
   43516   __ret = (uint64x1_t)(__p0 <= __p1);
   43517   return __ret;
   43518 }
   43519 #else
   43520 __ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
   43521   uint64x1_t __ret;
   43522   __ret = (uint64x1_t)(__p0 <= __p1);
   43523   return __ret;
   43524 }
   43525 #endif
   43526 
   43527 #ifdef __LITTLE_ENDIAN__
   43528 __ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
   43529   uint64_t __ret;
   43530   __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
   43531   return __ret;
   43532 }
   43533 #else
   43534 __ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
   43535   uint64_t __ret;
   43536   __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
   43537   return __ret;
   43538 }
   43539 #endif
   43540 
   43541 #ifdef __LITTLE_ENDIAN__
   43542 __ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
   43543   int64_t __ret;
   43544   __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
   43545   return __ret;
   43546 }
   43547 #else
   43548 __ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
   43549   int64_t __ret;
   43550   __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
   43551   return __ret;
   43552 }
   43553 #endif
   43554 
   43555 #ifdef __LITTLE_ENDIAN__
   43556 __ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
   43557   uint64_t __ret;
   43558   __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
   43559   return __ret;
   43560 }
   43561 #else
   43562 __ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
   43563   uint64_t __ret;
   43564   __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
   43565   return __ret;
   43566 }
   43567 #endif
   43568 
   43569 #ifdef __LITTLE_ENDIAN__
   43570 __ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
   43571   uint32_t __ret;
   43572   __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
   43573   return __ret;
   43574 }
   43575 #else
   43576 __ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
   43577   uint32_t __ret;
   43578   __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
   43579   return __ret;
   43580 }
   43581 #endif
   43582 
   43583 #ifdef __LITTLE_ENDIAN__
   43584 __ai uint8x16_t vclezq_s8(int8x16_t __p0) {
   43585   uint8x16_t __ret;
   43586   __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 48);
   43587   return __ret;
   43588 }
   43589 #else
   43590 __ai uint8x16_t vclezq_s8(int8x16_t __p0) {
   43591   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   43592   uint8x16_t __ret;
   43593   __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 48);
   43594   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   43595   return __ret;
   43596 }
   43597 #endif
   43598 
   43599 #ifdef __LITTLE_ENDIAN__
   43600 __ai uint64x2_t vclezq_f64(float64x2_t __p0) {
   43601   uint64x2_t __ret;
   43602   __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
   43603   return __ret;
   43604 }
   43605 #else
   43606 __ai uint64x2_t vclezq_f64(float64x2_t __p0) {
   43607   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43608   uint64x2_t __ret;
   43609   __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
   43610   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43611   return __ret;
   43612 }
   43613 #endif
   43614 
   43615 #ifdef __LITTLE_ENDIAN__
   43616 __ai uint32x4_t vclezq_f32(float32x4_t __p0) {
   43617   uint32x4_t __ret;
   43618   __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
   43619   return __ret;
   43620 }
   43621 #else
   43622 __ai uint32x4_t vclezq_f32(float32x4_t __p0) {
   43623   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   43624   uint32x4_t __ret;
   43625   __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
   43626   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   43627   return __ret;
   43628 }
   43629 #endif
   43630 
   43631 #ifdef __LITTLE_ENDIAN__
   43632 __ai uint32x4_t vclezq_s32(int32x4_t __p0) {
   43633   uint32x4_t __ret;
   43634   __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
   43635   return __ret;
   43636 }
   43637 #else
   43638 __ai uint32x4_t vclezq_s32(int32x4_t __p0) {
   43639   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   43640   uint32x4_t __ret;
   43641   __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
   43642   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   43643   return __ret;
   43644 }
   43645 #endif
   43646 
   43647 #ifdef __LITTLE_ENDIAN__
   43648 __ai uint64x2_t vclezq_s64(int64x2_t __p0) {
   43649   uint64x2_t __ret;
   43650   __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
   43651   return __ret;
   43652 }
   43653 #else
   43654 __ai uint64x2_t vclezq_s64(int64x2_t __p0) {
   43655   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43656   uint64x2_t __ret;
   43657   __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
   43658   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43659   return __ret;
   43660 }
   43661 #endif
   43662 
   43663 #ifdef __LITTLE_ENDIAN__
   43664 __ai uint16x8_t vclezq_s16(int16x8_t __p0) {
   43665   uint16x8_t __ret;
   43666   __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 49);
   43667   return __ret;
   43668 }
   43669 #else
   43670 __ai uint16x8_t vclezq_s16(int16x8_t __p0) {
   43671   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   43672   uint16x8_t __ret;
   43673   __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 49);
   43674   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   43675   return __ret;
   43676 }
   43677 #endif
   43678 
   43679 #ifdef __LITTLE_ENDIAN__
   43680 __ai uint8x8_t vclez_s8(int8x8_t __p0) {
   43681   uint8x8_t __ret;
   43682   __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__p0, 16);
   43683   return __ret;
   43684 }
   43685 #else
   43686 __ai uint8x8_t vclez_s8(int8x8_t __p0) {
   43687   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   43688   uint8x8_t __ret;
   43689   __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 16);
   43690   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   43691   return __ret;
   43692 }
   43693 #endif
   43694 
   43695 #ifdef __LITTLE_ENDIAN__
   43696 __ai uint64x1_t vclez_f64(float64x1_t __p0) {
   43697   uint64x1_t __ret;
   43698   __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
   43699   return __ret;
   43700 }
   43701 #else
   43702 __ai uint64x1_t vclez_f64(float64x1_t __p0) {
   43703   uint64x1_t __ret;
   43704   __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
   43705   return __ret;
   43706 }
   43707 #endif
   43708 
   43709 #ifdef __LITTLE_ENDIAN__
   43710 __ai uint32x2_t vclez_f32(float32x2_t __p0) {
   43711   uint32x2_t __ret;
   43712   __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
   43713   return __ret;
   43714 }
   43715 #else
   43716 __ai uint32x2_t vclez_f32(float32x2_t __p0) {
   43717   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43718   uint32x2_t __ret;
   43719   __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
   43720   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43721   return __ret;
   43722 }
   43723 #endif
   43724 
   43725 #ifdef __LITTLE_ENDIAN__
   43726 __ai uint32x2_t vclez_s32(int32x2_t __p0) {
   43727   uint32x2_t __ret;
   43728   __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
   43729   return __ret;
   43730 }
   43731 #else
   43732 __ai uint32x2_t vclez_s32(int32x2_t __p0) {
   43733   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43734   uint32x2_t __ret;
   43735   __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
   43736   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43737   return __ret;
   43738 }
   43739 #endif
   43740 
   43741 #ifdef __LITTLE_ENDIAN__
   43742 __ai uint64x1_t vclez_s64(int64x1_t __p0) {
   43743   uint64x1_t __ret;
   43744   __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
   43745   return __ret;
   43746 }
   43747 #else
   43748 __ai uint64x1_t vclez_s64(int64x1_t __p0) {
   43749   uint64x1_t __ret;
   43750   __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
   43751   return __ret;
   43752 }
   43753 #endif
   43754 
   43755 #ifdef __LITTLE_ENDIAN__
   43756 __ai uint16x4_t vclez_s16(int16x4_t __p0) {
   43757   uint16x4_t __ret;
   43758   __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__p0, 17);
   43759   return __ret;
   43760 }
   43761 #else
   43762 __ai uint16x4_t vclez_s16(int16x4_t __p0) {
   43763   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   43764   uint16x4_t __ret;
   43765   __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 17);
   43766   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   43767   return __ret;
   43768 }
   43769 #endif
   43770 
   43771 #ifdef __LITTLE_ENDIAN__
   43772 __ai int64_t vclezd_s64(int64_t __p0) {
   43773   int64_t __ret;
   43774   __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
   43775   return __ret;
   43776 }
   43777 #else
   43778 __ai int64_t vclezd_s64(int64_t __p0) {
   43779   int64_t __ret;
   43780   __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
   43781   return __ret;
   43782 }
   43783 #endif
   43784 
   43785 #ifdef __LITTLE_ENDIAN__
   43786 __ai uint64_t vclezd_f64(float64_t __p0) {
   43787   uint64_t __ret;
   43788   __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
   43789   return __ret;
   43790 }
   43791 #else
   43792 __ai uint64_t vclezd_f64(float64_t __p0) {
   43793   uint64_t __ret;
   43794   __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
   43795   return __ret;
   43796 }
   43797 #endif
   43798 
   43799 #ifdef __LITTLE_ENDIAN__
   43800 __ai uint32_t vclezs_f32(float32_t __p0) {
   43801   uint32_t __ret;
   43802   __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
   43803   return __ret;
   43804 }
   43805 #else
   43806 __ai uint32_t vclezs_f32(float32_t __p0) {
   43807   uint32_t __ret;
   43808   __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
   43809   return __ret;
   43810 }
   43811 #endif
   43812 
   43813 #ifdef __LITTLE_ENDIAN__
   43814 __ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   43815   uint64x2_t __ret;
   43816   __ret = (uint64x2_t)(__p0 < __p1);
   43817   return __ret;
   43818 }
   43819 #else
   43820 __ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   43821   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43822   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   43823   uint64x2_t __ret;
   43824   __ret = (uint64x2_t)(__rev0 < __rev1);
   43825   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43826   return __ret;
   43827 }
   43828 #endif
   43829 
   43830 #ifdef __LITTLE_ENDIAN__
   43831 __ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
   43832   uint64x2_t __ret;
   43833   __ret = (uint64x2_t)(__p0 < __p1);
   43834   return __ret;
   43835 }
   43836 #else
   43837 __ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
   43838   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43839   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   43840   uint64x2_t __ret;
   43841   __ret = (uint64x2_t)(__rev0 < __rev1);
   43842   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43843   return __ret;
   43844 }
   43845 #endif
   43846 
   43847 #ifdef __LITTLE_ENDIAN__
   43848 __ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
   43849   uint64x2_t __ret;
   43850   __ret = (uint64x2_t)(__p0 < __p1);
   43851   return __ret;
   43852 }
   43853 #else
   43854 __ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
   43855   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43856   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   43857   uint64x2_t __ret;
   43858   __ret = (uint64x2_t)(__rev0 < __rev1);
   43859   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43860   return __ret;
   43861 }
   43862 #endif
   43863 
   43864 #ifdef __LITTLE_ENDIAN__
   43865 __ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
   43866   uint64x1_t __ret;
   43867   __ret = (uint64x1_t)(__p0 < __p1);
   43868   return __ret;
   43869 }
   43870 #else
   43871 __ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
   43872   uint64x1_t __ret;
   43873   __ret = (uint64x1_t)(__p0 < __p1);
   43874   return __ret;
   43875 }
   43876 #endif
   43877 
   43878 #ifdef __LITTLE_ENDIAN__
   43879 __ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
   43880   uint64x1_t __ret;
   43881   __ret = (uint64x1_t)(__p0 < __p1);
   43882   return __ret;
   43883 }
   43884 #else
   43885 __ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
   43886   uint64x1_t __ret;
   43887   __ret = (uint64x1_t)(__p0 < __p1);
   43888   return __ret;
   43889 }
   43890 #endif
   43891 
   43892 #ifdef __LITTLE_ENDIAN__
   43893 __ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
   43894   uint64x1_t __ret;
   43895   __ret = (uint64x1_t)(__p0 < __p1);
   43896   return __ret;
   43897 }
   43898 #else
   43899 __ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
   43900   uint64x1_t __ret;
   43901   __ret = (uint64x1_t)(__p0 < __p1);
   43902   return __ret;
   43903 }
   43904 #endif
   43905 
   43906 #ifdef __LITTLE_ENDIAN__
   43907 __ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
   43908   uint64_t __ret;
   43909   __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
   43910   return __ret;
   43911 }
   43912 #else
   43913 __ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
   43914   uint64_t __ret;
   43915   __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
   43916   return __ret;
   43917 }
   43918 #endif
   43919 
   43920 #ifdef __LITTLE_ENDIAN__
   43921 __ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
   43922   int64_t __ret;
   43923   __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
   43924   return __ret;
   43925 }
   43926 #else
   43927 __ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
   43928   int64_t __ret;
   43929   __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
   43930   return __ret;
   43931 }
   43932 #endif
   43933 
   43934 #ifdef __LITTLE_ENDIAN__
   43935 __ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
   43936   uint64_t __ret;
   43937   __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
   43938   return __ret;
   43939 }
   43940 #else
   43941 __ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
   43942   uint64_t __ret;
   43943   __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
   43944   return __ret;
   43945 }
   43946 #endif
   43947 
   43948 #ifdef __LITTLE_ENDIAN__
   43949 __ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
   43950   uint32_t __ret;
   43951   __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
   43952   return __ret;
   43953 }
   43954 #else
   43955 __ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
   43956   uint32_t __ret;
   43957   __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
   43958   return __ret;
   43959 }
   43960 #endif
   43961 
   43962 #ifdef __LITTLE_ENDIAN__
   43963 __ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
   43964   uint8x16_t __ret;
   43965   __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 48);
   43966   return __ret;
   43967 }
   43968 #else
   43969 __ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
   43970   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   43971   uint8x16_t __ret;
   43972   __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 48);
   43973   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   43974   return __ret;
   43975 }
   43976 #endif
   43977 
   43978 #ifdef __LITTLE_ENDIAN__
   43979 __ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
   43980   uint64x2_t __ret;
   43981   __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
   43982   return __ret;
   43983 }
   43984 #else
   43985 __ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
   43986   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   43987   uint64x2_t __ret;
   43988   __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
   43989   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   43990   return __ret;
   43991 }
   43992 #endif
   43993 
   43994 #ifdef __LITTLE_ENDIAN__
   43995 __ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
   43996   uint32x4_t __ret;
   43997   __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
   43998   return __ret;
   43999 }
   44000 #else
   44001 __ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
   44002   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   44003   uint32x4_t __ret;
   44004   __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
   44005   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   44006   return __ret;
   44007 }
   44008 #endif
   44009 
   44010 #ifdef __LITTLE_ENDIAN__
   44011 __ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
   44012   uint32x4_t __ret;
   44013   __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
   44014   return __ret;
   44015 }
   44016 #else
   44017 __ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
   44018   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   44019   uint32x4_t __ret;
   44020   __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
   44021   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   44022   return __ret;
   44023 }
   44024 #endif
   44025 
   44026 #ifdef __LITTLE_ENDIAN__
   44027 __ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
   44028   uint64x2_t __ret;
   44029   __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
   44030   return __ret;
   44031 }
   44032 #else
   44033 __ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
   44034   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   44035   uint64x2_t __ret;
   44036   __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
   44037   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   44038   return __ret;
   44039 }
   44040 #endif
   44041 
   44042 #ifdef __LITTLE_ENDIAN__
   44043 __ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
   44044   uint16x8_t __ret;
   44045   __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 49);
   44046   return __ret;
   44047 }
   44048 #else
   44049 __ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
   44050   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   44051   uint16x8_t __ret;
   44052   __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 49);
   44053   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   44054   return __ret;
   44055 }
   44056 #endif
   44057 
   44058 #ifdef __LITTLE_ENDIAN__
   44059 __ai uint8x8_t vcltz_s8(int8x8_t __p0) {
   44060   uint8x8_t __ret;
   44061   __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 16);
   44062   return __ret;
   44063 }
   44064 #else
   44065 __ai uint8x8_t vcltz_s8(int8x8_t __p0) {
   44066   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   44067   uint8x8_t __ret;
   44068   __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 16);
   44069   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   44070   return __ret;
   44071 }
   44072 #endif
   44073 
   44074 #ifdef __LITTLE_ENDIAN__
   44075 __ai uint64x1_t vcltz_f64(float64x1_t __p0) {
   44076   uint64x1_t __ret;
   44077   __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
   44078   return __ret;
   44079 }
   44080 #else
   44081 __ai uint64x1_t vcltz_f64(float64x1_t __p0) {
   44082   uint64x1_t __ret;
   44083   __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
   44084   return __ret;
   44085 }
   44086 #endif
   44087 
   44088 #ifdef __LITTLE_ENDIAN__
   44089 __ai uint32x2_t vcltz_f32(float32x2_t __p0) {
   44090   uint32x2_t __ret;
   44091   __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
   44092   return __ret;
   44093 }
   44094 #else
   44095 __ai uint32x2_t vcltz_f32(float32x2_t __p0) {
   44096   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   44097   uint32x2_t __ret;
   44098   __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
   44099   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   44100   return __ret;
   44101 }
   44102 #endif
   44103 
   44104 #ifdef __LITTLE_ENDIAN__
   44105 __ai uint32x2_t vcltz_s32(int32x2_t __p0) {
   44106   uint32x2_t __ret;
   44107   __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
   44108   return __ret;
   44109 }
   44110 #else
   44111 __ai uint32x2_t vcltz_s32(int32x2_t __p0) {
   44112   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   44113   uint32x2_t __ret;
   44114   __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
   44115   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   44116   return __ret;
   44117 }
   44118 #endif
   44119 
   44120 #ifdef __LITTLE_ENDIAN__
   44121 __ai uint64x1_t vcltz_s64(int64x1_t __p0) {
   44122   uint64x1_t __ret;
   44123   __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
   44124   return __ret;
   44125 }
   44126 #else
   44127 __ai uint64x1_t vcltz_s64(int64x1_t __p0) {
   44128   uint64x1_t __ret;
   44129   __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
   44130   return __ret;
   44131 }
   44132 #endif
   44133 
   44134 #ifdef __LITTLE_ENDIAN__
   44135 __ai uint16x4_t vcltz_s16(int16x4_t __p0) {
   44136   uint16x4_t __ret;
   44137   __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 17);
   44138   return __ret;
   44139 }
   44140 #else
   44141 __ai uint16x4_t vcltz_s16(int16x4_t __p0) {
   44142   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   44143   uint16x4_t __ret;
   44144   __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 17);
   44145   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   44146   return __ret;
   44147 }
   44148 #endif
   44149 
   44150 #ifdef __LITTLE_ENDIAN__
   44151 __ai int64_t vcltzd_s64(int64_t __p0) {
   44152   int64_t __ret;
   44153   __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
   44154   return __ret;
   44155 }
   44156 #else
   44157 __ai int64_t vcltzd_s64(int64_t __p0) {
   44158   int64_t __ret;
   44159   __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
   44160   return __ret;
   44161 }
   44162 #endif
   44163 
   44164 #ifdef __LITTLE_ENDIAN__
   44165 __ai uint64_t vcltzd_f64(float64_t __p0) {
   44166   uint64_t __ret;
   44167   __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
   44168   return __ret;
   44169 }
   44170 #else
   44171 __ai uint64_t vcltzd_f64(float64_t __p0) {
   44172   uint64_t __ret;
   44173   __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
   44174   return __ret;
   44175 }
   44176 #endif
   44177 
   44178 #ifdef __LITTLE_ENDIAN__
   44179 __ai uint32_t vcltzs_f32(float32_t __p0) {
   44180   uint32_t __ret;
   44181   __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
   44182   return __ret;
   44183 }
   44184 #else
   44185 __ai uint32_t vcltzs_f32(float32_t __p0) {
   44186   uint32_t __ret;
   44187   __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
   44188   return __ret;
   44189 }
   44190 #endif
   44191 
   44192 #ifdef __LITTLE_ENDIAN__
   44193 __ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
   44194   poly64x2_t __ret;
   44195   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
   44196   return __ret;
   44197 }
   44198 #else
   44199 __ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
   44200   poly64x2_t __ret;
   44201   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
   44202   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   44203   return __ret;
   44204 }
   44205 #endif
   44206 
   44207 #ifdef __LITTLE_ENDIAN__
   44208 __ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
   44209   float64x2_t __ret;
   44210   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
   44211   return __ret;
   44212 }
   44213 #else
   44214 __ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
   44215   float64x2_t __ret;
   44216   __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
   44217   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   44218   return __ret;
   44219 }
   44220 #endif
   44221 
   44222 #ifdef __LITTLE_ENDIAN__
   44223 #define vcopyq_lane_p8(__p0_0, __p1_0, __p2_0, __p3_0) __extension__ ({ \
   44224   poly8x16_t __s0_0 = __p0_0; \
   44225   poly8x8_t __s2_0 = __p2_0; \
   44226   poly8x16_t __ret_0; \
   44227   __ret_0 = vsetq_lane_p8(vget_lane_p8(__s2_0, __p3_0), __s0_0, __p1_0); \
   44228   __ret_0; \
   44229 })
   44230 #else
   44231 #define vcopyq_lane_p8(__p0_1, __p1_1, __p2_1, __p3_1) __extension__ ({ \
   44232   poly8x16_t __s0_1 = __p0_1; \
   44233   poly8x8_t __s2_1 = __p2_1; \
   44234   poly8x16_t __rev0_1;  __rev0_1 = __builtin_shufflevector(__s0_1, __s0_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44235   poly8x8_t __rev2_1;  __rev2_1 = __builtin_shufflevector(__s2_1, __s2_1, 7, 6, 5, 4, 3, 2, 1, 0); \
   44236   poly8x16_t __ret_1; \
   44237   __ret_1 = __noswap_vsetq_lane_p8(__noswap_vget_lane_p8(__rev2_1, __p3_1), __rev0_1, __p1_1); \
   44238   __ret_1 = __builtin_shufflevector(__ret_1, __ret_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44239   __ret_1; \
   44240 })
   44241 #endif
   44242 
   44243 #ifdef __LITTLE_ENDIAN__
   44244 #define vcopyq_lane_p16(__p0_2, __p1_2, __p2_2, __p3_2) __extension__ ({ \
   44245   poly16x8_t __s0_2 = __p0_2; \
   44246   poly16x4_t __s2_2 = __p2_2; \
   44247   poly16x8_t __ret_2; \
   44248   __ret_2 = vsetq_lane_p16(vget_lane_p16(__s2_2, __p3_2), __s0_2, __p1_2); \
   44249   __ret_2; \
   44250 })
   44251 #else
   44252 #define vcopyq_lane_p16(__p0_3, __p1_3, __p2_3, __p3_3) __extension__ ({ \
   44253   poly16x8_t __s0_3 = __p0_3; \
   44254   poly16x4_t __s2_3 = __p2_3; \
   44255   poly16x8_t __rev0_3;  __rev0_3 = __builtin_shufflevector(__s0_3, __s0_3, 7, 6, 5, 4, 3, 2, 1, 0); \
   44256   poly16x4_t __rev2_3;  __rev2_3 = __builtin_shufflevector(__s2_3, __s2_3, 3, 2, 1, 0); \
   44257   poly16x8_t __ret_3; \
   44258   __ret_3 = __noswap_vsetq_lane_p16(__noswap_vget_lane_p16(__rev2_3, __p3_3), __rev0_3, __p1_3); \
   44259   __ret_3 = __builtin_shufflevector(__ret_3, __ret_3, 7, 6, 5, 4, 3, 2, 1, 0); \
   44260   __ret_3; \
   44261 })
   44262 #endif
   44263 
   44264 #ifdef __LITTLE_ENDIAN__
   44265 #define vcopyq_lane_u8(__p0_4, __p1_4, __p2_4, __p3_4) __extension__ ({ \
   44266   uint8x16_t __s0_4 = __p0_4; \
   44267   uint8x8_t __s2_4 = __p2_4; \
   44268   uint8x16_t __ret_4; \
   44269   __ret_4 = vsetq_lane_u8(vget_lane_u8(__s2_4, __p3_4), __s0_4, __p1_4); \
   44270   __ret_4; \
   44271 })
   44272 #else
   44273 #define vcopyq_lane_u8(__p0_5, __p1_5, __p2_5, __p3_5) __extension__ ({ \
   44274   uint8x16_t __s0_5 = __p0_5; \
   44275   uint8x8_t __s2_5 = __p2_5; \
   44276   uint8x16_t __rev0_5;  __rev0_5 = __builtin_shufflevector(__s0_5, __s0_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44277   uint8x8_t __rev2_5;  __rev2_5 = __builtin_shufflevector(__s2_5, __s2_5, 7, 6, 5, 4, 3, 2, 1, 0); \
   44278   uint8x16_t __ret_5; \
   44279   __ret_5 = __noswap_vsetq_lane_u8(__noswap_vget_lane_u8(__rev2_5, __p3_5), __rev0_5, __p1_5); \
   44280   __ret_5 = __builtin_shufflevector(__ret_5, __ret_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44281   __ret_5; \
   44282 })
   44283 #endif
   44284 
   44285 #ifdef __LITTLE_ENDIAN__
   44286 #define vcopyq_lane_u32(__p0_6, __p1_6, __p2_6, __p3_6) __extension__ ({ \
   44287   uint32x4_t __s0_6 = __p0_6; \
   44288   uint32x2_t __s2_6 = __p2_6; \
   44289   uint32x4_t __ret_6; \
   44290   __ret_6 = vsetq_lane_u32(vget_lane_u32(__s2_6, __p3_6), __s0_6, __p1_6); \
   44291   __ret_6; \
   44292 })
   44293 #else
   44294 #define vcopyq_lane_u32(__p0_7, __p1_7, __p2_7, __p3_7) __extension__ ({ \
   44295   uint32x4_t __s0_7 = __p0_7; \
   44296   uint32x2_t __s2_7 = __p2_7; \
   44297   uint32x4_t __rev0_7;  __rev0_7 = __builtin_shufflevector(__s0_7, __s0_7, 3, 2, 1, 0); \
   44298   uint32x2_t __rev2_7;  __rev2_7 = __builtin_shufflevector(__s2_7, __s2_7, 1, 0); \
   44299   uint32x4_t __ret_7; \
   44300   __ret_7 = __noswap_vsetq_lane_u32(__noswap_vget_lane_u32(__rev2_7, __p3_7), __rev0_7, __p1_7); \
   44301   __ret_7 = __builtin_shufflevector(__ret_7, __ret_7, 3, 2, 1, 0); \
   44302   __ret_7; \
   44303 })
   44304 #endif
   44305 
   44306 #ifdef __LITTLE_ENDIAN__
   44307 #define vcopyq_lane_u64(__p0_8, __p1_8, __p2_8, __p3_8) __extension__ ({ \
   44308   uint64x2_t __s0_8 = __p0_8; \
   44309   uint64x1_t __s2_8 = __p2_8; \
   44310   uint64x2_t __ret_8; \
   44311   __ret_8 = vsetq_lane_u64(vget_lane_u64(__s2_8, __p3_8), __s0_8, __p1_8); \
   44312   __ret_8; \
   44313 })
   44314 #else
   44315 #define vcopyq_lane_u64(__p0_9, __p1_9, __p2_9, __p3_9) __extension__ ({ \
   44316   uint64x2_t __s0_9 = __p0_9; \
   44317   uint64x1_t __s2_9 = __p2_9; \
   44318   uint64x2_t __rev0_9;  __rev0_9 = __builtin_shufflevector(__s0_9, __s0_9, 1, 0); \
   44319   uint64x2_t __ret_9; \
   44320   __ret_9 = __noswap_vsetq_lane_u64(__noswap_vget_lane_u64(__s2_9, __p3_9), __rev0_9, __p1_9); \
   44321   __ret_9 = __builtin_shufflevector(__ret_9, __ret_9, 1, 0); \
   44322   __ret_9; \
   44323 })
   44324 #endif
   44325 
   44326 #ifdef __LITTLE_ENDIAN__
   44327 #define vcopyq_lane_u16(__p0_10, __p1_10, __p2_10, __p3_10) __extension__ ({ \
   44328   uint16x8_t __s0_10 = __p0_10; \
   44329   uint16x4_t __s2_10 = __p2_10; \
   44330   uint16x8_t __ret_10; \
   44331   __ret_10 = vsetq_lane_u16(vget_lane_u16(__s2_10, __p3_10), __s0_10, __p1_10); \
   44332   __ret_10; \
   44333 })
   44334 #else
   44335 #define vcopyq_lane_u16(__p0_11, __p1_11, __p2_11, __p3_11) __extension__ ({ \
   44336   uint16x8_t __s0_11 = __p0_11; \
   44337   uint16x4_t __s2_11 = __p2_11; \
   44338   uint16x8_t __rev0_11;  __rev0_11 = __builtin_shufflevector(__s0_11, __s0_11, 7, 6, 5, 4, 3, 2, 1, 0); \
   44339   uint16x4_t __rev2_11;  __rev2_11 = __builtin_shufflevector(__s2_11, __s2_11, 3, 2, 1, 0); \
   44340   uint16x8_t __ret_11; \
   44341   __ret_11 = __noswap_vsetq_lane_u16(__noswap_vget_lane_u16(__rev2_11, __p3_11), __rev0_11, __p1_11); \
   44342   __ret_11 = __builtin_shufflevector(__ret_11, __ret_11, 7, 6, 5, 4, 3, 2, 1, 0); \
   44343   __ret_11; \
   44344 })
   44345 #endif
   44346 
   44347 #ifdef __LITTLE_ENDIAN__
   44348 #define vcopyq_lane_s8(__p0_12, __p1_12, __p2_12, __p3_12) __extension__ ({ \
   44349   int8x16_t __s0_12 = __p0_12; \
   44350   int8x8_t __s2_12 = __p2_12; \
   44351   int8x16_t __ret_12; \
   44352   __ret_12 = vsetq_lane_s8(vget_lane_s8(__s2_12, __p3_12), __s0_12, __p1_12); \
   44353   __ret_12; \
   44354 })
   44355 #else
   44356 #define vcopyq_lane_s8(__p0_13, __p1_13, __p2_13, __p3_13) __extension__ ({ \
   44357   int8x16_t __s0_13 = __p0_13; \
   44358   int8x8_t __s2_13 = __p2_13; \
   44359   int8x16_t __rev0_13;  __rev0_13 = __builtin_shufflevector(__s0_13, __s0_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44360   int8x8_t __rev2_13;  __rev2_13 = __builtin_shufflevector(__s2_13, __s2_13, 7, 6, 5, 4, 3, 2, 1, 0); \
   44361   int8x16_t __ret_13; \
   44362   __ret_13 = __noswap_vsetq_lane_s8(__noswap_vget_lane_s8(__rev2_13, __p3_13), __rev0_13, __p1_13); \
   44363   __ret_13 = __builtin_shufflevector(__ret_13, __ret_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44364   __ret_13; \
   44365 })
   44366 #endif
   44367 
   44368 #ifdef __LITTLE_ENDIAN__
   44369 #define vcopyq_lane_f32(__p0_14, __p1_14, __p2_14, __p3_14) __extension__ ({ \
   44370   float32x4_t __s0_14 = __p0_14; \
   44371   float32x2_t __s2_14 = __p2_14; \
   44372   float32x4_t __ret_14; \
   44373   __ret_14 = vsetq_lane_f32(vget_lane_f32(__s2_14, __p3_14), __s0_14, __p1_14); \
   44374   __ret_14; \
   44375 })
   44376 #else
   44377 #define vcopyq_lane_f32(__p0_15, __p1_15, __p2_15, __p3_15) __extension__ ({ \
   44378   float32x4_t __s0_15 = __p0_15; \
   44379   float32x2_t __s2_15 = __p2_15; \
   44380   float32x4_t __rev0_15;  __rev0_15 = __builtin_shufflevector(__s0_15, __s0_15, 3, 2, 1, 0); \
   44381   float32x2_t __rev2_15;  __rev2_15 = __builtin_shufflevector(__s2_15, __s2_15, 1, 0); \
   44382   float32x4_t __ret_15; \
   44383   __ret_15 = __noswap_vsetq_lane_f32(__noswap_vget_lane_f32(__rev2_15, __p3_15), __rev0_15, __p1_15); \
   44384   __ret_15 = __builtin_shufflevector(__ret_15, __ret_15, 3, 2, 1, 0); \
   44385   __ret_15; \
   44386 })
   44387 #endif
   44388 
   44389 #ifdef __LITTLE_ENDIAN__
   44390 #define vcopyq_lane_s32(__p0_16, __p1_16, __p2_16, __p3_16) __extension__ ({ \
   44391   int32x4_t __s0_16 = __p0_16; \
   44392   int32x2_t __s2_16 = __p2_16; \
   44393   int32x4_t __ret_16; \
   44394   __ret_16 = vsetq_lane_s32(vget_lane_s32(__s2_16, __p3_16), __s0_16, __p1_16); \
   44395   __ret_16; \
   44396 })
   44397 #else
   44398 #define vcopyq_lane_s32(__p0_17, __p1_17, __p2_17, __p3_17) __extension__ ({ \
   44399   int32x4_t __s0_17 = __p0_17; \
   44400   int32x2_t __s2_17 = __p2_17; \
   44401   int32x4_t __rev0_17;  __rev0_17 = __builtin_shufflevector(__s0_17, __s0_17, 3, 2, 1, 0); \
   44402   int32x2_t __rev2_17;  __rev2_17 = __builtin_shufflevector(__s2_17, __s2_17, 1, 0); \
   44403   int32x4_t __ret_17; \
   44404   __ret_17 = __noswap_vsetq_lane_s32(__noswap_vget_lane_s32(__rev2_17, __p3_17), __rev0_17, __p1_17); \
   44405   __ret_17 = __builtin_shufflevector(__ret_17, __ret_17, 3, 2, 1, 0); \
   44406   __ret_17; \
   44407 })
   44408 #endif
   44409 
   44410 #ifdef __LITTLE_ENDIAN__
   44411 #define vcopyq_lane_s64(__p0_18, __p1_18, __p2_18, __p3_18) __extension__ ({ \
   44412   int64x2_t __s0_18 = __p0_18; \
   44413   int64x1_t __s2_18 = __p2_18; \
   44414   int64x2_t __ret_18; \
   44415   __ret_18 = vsetq_lane_s64(vget_lane_s64(__s2_18, __p3_18), __s0_18, __p1_18); \
   44416   __ret_18; \
   44417 })
   44418 #else
   44419 #define vcopyq_lane_s64(__p0_19, __p1_19, __p2_19, __p3_19) __extension__ ({ \
   44420   int64x2_t __s0_19 = __p0_19; \
   44421   int64x1_t __s2_19 = __p2_19; \
   44422   int64x2_t __rev0_19;  __rev0_19 = __builtin_shufflevector(__s0_19, __s0_19, 1, 0); \
   44423   int64x2_t __ret_19; \
   44424   __ret_19 = __noswap_vsetq_lane_s64(__noswap_vget_lane_s64(__s2_19, __p3_19), __rev0_19, __p1_19); \
   44425   __ret_19 = __builtin_shufflevector(__ret_19, __ret_19, 1, 0); \
   44426   __ret_19; \
   44427 })
   44428 #endif
   44429 
   44430 #ifdef __LITTLE_ENDIAN__
   44431 #define vcopyq_lane_s16(__p0_20, __p1_20, __p2_20, __p3_20) __extension__ ({ \
   44432   int16x8_t __s0_20 = __p0_20; \
   44433   int16x4_t __s2_20 = __p2_20; \
   44434   int16x8_t __ret_20; \
   44435   __ret_20 = vsetq_lane_s16(vget_lane_s16(__s2_20, __p3_20), __s0_20, __p1_20); \
   44436   __ret_20; \
   44437 })
   44438 #else
   44439 #define vcopyq_lane_s16(__p0_21, __p1_21, __p2_21, __p3_21) __extension__ ({ \
   44440   int16x8_t __s0_21 = __p0_21; \
   44441   int16x4_t __s2_21 = __p2_21; \
   44442   int16x8_t __rev0_21;  __rev0_21 = __builtin_shufflevector(__s0_21, __s0_21, 7, 6, 5, 4, 3, 2, 1, 0); \
   44443   int16x4_t __rev2_21;  __rev2_21 = __builtin_shufflevector(__s2_21, __s2_21, 3, 2, 1, 0); \
   44444   int16x8_t __ret_21; \
   44445   __ret_21 = __noswap_vsetq_lane_s16(__noswap_vget_lane_s16(__rev2_21, __p3_21), __rev0_21, __p1_21); \
   44446   __ret_21 = __builtin_shufflevector(__ret_21, __ret_21, 7, 6, 5, 4, 3, 2, 1, 0); \
   44447   __ret_21; \
   44448 })
   44449 #endif
   44450 
   44451 #ifdef __LITTLE_ENDIAN__
   44452 #define vcopy_lane_p8(__p0_22, __p1_22, __p2_22, __p3_22) __extension__ ({ \
   44453   poly8x8_t __s0_22 = __p0_22; \
   44454   poly8x8_t __s2_22 = __p2_22; \
   44455   poly8x8_t __ret_22; \
   44456   __ret_22 = vset_lane_p8(vget_lane_p8(__s2_22, __p3_22), __s0_22, __p1_22); \
   44457   __ret_22; \
   44458 })
   44459 #else
   44460 #define vcopy_lane_p8(__p0_23, __p1_23, __p2_23, __p3_23) __extension__ ({ \
   44461   poly8x8_t __s0_23 = __p0_23; \
   44462   poly8x8_t __s2_23 = __p2_23; \
   44463   poly8x8_t __rev0_23;  __rev0_23 = __builtin_shufflevector(__s0_23, __s0_23, 7, 6, 5, 4, 3, 2, 1, 0); \
   44464   poly8x8_t __rev2_23;  __rev2_23 = __builtin_shufflevector(__s2_23, __s2_23, 7, 6, 5, 4, 3, 2, 1, 0); \
   44465   poly8x8_t __ret_23; \
   44466   __ret_23 = __noswap_vset_lane_p8(__noswap_vget_lane_p8(__rev2_23, __p3_23), __rev0_23, __p1_23); \
   44467   __ret_23 = __builtin_shufflevector(__ret_23, __ret_23, 7, 6, 5, 4, 3, 2, 1, 0); \
   44468   __ret_23; \
   44469 })
   44470 #endif
   44471 
   44472 #ifdef __LITTLE_ENDIAN__
   44473 #define vcopy_lane_p16(__p0_24, __p1_24, __p2_24, __p3_24) __extension__ ({ \
   44474   poly16x4_t __s0_24 = __p0_24; \
   44475   poly16x4_t __s2_24 = __p2_24; \
   44476   poly16x4_t __ret_24; \
   44477   __ret_24 = vset_lane_p16(vget_lane_p16(__s2_24, __p3_24), __s0_24, __p1_24); \
   44478   __ret_24; \
   44479 })
   44480 #else
   44481 #define vcopy_lane_p16(__p0_25, __p1_25, __p2_25, __p3_25) __extension__ ({ \
   44482   poly16x4_t __s0_25 = __p0_25; \
   44483   poly16x4_t __s2_25 = __p2_25; \
   44484   poly16x4_t __rev0_25;  __rev0_25 = __builtin_shufflevector(__s0_25, __s0_25, 3, 2, 1, 0); \
   44485   poly16x4_t __rev2_25;  __rev2_25 = __builtin_shufflevector(__s2_25, __s2_25, 3, 2, 1, 0); \
   44486   poly16x4_t __ret_25; \
   44487   __ret_25 = __noswap_vset_lane_p16(__noswap_vget_lane_p16(__rev2_25, __p3_25), __rev0_25, __p1_25); \
   44488   __ret_25 = __builtin_shufflevector(__ret_25, __ret_25, 3, 2, 1, 0); \
   44489   __ret_25; \
   44490 })
   44491 #endif
   44492 
   44493 #ifdef __LITTLE_ENDIAN__
   44494 #define vcopy_lane_u8(__p0_26, __p1_26, __p2_26, __p3_26) __extension__ ({ \
   44495   uint8x8_t __s0_26 = __p0_26; \
   44496   uint8x8_t __s2_26 = __p2_26; \
   44497   uint8x8_t __ret_26; \
   44498   __ret_26 = vset_lane_u8(vget_lane_u8(__s2_26, __p3_26), __s0_26, __p1_26); \
   44499   __ret_26; \
   44500 })
   44501 #else
   44502 #define vcopy_lane_u8(__p0_27, __p1_27, __p2_27, __p3_27) __extension__ ({ \
   44503   uint8x8_t __s0_27 = __p0_27; \
   44504   uint8x8_t __s2_27 = __p2_27; \
   44505   uint8x8_t __rev0_27;  __rev0_27 = __builtin_shufflevector(__s0_27, __s0_27, 7, 6, 5, 4, 3, 2, 1, 0); \
   44506   uint8x8_t __rev2_27;  __rev2_27 = __builtin_shufflevector(__s2_27, __s2_27, 7, 6, 5, 4, 3, 2, 1, 0); \
   44507   uint8x8_t __ret_27; \
   44508   __ret_27 = __noswap_vset_lane_u8(__noswap_vget_lane_u8(__rev2_27, __p3_27), __rev0_27, __p1_27); \
   44509   __ret_27 = __builtin_shufflevector(__ret_27, __ret_27, 7, 6, 5, 4, 3, 2, 1, 0); \
   44510   __ret_27; \
   44511 })
   44512 #endif
   44513 
   44514 #ifdef __LITTLE_ENDIAN__
   44515 #define vcopy_lane_u32(__p0_28, __p1_28, __p2_28, __p3_28) __extension__ ({ \
   44516   uint32x2_t __s0_28 = __p0_28; \
   44517   uint32x2_t __s2_28 = __p2_28; \
   44518   uint32x2_t __ret_28; \
   44519   __ret_28 = vset_lane_u32(vget_lane_u32(__s2_28, __p3_28), __s0_28, __p1_28); \
   44520   __ret_28; \
   44521 })
   44522 #else
   44523 #define vcopy_lane_u32(__p0_29, __p1_29, __p2_29, __p3_29) __extension__ ({ \
   44524   uint32x2_t __s0_29 = __p0_29; \
   44525   uint32x2_t __s2_29 = __p2_29; \
   44526   uint32x2_t __rev0_29;  __rev0_29 = __builtin_shufflevector(__s0_29, __s0_29, 1, 0); \
   44527   uint32x2_t __rev2_29;  __rev2_29 = __builtin_shufflevector(__s2_29, __s2_29, 1, 0); \
   44528   uint32x2_t __ret_29; \
   44529   __ret_29 = __noswap_vset_lane_u32(__noswap_vget_lane_u32(__rev2_29, __p3_29), __rev0_29, __p1_29); \
   44530   __ret_29 = __builtin_shufflevector(__ret_29, __ret_29, 1, 0); \
   44531   __ret_29; \
   44532 })
   44533 #endif
   44534 
   44535 #ifdef __LITTLE_ENDIAN__
   44536 #define vcopy_lane_u64(__p0_30, __p1_30, __p2_30, __p3_30) __extension__ ({ \
   44537   uint64x1_t __s0_30 = __p0_30; \
   44538   uint64x1_t __s2_30 = __p2_30; \
   44539   uint64x1_t __ret_30; \
   44540   __ret_30 = vset_lane_u64(vget_lane_u64(__s2_30, __p3_30), __s0_30, __p1_30); \
   44541   __ret_30; \
   44542 })
   44543 #else
   44544 #define vcopy_lane_u64(__p0_31, __p1_31, __p2_31, __p3_31) __extension__ ({ \
   44545   uint64x1_t __s0_31 = __p0_31; \
   44546   uint64x1_t __s2_31 = __p2_31; \
   44547   uint64x1_t __ret_31; \
   44548   __ret_31 = __noswap_vset_lane_u64(__noswap_vget_lane_u64(__s2_31, __p3_31), __s0_31, __p1_31); \
   44549   __ret_31; \
   44550 })
   44551 #endif
   44552 
   44553 #ifdef __LITTLE_ENDIAN__
   44554 #define vcopy_lane_u16(__p0_32, __p1_32, __p2_32, __p3_32) __extension__ ({ \
   44555   uint16x4_t __s0_32 = __p0_32; \
   44556   uint16x4_t __s2_32 = __p2_32; \
   44557   uint16x4_t __ret_32; \
   44558   __ret_32 = vset_lane_u16(vget_lane_u16(__s2_32, __p3_32), __s0_32, __p1_32); \
   44559   __ret_32; \
   44560 })
   44561 #else
   44562 #define vcopy_lane_u16(__p0_33, __p1_33, __p2_33, __p3_33) __extension__ ({ \
   44563   uint16x4_t __s0_33 = __p0_33; \
   44564   uint16x4_t __s2_33 = __p2_33; \
   44565   uint16x4_t __rev0_33;  __rev0_33 = __builtin_shufflevector(__s0_33, __s0_33, 3, 2, 1, 0); \
   44566   uint16x4_t __rev2_33;  __rev2_33 = __builtin_shufflevector(__s2_33, __s2_33, 3, 2, 1, 0); \
   44567   uint16x4_t __ret_33; \
   44568   __ret_33 = __noswap_vset_lane_u16(__noswap_vget_lane_u16(__rev2_33, __p3_33), __rev0_33, __p1_33); \
   44569   __ret_33 = __builtin_shufflevector(__ret_33, __ret_33, 3, 2, 1, 0); \
   44570   __ret_33; \
   44571 })
   44572 #endif
   44573 
   44574 #ifdef __LITTLE_ENDIAN__
   44575 #define vcopy_lane_s8(__p0_34, __p1_34, __p2_34, __p3_34) __extension__ ({ \
   44576   int8x8_t __s0_34 = __p0_34; \
   44577   int8x8_t __s2_34 = __p2_34; \
   44578   int8x8_t __ret_34; \
   44579   __ret_34 = vset_lane_s8(vget_lane_s8(__s2_34, __p3_34), __s0_34, __p1_34); \
   44580   __ret_34; \
   44581 })
   44582 #else
   44583 #define vcopy_lane_s8(__p0_35, __p1_35, __p2_35, __p3_35) __extension__ ({ \
   44584   int8x8_t __s0_35 = __p0_35; \
   44585   int8x8_t __s2_35 = __p2_35; \
   44586   int8x8_t __rev0_35;  __rev0_35 = __builtin_shufflevector(__s0_35, __s0_35, 7, 6, 5, 4, 3, 2, 1, 0); \
   44587   int8x8_t __rev2_35;  __rev2_35 = __builtin_shufflevector(__s2_35, __s2_35, 7, 6, 5, 4, 3, 2, 1, 0); \
   44588   int8x8_t __ret_35; \
   44589   __ret_35 = __noswap_vset_lane_s8(__noswap_vget_lane_s8(__rev2_35, __p3_35), __rev0_35, __p1_35); \
   44590   __ret_35 = __builtin_shufflevector(__ret_35, __ret_35, 7, 6, 5, 4, 3, 2, 1, 0); \
   44591   __ret_35; \
   44592 })
   44593 #endif
   44594 
   44595 #ifdef __LITTLE_ENDIAN__
   44596 #define vcopy_lane_f32(__p0_36, __p1_36, __p2_36, __p3_36) __extension__ ({ \
   44597   float32x2_t __s0_36 = __p0_36; \
   44598   float32x2_t __s2_36 = __p2_36; \
   44599   float32x2_t __ret_36; \
   44600   __ret_36 = vset_lane_f32(vget_lane_f32(__s2_36, __p3_36), __s0_36, __p1_36); \
   44601   __ret_36; \
   44602 })
   44603 #else
   44604 #define vcopy_lane_f32(__p0_37, __p1_37, __p2_37, __p3_37) __extension__ ({ \
   44605   float32x2_t __s0_37 = __p0_37; \
   44606   float32x2_t __s2_37 = __p2_37; \
   44607   float32x2_t __rev0_37;  __rev0_37 = __builtin_shufflevector(__s0_37, __s0_37, 1, 0); \
   44608   float32x2_t __rev2_37;  __rev2_37 = __builtin_shufflevector(__s2_37, __s2_37, 1, 0); \
   44609   float32x2_t __ret_37; \
   44610   __ret_37 = __noswap_vset_lane_f32(__noswap_vget_lane_f32(__rev2_37, __p3_37), __rev0_37, __p1_37); \
   44611   __ret_37 = __builtin_shufflevector(__ret_37, __ret_37, 1, 0); \
   44612   __ret_37; \
   44613 })
   44614 #endif
   44615 
   44616 #ifdef __LITTLE_ENDIAN__
   44617 #define vcopy_lane_s32(__p0_38, __p1_38, __p2_38, __p3_38) __extension__ ({ \
   44618   int32x2_t __s0_38 = __p0_38; \
   44619   int32x2_t __s2_38 = __p2_38; \
   44620   int32x2_t __ret_38; \
   44621   __ret_38 = vset_lane_s32(vget_lane_s32(__s2_38, __p3_38), __s0_38, __p1_38); \
   44622   __ret_38; \
   44623 })
   44624 #else
   44625 #define vcopy_lane_s32(__p0_39, __p1_39, __p2_39, __p3_39) __extension__ ({ \
   44626   int32x2_t __s0_39 = __p0_39; \
   44627   int32x2_t __s2_39 = __p2_39; \
   44628   int32x2_t __rev0_39;  __rev0_39 = __builtin_shufflevector(__s0_39, __s0_39, 1, 0); \
   44629   int32x2_t __rev2_39;  __rev2_39 = __builtin_shufflevector(__s2_39, __s2_39, 1, 0); \
   44630   int32x2_t __ret_39; \
   44631   __ret_39 = __noswap_vset_lane_s32(__noswap_vget_lane_s32(__rev2_39, __p3_39), __rev0_39, __p1_39); \
   44632   __ret_39 = __builtin_shufflevector(__ret_39, __ret_39, 1, 0); \
   44633   __ret_39; \
   44634 })
   44635 #endif
   44636 
   44637 #ifdef __LITTLE_ENDIAN__
   44638 #define vcopy_lane_s64(__p0_40, __p1_40, __p2_40, __p3_40) __extension__ ({ \
   44639   int64x1_t __s0_40 = __p0_40; \
   44640   int64x1_t __s2_40 = __p2_40; \
   44641   int64x1_t __ret_40; \
   44642   __ret_40 = vset_lane_s64(vget_lane_s64(__s2_40, __p3_40), __s0_40, __p1_40); \
   44643   __ret_40; \
   44644 })
   44645 #else
   44646 #define vcopy_lane_s64(__p0_41, __p1_41, __p2_41, __p3_41) __extension__ ({ \
   44647   int64x1_t __s0_41 = __p0_41; \
   44648   int64x1_t __s2_41 = __p2_41; \
   44649   int64x1_t __ret_41; \
   44650   __ret_41 = __noswap_vset_lane_s64(__noswap_vget_lane_s64(__s2_41, __p3_41), __s0_41, __p1_41); \
   44651   __ret_41; \
   44652 })
   44653 #endif
   44654 
   44655 #ifdef __LITTLE_ENDIAN__
   44656 #define vcopy_lane_s16(__p0_42, __p1_42, __p2_42, __p3_42) __extension__ ({ \
   44657   int16x4_t __s0_42 = __p0_42; \
   44658   int16x4_t __s2_42 = __p2_42; \
   44659   int16x4_t __ret_42; \
   44660   __ret_42 = vset_lane_s16(vget_lane_s16(__s2_42, __p3_42), __s0_42, __p1_42); \
   44661   __ret_42; \
   44662 })
   44663 #else
   44664 #define vcopy_lane_s16(__p0_43, __p1_43, __p2_43, __p3_43) __extension__ ({ \
   44665   int16x4_t __s0_43 = __p0_43; \
   44666   int16x4_t __s2_43 = __p2_43; \
   44667   int16x4_t __rev0_43;  __rev0_43 = __builtin_shufflevector(__s0_43, __s0_43, 3, 2, 1, 0); \
   44668   int16x4_t __rev2_43;  __rev2_43 = __builtin_shufflevector(__s2_43, __s2_43, 3, 2, 1, 0); \
   44669   int16x4_t __ret_43; \
   44670   __ret_43 = __noswap_vset_lane_s16(__noswap_vget_lane_s16(__rev2_43, __p3_43), __rev0_43, __p1_43); \
   44671   __ret_43 = __builtin_shufflevector(__ret_43, __ret_43, 3, 2, 1, 0); \
   44672   __ret_43; \
   44673 })
   44674 #endif
   44675 
   44676 #ifdef __LITTLE_ENDIAN__
   44677 #define vcopyq_laneq_p8(__p0_44, __p1_44, __p2_44, __p3_44) __extension__ ({ \
   44678   poly8x16_t __s0_44 = __p0_44; \
   44679   poly8x16_t __s2_44 = __p2_44; \
   44680   poly8x16_t __ret_44; \
   44681   __ret_44 = vsetq_lane_p8(vgetq_lane_p8(__s2_44, __p3_44), __s0_44, __p1_44); \
   44682   __ret_44; \
   44683 })
   44684 #else
   44685 #define vcopyq_laneq_p8(__p0_45, __p1_45, __p2_45, __p3_45) __extension__ ({ \
   44686   poly8x16_t __s0_45 = __p0_45; \
   44687   poly8x16_t __s2_45 = __p2_45; \
   44688   poly8x16_t __rev0_45;  __rev0_45 = __builtin_shufflevector(__s0_45, __s0_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44689   poly8x16_t __rev2_45;  __rev2_45 = __builtin_shufflevector(__s2_45, __s2_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44690   poly8x16_t __ret_45; \
   44691   __ret_45 = __noswap_vsetq_lane_p8(__noswap_vgetq_lane_p8(__rev2_45, __p3_45), __rev0_45, __p1_45); \
   44692   __ret_45 = __builtin_shufflevector(__ret_45, __ret_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44693   __ret_45; \
   44694 })
   44695 #endif
   44696 
   44697 #ifdef __LITTLE_ENDIAN__
   44698 #define vcopyq_laneq_p16(__p0_46, __p1_46, __p2_46, __p3_46) __extension__ ({ \
   44699   poly16x8_t __s0_46 = __p0_46; \
   44700   poly16x8_t __s2_46 = __p2_46; \
   44701   poly16x8_t __ret_46; \
   44702   __ret_46 = vsetq_lane_p16(vgetq_lane_p16(__s2_46, __p3_46), __s0_46, __p1_46); \
   44703   __ret_46; \
   44704 })
   44705 #else
   44706 #define vcopyq_laneq_p16(__p0_47, __p1_47, __p2_47, __p3_47) __extension__ ({ \
   44707   poly16x8_t __s0_47 = __p0_47; \
   44708   poly16x8_t __s2_47 = __p2_47; \
   44709   poly16x8_t __rev0_47;  __rev0_47 = __builtin_shufflevector(__s0_47, __s0_47, 7, 6, 5, 4, 3, 2, 1, 0); \
   44710   poly16x8_t __rev2_47;  __rev2_47 = __builtin_shufflevector(__s2_47, __s2_47, 7, 6, 5, 4, 3, 2, 1, 0); \
   44711   poly16x8_t __ret_47; \
   44712   __ret_47 = __noswap_vsetq_lane_p16(__noswap_vgetq_lane_p16(__rev2_47, __p3_47), __rev0_47, __p1_47); \
   44713   __ret_47 = __builtin_shufflevector(__ret_47, __ret_47, 7, 6, 5, 4, 3, 2, 1, 0); \
   44714   __ret_47; \
   44715 })
   44716 #endif
   44717 
   44718 #ifdef __LITTLE_ENDIAN__
   44719 #define vcopyq_laneq_u8(__p0_48, __p1_48, __p2_48, __p3_48) __extension__ ({ \
   44720   uint8x16_t __s0_48 = __p0_48; \
   44721   uint8x16_t __s2_48 = __p2_48; \
   44722   uint8x16_t __ret_48; \
   44723   __ret_48 = vsetq_lane_u8(vgetq_lane_u8(__s2_48, __p3_48), __s0_48, __p1_48); \
   44724   __ret_48; \
   44725 })
   44726 #else
   44727 #define vcopyq_laneq_u8(__p0_49, __p1_49, __p2_49, __p3_49) __extension__ ({ \
   44728   uint8x16_t __s0_49 = __p0_49; \
   44729   uint8x16_t __s2_49 = __p2_49; \
   44730   uint8x16_t __rev0_49;  __rev0_49 = __builtin_shufflevector(__s0_49, __s0_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44731   uint8x16_t __rev2_49;  __rev2_49 = __builtin_shufflevector(__s2_49, __s2_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44732   uint8x16_t __ret_49; \
   44733   __ret_49 = __noswap_vsetq_lane_u8(__noswap_vgetq_lane_u8(__rev2_49, __p3_49), __rev0_49, __p1_49); \
   44734   __ret_49 = __builtin_shufflevector(__ret_49, __ret_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44735   __ret_49; \
   44736 })
   44737 #endif
   44738 
   44739 #ifdef __LITTLE_ENDIAN__
   44740 #define vcopyq_laneq_u32(__p0_50, __p1_50, __p2_50, __p3_50) __extension__ ({ \
   44741   uint32x4_t __s0_50 = __p0_50; \
   44742   uint32x4_t __s2_50 = __p2_50; \
   44743   uint32x4_t __ret_50; \
   44744   __ret_50 = vsetq_lane_u32(vgetq_lane_u32(__s2_50, __p3_50), __s0_50, __p1_50); \
   44745   __ret_50; \
   44746 })
   44747 #else
   44748 #define vcopyq_laneq_u32(__p0_51, __p1_51, __p2_51, __p3_51) __extension__ ({ \
   44749   uint32x4_t __s0_51 = __p0_51; \
   44750   uint32x4_t __s2_51 = __p2_51; \
   44751   uint32x4_t __rev0_51;  __rev0_51 = __builtin_shufflevector(__s0_51, __s0_51, 3, 2, 1, 0); \
   44752   uint32x4_t __rev2_51;  __rev2_51 = __builtin_shufflevector(__s2_51, __s2_51, 3, 2, 1, 0); \
   44753   uint32x4_t __ret_51; \
   44754   __ret_51 = __noswap_vsetq_lane_u32(__noswap_vgetq_lane_u32(__rev2_51, __p3_51), __rev0_51, __p1_51); \
   44755   __ret_51 = __builtin_shufflevector(__ret_51, __ret_51, 3, 2, 1, 0); \
   44756   __ret_51; \
   44757 })
   44758 #endif
   44759 
   44760 #ifdef __LITTLE_ENDIAN__
   44761 #define vcopyq_laneq_u64(__p0_52, __p1_52, __p2_52, __p3_52) __extension__ ({ \
   44762   uint64x2_t __s0_52 = __p0_52; \
   44763   uint64x2_t __s2_52 = __p2_52; \
   44764   uint64x2_t __ret_52; \
   44765   __ret_52 = vsetq_lane_u64(vgetq_lane_u64(__s2_52, __p3_52), __s0_52, __p1_52); \
   44766   __ret_52; \
   44767 })
   44768 #else
   44769 #define vcopyq_laneq_u64(__p0_53, __p1_53, __p2_53, __p3_53) __extension__ ({ \
   44770   uint64x2_t __s0_53 = __p0_53; \
   44771   uint64x2_t __s2_53 = __p2_53; \
   44772   uint64x2_t __rev0_53;  __rev0_53 = __builtin_shufflevector(__s0_53, __s0_53, 1, 0); \
   44773   uint64x2_t __rev2_53;  __rev2_53 = __builtin_shufflevector(__s2_53, __s2_53, 1, 0); \
   44774   uint64x2_t __ret_53; \
   44775   __ret_53 = __noswap_vsetq_lane_u64(__noswap_vgetq_lane_u64(__rev2_53, __p3_53), __rev0_53, __p1_53); \
   44776   __ret_53 = __builtin_shufflevector(__ret_53, __ret_53, 1, 0); \
   44777   __ret_53; \
   44778 })
   44779 #endif
   44780 
   44781 #ifdef __LITTLE_ENDIAN__
   44782 #define vcopyq_laneq_u16(__p0_54, __p1_54, __p2_54, __p3_54) __extension__ ({ \
   44783   uint16x8_t __s0_54 = __p0_54; \
   44784   uint16x8_t __s2_54 = __p2_54; \
   44785   uint16x8_t __ret_54; \
   44786   __ret_54 = vsetq_lane_u16(vgetq_lane_u16(__s2_54, __p3_54), __s0_54, __p1_54); \
   44787   __ret_54; \
   44788 })
   44789 #else
   44790 #define vcopyq_laneq_u16(__p0_55, __p1_55, __p2_55, __p3_55) __extension__ ({ \
   44791   uint16x8_t __s0_55 = __p0_55; \
   44792   uint16x8_t __s2_55 = __p2_55; \
   44793   uint16x8_t __rev0_55;  __rev0_55 = __builtin_shufflevector(__s0_55, __s0_55, 7, 6, 5, 4, 3, 2, 1, 0); \
   44794   uint16x8_t __rev2_55;  __rev2_55 = __builtin_shufflevector(__s2_55, __s2_55, 7, 6, 5, 4, 3, 2, 1, 0); \
   44795   uint16x8_t __ret_55; \
   44796   __ret_55 = __noswap_vsetq_lane_u16(__noswap_vgetq_lane_u16(__rev2_55, __p3_55), __rev0_55, __p1_55); \
   44797   __ret_55 = __builtin_shufflevector(__ret_55, __ret_55, 7, 6, 5, 4, 3, 2, 1, 0); \
   44798   __ret_55; \
   44799 })
   44800 #endif
   44801 
   44802 #ifdef __LITTLE_ENDIAN__
   44803 #define vcopyq_laneq_s8(__p0_56, __p1_56, __p2_56, __p3_56) __extension__ ({ \
   44804   int8x16_t __s0_56 = __p0_56; \
   44805   int8x16_t __s2_56 = __p2_56; \
   44806   int8x16_t __ret_56; \
   44807   __ret_56 = vsetq_lane_s8(vgetq_lane_s8(__s2_56, __p3_56), __s0_56, __p1_56); \
   44808   __ret_56; \
   44809 })
   44810 #else
   44811 #define vcopyq_laneq_s8(__p0_57, __p1_57, __p2_57, __p3_57) __extension__ ({ \
   44812   int8x16_t __s0_57 = __p0_57; \
   44813   int8x16_t __s2_57 = __p2_57; \
   44814   int8x16_t __rev0_57;  __rev0_57 = __builtin_shufflevector(__s0_57, __s0_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44815   int8x16_t __rev2_57;  __rev2_57 = __builtin_shufflevector(__s2_57, __s2_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44816   int8x16_t __ret_57; \
   44817   __ret_57 = __noswap_vsetq_lane_s8(__noswap_vgetq_lane_s8(__rev2_57, __p3_57), __rev0_57, __p1_57); \
   44818   __ret_57 = __builtin_shufflevector(__ret_57, __ret_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44819   __ret_57; \
   44820 })
   44821 #endif
   44822 
   44823 #ifdef __LITTLE_ENDIAN__
   44824 #define vcopyq_laneq_f32(__p0_58, __p1_58, __p2_58, __p3_58) __extension__ ({ \
   44825   float32x4_t __s0_58 = __p0_58; \
   44826   float32x4_t __s2_58 = __p2_58; \
   44827   float32x4_t __ret_58; \
   44828   __ret_58 = vsetq_lane_f32(vgetq_lane_f32(__s2_58, __p3_58), __s0_58, __p1_58); \
   44829   __ret_58; \
   44830 })
   44831 #else
   44832 #define vcopyq_laneq_f32(__p0_59, __p1_59, __p2_59, __p3_59) __extension__ ({ \
   44833   float32x4_t __s0_59 = __p0_59; \
   44834   float32x4_t __s2_59 = __p2_59; \
   44835   float32x4_t __rev0_59;  __rev0_59 = __builtin_shufflevector(__s0_59, __s0_59, 3, 2, 1, 0); \
   44836   float32x4_t __rev2_59;  __rev2_59 = __builtin_shufflevector(__s2_59, __s2_59, 3, 2, 1, 0); \
   44837   float32x4_t __ret_59; \
   44838   __ret_59 = __noswap_vsetq_lane_f32(__noswap_vgetq_lane_f32(__rev2_59, __p3_59), __rev0_59, __p1_59); \
   44839   __ret_59 = __builtin_shufflevector(__ret_59, __ret_59, 3, 2, 1, 0); \
   44840   __ret_59; \
   44841 })
   44842 #endif
   44843 
   44844 #ifdef __LITTLE_ENDIAN__
   44845 #define vcopyq_laneq_s32(__p0_60, __p1_60, __p2_60, __p3_60) __extension__ ({ \
   44846   int32x4_t __s0_60 = __p0_60; \
   44847   int32x4_t __s2_60 = __p2_60; \
   44848   int32x4_t __ret_60; \
   44849   __ret_60 = vsetq_lane_s32(vgetq_lane_s32(__s2_60, __p3_60), __s0_60, __p1_60); \
   44850   __ret_60; \
   44851 })
   44852 #else
   44853 #define vcopyq_laneq_s32(__p0_61, __p1_61, __p2_61, __p3_61) __extension__ ({ \
   44854   int32x4_t __s0_61 = __p0_61; \
   44855   int32x4_t __s2_61 = __p2_61; \
   44856   int32x4_t __rev0_61;  __rev0_61 = __builtin_shufflevector(__s0_61, __s0_61, 3, 2, 1, 0); \
   44857   int32x4_t __rev2_61;  __rev2_61 = __builtin_shufflevector(__s2_61, __s2_61, 3, 2, 1, 0); \
   44858   int32x4_t __ret_61; \
   44859   __ret_61 = __noswap_vsetq_lane_s32(__noswap_vgetq_lane_s32(__rev2_61, __p3_61), __rev0_61, __p1_61); \
   44860   __ret_61 = __builtin_shufflevector(__ret_61, __ret_61, 3, 2, 1, 0); \
   44861   __ret_61; \
   44862 })
   44863 #endif
   44864 
   44865 #ifdef __LITTLE_ENDIAN__
   44866 #define vcopyq_laneq_s64(__p0_62, __p1_62, __p2_62, __p3_62) __extension__ ({ \
   44867   int64x2_t __s0_62 = __p0_62; \
   44868   int64x2_t __s2_62 = __p2_62; \
   44869   int64x2_t __ret_62; \
   44870   __ret_62 = vsetq_lane_s64(vgetq_lane_s64(__s2_62, __p3_62), __s0_62, __p1_62); \
   44871   __ret_62; \
   44872 })
   44873 #else
   44874 #define vcopyq_laneq_s64(__p0_63, __p1_63, __p2_63, __p3_63) __extension__ ({ \
   44875   int64x2_t __s0_63 = __p0_63; \
   44876   int64x2_t __s2_63 = __p2_63; \
   44877   int64x2_t __rev0_63;  __rev0_63 = __builtin_shufflevector(__s0_63, __s0_63, 1, 0); \
   44878   int64x2_t __rev2_63;  __rev2_63 = __builtin_shufflevector(__s2_63, __s2_63, 1, 0); \
   44879   int64x2_t __ret_63; \
   44880   __ret_63 = __noswap_vsetq_lane_s64(__noswap_vgetq_lane_s64(__rev2_63, __p3_63), __rev0_63, __p1_63); \
   44881   __ret_63 = __builtin_shufflevector(__ret_63, __ret_63, 1, 0); \
   44882   __ret_63; \
   44883 })
   44884 #endif
   44885 
   44886 #ifdef __LITTLE_ENDIAN__
   44887 #define vcopyq_laneq_s16(__p0_64, __p1_64, __p2_64, __p3_64) __extension__ ({ \
   44888   int16x8_t __s0_64 = __p0_64; \
   44889   int16x8_t __s2_64 = __p2_64; \
   44890   int16x8_t __ret_64; \
   44891   __ret_64 = vsetq_lane_s16(vgetq_lane_s16(__s2_64, __p3_64), __s0_64, __p1_64); \
   44892   __ret_64; \
   44893 })
   44894 #else
   44895 #define vcopyq_laneq_s16(__p0_65, __p1_65, __p2_65, __p3_65) __extension__ ({ \
   44896   int16x8_t __s0_65 = __p0_65; \
   44897   int16x8_t __s2_65 = __p2_65; \
   44898   int16x8_t __rev0_65;  __rev0_65 = __builtin_shufflevector(__s0_65, __s0_65, 7, 6, 5, 4, 3, 2, 1, 0); \
   44899   int16x8_t __rev2_65;  __rev2_65 = __builtin_shufflevector(__s2_65, __s2_65, 7, 6, 5, 4, 3, 2, 1, 0); \
   44900   int16x8_t __ret_65; \
   44901   __ret_65 = __noswap_vsetq_lane_s16(__noswap_vgetq_lane_s16(__rev2_65, __p3_65), __rev0_65, __p1_65); \
   44902   __ret_65 = __builtin_shufflevector(__ret_65, __ret_65, 7, 6, 5, 4, 3, 2, 1, 0); \
   44903   __ret_65; \
   44904 })
   44905 #endif
   44906 
   44907 #ifdef __LITTLE_ENDIAN__
   44908 #define vcopy_laneq_p8(__p0_66, __p1_66, __p2_66, __p3_66) __extension__ ({ \
   44909   poly8x8_t __s0_66 = __p0_66; \
   44910   poly8x16_t __s2_66 = __p2_66; \
   44911   poly8x8_t __ret_66; \
   44912   __ret_66 = vset_lane_p8(vgetq_lane_p8(__s2_66, __p3_66), __s0_66, __p1_66); \
   44913   __ret_66; \
   44914 })
   44915 #else
   44916 #define vcopy_laneq_p8(__p0_67, __p1_67, __p2_67, __p3_67) __extension__ ({ \
   44917   poly8x8_t __s0_67 = __p0_67; \
   44918   poly8x16_t __s2_67 = __p2_67; \
   44919   poly8x8_t __rev0_67;  __rev0_67 = __builtin_shufflevector(__s0_67, __s0_67, 7, 6, 5, 4, 3, 2, 1, 0); \
   44920   poly8x16_t __rev2_67;  __rev2_67 = __builtin_shufflevector(__s2_67, __s2_67, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44921   poly8x8_t __ret_67; \
   44922   __ret_67 = __noswap_vset_lane_p8(__noswap_vgetq_lane_p8(__rev2_67, __p3_67), __rev0_67, __p1_67); \
   44923   __ret_67 = __builtin_shufflevector(__ret_67, __ret_67, 7, 6, 5, 4, 3, 2, 1, 0); \
   44924   __ret_67; \
   44925 })
   44926 #endif
   44927 
   44928 #ifdef __LITTLE_ENDIAN__
   44929 #define vcopy_laneq_p16(__p0_68, __p1_68, __p2_68, __p3_68) __extension__ ({ \
   44930   poly16x4_t __s0_68 = __p0_68; \
   44931   poly16x8_t __s2_68 = __p2_68; \
   44932   poly16x4_t __ret_68; \
   44933   __ret_68 = vset_lane_p16(vgetq_lane_p16(__s2_68, __p3_68), __s0_68, __p1_68); \
   44934   __ret_68; \
   44935 })
   44936 #else
   44937 #define vcopy_laneq_p16(__p0_69, __p1_69, __p2_69, __p3_69) __extension__ ({ \
   44938   poly16x4_t __s0_69 = __p0_69; \
   44939   poly16x8_t __s2_69 = __p2_69; \
   44940   poly16x4_t __rev0_69;  __rev0_69 = __builtin_shufflevector(__s0_69, __s0_69, 3, 2, 1, 0); \
   44941   poly16x8_t __rev2_69;  __rev2_69 = __builtin_shufflevector(__s2_69, __s2_69, 7, 6, 5, 4, 3, 2, 1, 0); \
   44942   poly16x4_t __ret_69; \
   44943   __ret_69 = __noswap_vset_lane_p16(__noswap_vgetq_lane_p16(__rev2_69, __p3_69), __rev0_69, __p1_69); \
   44944   __ret_69 = __builtin_shufflevector(__ret_69, __ret_69, 3, 2, 1, 0); \
   44945   __ret_69; \
   44946 })
   44947 #endif
   44948 
   44949 #ifdef __LITTLE_ENDIAN__
   44950 #define vcopy_laneq_u8(__p0_70, __p1_70, __p2_70, __p3_70) __extension__ ({ \
   44951   uint8x8_t __s0_70 = __p0_70; \
   44952   uint8x16_t __s2_70 = __p2_70; \
   44953   uint8x8_t __ret_70; \
   44954   __ret_70 = vset_lane_u8(vgetq_lane_u8(__s2_70, __p3_70), __s0_70, __p1_70); \
   44955   __ret_70; \
   44956 })
   44957 #else
   44958 #define vcopy_laneq_u8(__p0_71, __p1_71, __p2_71, __p3_71) __extension__ ({ \
   44959   uint8x8_t __s0_71 = __p0_71; \
   44960   uint8x16_t __s2_71 = __p2_71; \
   44961   uint8x8_t __rev0_71;  __rev0_71 = __builtin_shufflevector(__s0_71, __s0_71, 7, 6, 5, 4, 3, 2, 1, 0); \
   44962   uint8x16_t __rev2_71;  __rev2_71 = __builtin_shufflevector(__s2_71, __s2_71, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   44963   uint8x8_t __ret_71; \
   44964   __ret_71 = __noswap_vset_lane_u8(__noswap_vgetq_lane_u8(__rev2_71, __p3_71), __rev0_71, __p1_71); \
   44965   __ret_71 = __builtin_shufflevector(__ret_71, __ret_71, 7, 6, 5, 4, 3, 2, 1, 0); \
   44966   __ret_71; \
   44967 })
   44968 #endif
   44969 
   44970 #ifdef __LITTLE_ENDIAN__
   44971 #define vcopy_laneq_u32(__p0_72, __p1_72, __p2_72, __p3_72) __extension__ ({ \
   44972   uint32x2_t __s0_72 = __p0_72; \
   44973   uint32x4_t __s2_72 = __p2_72; \
   44974   uint32x2_t __ret_72; \
   44975   __ret_72 = vset_lane_u32(vgetq_lane_u32(__s2_72, __p3_72), __s0_72, __p1_72); \
   44976   __ret_72; \
   44977 })
   44978 #else
   44979 #define vcopy_laneq_u32(__p0_73, __p1_73, __p2_73, __p3_73) __extension__ ({ \
   44980   uint32x2_t __s0_73 = __p0_73; \
   44981   uint32x4_t __s2_73 = __p2_73; \
   44982   uint32x2_t __rev0_73;  __rev0_73 = __builtin_shufflevector(__s0_73, __s0_73, 1, 0); \
   44983   uint32x4_t __rev2_73;  __rev2_73 = __builtin_shufflevector(__s2_73, __s2_73, 3, 2, 1, 0); \
   44984   uint32x2_t __ret_73; \
   44985   __ret_73 = __noswap_vset_lane_u32(__noswap_vgetq_lane_u32(__rev2_73, __p3_73), __rev0_73, __p1_73); \
   44986   __ret_73 = __builtin_shufflevector(__ret_73, __ret_73, 1, 0); \
   44987   __ret_73; \
   44988 })
   44989 #endif
   44990 
   44991 #ifdef __LITTLE_ENDIAN__
   44992 #define vcopy_laneq_u64(__p0_74, __p1_74, __p2_74, __p3_74) __extension__ ({ \
   44993   uint64x1_t __s0_74 = __p0_74; \
   44994   uint64x2_t __s2_74 = __p2_74; \
   44995   uint64x1_t __ret_74; \
   44996   __ret_74 = vset_lane_u64(vgetq_lane_u64(__s2_74, __p3_74), __s0_74, __p1_74); \
   44997   __ret_74; \
   44998 })
   44999 #else
   45000 #define vcopy_laneq_u64(__p0_75, __p1_75, __p2_75, __p3_75) __extension__ ({ \
   45001   uint64x1_t __s0_75 = __p0_75; \
   45002   uint64x2_t __s2_75 = __p2_75; \
   45003   uint64x2_t __rev2_75;  __rev2_75 = __builtin_shufflevector(__s2_75, __s2_75, 1, 0); \
   45004   uint64x1_t __ret_75; \
   45005   __ret_75 = __noswap_vset_lane_u64(__noswap_vgetq_lane_u64(__rev2_75, __p3_75), __s0_75, __p1_75); \
   45006   __ret_75; \
   45007 })
   45008 #endif
   45009 
   45010 #ifdef __LITTLE_ENDIAN__
   45011 #define vcopy_laneq_u16(__p0_76, __p1_76, __p2_76, __p3_76) __extension__ ({ \
   45012   uint16x4_t __s0_76 = __p0_76; \
   45013   uint16x8_t __s2_76 = __p2_76; \
   45014   uint16x4_t __ret_76; \
   45015   __ret_76 = vset_lane_u16(vgetq_lane_u16(__s2_76, __p3_76), __s0_76, __p1_76); \
   45016   __ret_76; \
   45017 })
   45018 #else
   45019 #define vcopy_laneq_u16(__p0_77, __p1_77, __p2_77, __p3_77) __extension__ ({ \
   45020   uint16x4_t __s0_77 = __p0_77; \
   45021   uint16x8_t __s2_77 = __p2_77; \
   45022   uint16x4_t __rev0_77;  __rev0_77 = __builtin_shufflevector(__s0_77, __s0_77, 3, 2, 1, 0); \
   45023   uint16x8_t __rev2_77;  __rev2_77 = __builtin_shufflevector(__s2_77, __s2_77, 7, 6, 5, 4, 3, 2, 1, 0); \
   45024   uint16x4_t __ret_77; \
   45025   __ret_77 = __noswap_vset_lane_u16(__noswap_vgetq_lane_u16(__rev2_77, __p3_77), __rev0_77, __p1_77); \
   45026   __ret_77 = __builtin_shufflevector(__ret_77, __ret_77, 3, 2, 1, 0); \
   45027   __ret_77; \
   45028 })
   45029 #endif
   45030 
   45031 #ifdef __LITTLE_ENDIAN__
   45032 #define vcopy_laneq_s8(__p0_78, __p1_78, __p2_78, __p3_78) __extension__ ({ \
   45033   int8x8_t __s0_78 = __p0_78; \
   45034   int8x16_t __s2_78 = __p2_78; \
   45035   int8x8_t __ret_78; \
   45036   __ret_78 = vset_lane_s8(vgetq_lane_s8(__s2_78, __p3_78), __s0_78, __p1_78); \
   45037   __ret_78; \
   45038 })
   45039 #else
   45040 #define vcopy_laneq_s8(__p0_79, __p1_79, __p2_79, __p3_79) __extension__ ({ \
   45041   int8x8_t __s0_79 = __p0_79; \
   45042   int8x16_t __s2_79 = __p2_79; \
   45043   int8x8_t __rev0_79;  __rev0_79 = __builtin_shufflevector(__s0_79, __s0_79, 7, 6, 5, 4, 3, 2, 1, 0); \
   45044   int8x16_t __rev2_79;  __rev2_79 = __builtin_shufflevector(__s2_79, __s2_79, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   45045   int8x8_t __ret_79; \
   45046   __ret_79 = __noswap_vset_lane_s8(__noswap_vgetq_lane_s8(__rev2_79, __p3_79), __rev0_79, __p1_79); \
   45047   __ret_79 = __builtin_shufflevector(__ret_79, __ret_79, 7, 6, 5, 4, 3, 2, 1, 0); \
   45048   __ret_79; \
   45049 })
   45050 #endif
   45051 
   45052 #ifdef __LITTLE_ENDIAN__
   45053 #define vcopy_laneq_f32(__p0_80, __p1_80, __p2_80, __p3_80) __extension__ ({ \
   45054   float32x2_t __s0_80 = __p0_80; \
   45055   float32x4_t __s2_80 = __p2_80; \
   45056   float32x2_t __ret_80; \
   45057   __ret_80 = vset_lane_f32(vgetq_lane_f32(__s2_80, __p3_80), __s0_80, __p1_80); \
   45058   __ret_80; \
   45059 })
   45060 #else
   45061 #define vcopy_laneq_f32(__p0_81, __p1_81, __p2_81, __p3_81) __extension__ ({ \
   45062   float32x2_t __s0_81 = __p0_81; \
   45063   float32x4_t __s2_81 = __p2_81; \
   45064   float32x2_t __rev0_81;  __rev0_81 = __builtin_shufflevector(__s0_81, __s0_81, 1, 0); \
   45065   float32x4_t __rev2_81;  __rev2_81 = __builtin_shufflevector(__s2_81, __s2_81, 3, 2, 1, 0); \
   45066   float32x2_t __ret_81; \
   45067   __ret_81 = __noswap_vset_lane_f32(__noswap_vgetq_lane_f32(__rev2_81, __p3_81), __rev0_81, __p1_81); \
   45068   __ret_81 = __builtin_shufflevector(__ret_81, __ret_81, 1, 0); \
   45069   __ret_81; \
   45070 })
   45071 #endif
   45072 
   45073 #ifdef __LITTLE_ENDIAN__
   45074 #define vcopy_laneq_s32(__p0_82, __p1_82, __p2_82, __p3_82) __extension__ ({ \
   45075   int32x2_t __s0_82 = __p0_82; \
   45076   int32x4_t __s2_82 = __p2_82; \
   45077   int32x2_t __ret_82; \
   45078   __ret_82 = vset_lane_s32(vgetq_lane_s32(__s2_82, __p3_82), __s0_82, __p1_82); \
   45079   __ret_82; \
   45080 })
   45081 #else
   45082 #define vcopy_laneq_s32(__p0_83, __p1_83, __p2_83, __p3_83) __extension__ ({ \
   45083   int32x2_t __s0_83 = __p0_83; \
   45084   int32x4_t __s2_83 = __p2_83; \
   45085   int32x2_t __rev0_83;  __rev0_83 = __builtin_shufflevector(__s0_83, __s0_83, 1, 0); \
   45086   int32x4_t __rev2_83;  __rev2_83 = __builtin_shufflevector(__s2_83, __s2_83, 3, 2, 1, 0); \
   45087   int32x2_t __ret_83; \
   45088   __ret_83 = __noswap_vset_lane_s32(__noswap_vgetq_lane_s32(__rev2_83, __p3_83), __rev0_83, __p1_83); \
   45089   __ret_83 = __builtin_shufflevector(__ret_83, __ret_83, 1, 0); \
   45090   __ret_83; \
   45091 })
   45092 #endif
   45093 
   45094 #ifdef __LITTLE_ENDIAN__
   45095 #define vcopy_laneq_s64(__p0_84, __p1_84, __p2_84, __p3_84) __extension__ ({ \
   45096   int64x1_t __s0_84 = __p0_84; \
   45097   int64x2_t __s2_84 = __p2_84; \
   45098   int64x1_t __ret_84; \
   45099   __ret_84 = vset_lane_s64(vgetq_lane_s64(__s2_84, __p3_84), __s0_84, __p1_84); \
   45100   __ret_84; \
   45101 })
   45102 #else
   45103 #define vcopy_laneq_s64(__p0_85, __p1_85, __p2_85, __p3_85) __extension__ ({ \
   45104   int64x1_t __s0_85 = __p0_85; \
   45105   int64x2_t __s2_85 = __p2_85; \
   45106   int64x2_t __rev2_85;  __rev2_85 = __builtin_shufflevector(__s2_85, __s2_85, 1, 0); \
   45107   int64x1_t __ret_85; \
   45108   __ret_85 = __noswap_vset_lane_s64(__noswap_vgetq_lane_s64(__rev2_85, __p3_85), __s0_85, __p1_85); \
   45109   __ret_85; \
   45110 })
   45111 #endif
   45112 
   45113 #ifdef __LITTLE_ENDIAN__
   45114 #define vcopy_laneq_s16(__p0_86, __p1_86, __p2_86, __p3_86) __extension__ ({ \
   45115   int16x4_t __s0_86 = __p0_86; \
   45116   int16x8_t __s2_86 = __p2_86; \
   45117   int16x4_t __ret_86; \
   45118   __ret_86 = vset_lane_s16(vgetq_lane_s16(__s2_86, __p3_86), __s0_86, __p1_86); \
   45119   __ret_86; \
   45120 })
   45121 #else
   45122 #define vcopy_laneq_s16(__p0_87, __p1_87, __p2_87, __p3_87) __extension__ ({ \
   45123   int16x4_t __s0_87 = __p0_87; \
   45124   int16x8_t __s2_87 = __p2_87; \
   45125   int16x4_t __rev0_87;  __rev0_87 = __builtin_shufflevector(__s0_87, __s0_87, 3, 2, 1, 0); \
   45126   int16x8_t __rev2_87;  __rev2_87 = __builtin_shufflevector(__s2_87, __s2_87, 7, 6, 5, 4, 3, 2, 1, 0); \
   45127   int16x4_t __ret_87; \
   45128   __ret_87 = __noswap_vset_lane_s16(__noswap_vgetq_lane_s16(__rev2_87, __p3_87), __rev0_87, __p1_87); \
   45129   __ret_87 = __builtin_shufflevector(__ret_87, __ret_87, 3, 2, 1, 0); \
   45130   __ret_87; \
   45131 })
   45132 #endif
   45133 
   45134 #ifdef __LITTLE_ENDIAN__
   45135 __ai poly64x1_t vcreate_p64(uint64_t __p0) {
   45136   poly64x1_t __ret;
   45137   __ret = (poly64x1_t)(__p0);
   45138   return __ret;
   45139 }
   45140 #else
   45141 __ai poly64x1_t vcreate_p64(uint64_t __p0) {
   45142   poly64x1_t __ret;
   45143   __ret = (poly64x1_t)(__p0);
   45144   return __ret;
   45145 }
   45146 #endif
   45147 
   45148 #ifdef __LITTLE_ENDIAN__
   45149 __ai float64x1_t vcreate_f64(uint64_t __p0) {
   45150   float64x1_t __ret;
   45151   __ret = (float64x1_t)(__p0);
   45152   return __ret;
   45153 }
   45154 #else
   45155 __ai float64x1_t vcreate_f64(uint64_t __p0) {
   45156   float64x1_t __ret;
   45157   __ret = (float64x1_t)(__p0);
   45158   return __ret;
   45159 }
   45160 #endif
   45161 
   45162 #ifdef __LITTLE_ENDIAN__
   45163 __ai float32_t vcvts_f32_s32(int32_t __p0) {
   45164   float32_t __ret;
   45165   __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
   45166   return __ret;
   45167 }
   45168 #else
   45169 __ai float32_t vcvts_f32_s32(int32_t __p0) {
   45170   float32_t __ret;
   45171   __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
   45172   return __ret;
   45173 }
   45174 #endif
   45175 
   45176 #ifdef __LITTLE_ENDIAN__
   45177 __ai float32_t vcvts_f32_u32(uint32_t __p0) {
   45178   float32_t __ret;
   45179   __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
   45180   return __ret;
   45181 }
   45182 #else
   45183 __ai float32_t vcvts_f32_u32(uint32_t __p0) {
   45184   float32_t __ret;
   45185   __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
   45186   return __ret;
   45187 }
   45188 #endif
   45189 
   45190 #ifdef __LITTLE_ENDIAN__
   45191 __ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
   45192   float32x2_t __ret;
   45193   __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
   45194   return __ret;
   45195 }
   45196 #else
   45197 __ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
   45198   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   45199   float32x2_t __ret;
   45200   __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__rev0, 9);
   45201   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   45202   return __ret;
   45203 }
   45204 __ai float32x2_t __noswap_vcvt_f32_f64(float64x2_t __p0) {
   45205   float32x2_t __ret;
   45206   __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
   45207   return __ret;
   45208 }
   45209 #endif
   45210 
   45211 #ifdef __LITTLE_ENDIAN__
   45212 __ai float64_t vcvtd_f64_s64(int64_t __p0) {
   45213   float64_t __ret;
   45214   __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
   45215   return __ret;
   45216 }
   45217 #else
   45218 __ai float64_t vcvtd_f64_s64(int64_t __p0) {
   45219   float64_t __ret;
   45220   __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
   45221   return __ret;
   45222 }
   45223 #endif
   45224 
   45225 #ifdef __LITTLE_ENDIAN__
   45226 __ai float64_t vcvtd_f64_u64(uint64_t __p0) {
   45227   float64_t __ret;
   45228   __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
   45229   return __ret;
   45230 }
   45231 #else
   45232 __ai float64_t vcvtd_f64_u64(uint64_t __p0) {
   45233   float64_t __ret;
   45234   __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
   45235   return __ret;
   45236 }
   45237 #endif
   45238 
   45239 #ifdef __LITTLE_ENDIAN__
   45240 __ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
   45241   float64x2_t __ret;
   45242   __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 51);
   45243   return __ret;
   45244 }
   45245 #else
   45246 __ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
   45247   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   45248   float64x2_t __ret;
   45249   __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 51);
   45250   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   45251   return __ret;
   45252 }
   45253 #endif
   45254 
   45255 #ifdef __LITTLE_ENDIAN__
   45256 __ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
   45257   float64x2_t __ret;
   45258   __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 35);
   45259   return __ret;
   45260 }
   45261 #else
   45262 __ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
   45263   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   45264   float64x2_t __ret;
   45265   __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 35);
   45266   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   45267   return __ret;
   45268 }
   45269 #endif
   45270 
   45271 #ifdef __LITTLE_ENDIAN__
   45272 __ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
   45273   float64x1_t __ret;
   45274   __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
   45275   return __ret;
   45276 }
   45277 #else
   45278 __ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
   45279   float64x1_t __ret;
   45280   __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
   45281   return __ret;
   45282 }
   45283 #endif
   45284 
   45285 #ifdef __LITTLE_ENDIAN__
   45286 __ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
   45287   float64x1_t __ret;
   45288   __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
   45289   return __ret;
   45290 }
   45291 #else
   45292 __ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
   45293   float64x1_t __ret;
   45294   __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
   45295   return __ret;
   45296 }
   45297 #endif
   45298 
   45299 #ifdef __LITTLE_ENDIAN__
   45300 __ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
   45301   float64x2_t __ret;
   45302   __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
   45303   return __ret;
   45304 }
   45305 #else
   45306 __ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
   45307   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   45308   float64x2_t __ret;
   45309   __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__rev0, 42);
   45310   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   45311   return __ret;
   45312 }
   45313 __ai float64x2_t __noswap_vcvt_f64_f32(float32x2_t __p0) {
   45314   float64x2_t __ret;
   45315   __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
   45316   return __ret;
   45317 }
   45318 #endif
   45319 
   45320 #ifdef __LITTLE_ENDIAN__
   45321 __ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
   45322   float16x8_t __ret;
   45323   __ret = vcombine_f16(__p0, vcvt_f16_f32(__p1));
   45324   return __ret;
   45325 }
   45326 #else
   45327 __ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
   45328   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   45329   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   45330   float16x8_t __ret;
   45331   __ret = __noswap_vcombine_f16(__rev0, __noswap_vcvt_f16_f32(__rev1));
   45332   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   45333   return __ret;
   45334 }
   45335 #endif
   45336 
   45337 #ifdef __LITTLE_ENDIAN__
   45338 __ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
   45339   float32x4_t __ret;
   45340   __ret = vcvt_f32_f16(vget_high_f16(__p0));
   45341   return __ret;
   45342 }
   45343 #else
   45344 __ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
   45345   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   45346   float32x4_t __ret;
   45347   __ret = __noswap_vcvt_f32_f16(__noswap_vget_high_f16(__rev0));
   45348   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   45349   return __ret;
   45350 }
   45351 #endif
   45352 
   45353 #ifdef __LITTLE_ENDIAN__
   45354 __ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
   45355   float32x4_t __ret;
   45356   __ret = vcombine_f32(__p0, vcvt_f32_f64(__p1));
   45357   return __ret;
   45358 }
   45359 #else
   45360 __ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
   45361   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   45362   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   45363   float32x4_t __ret;
   45364   __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvt_f32_f64(__rev1));
   45365   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   45366   return __ret;
   45367 }
   45368 #endif
   45369 
   45370 #ifdef __LITTLE_ENDIAN__
   45371 __ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
   45372   float64x2_t __ret;
   45373   __ret = vcvt_f64_f32(vget_high_f32(__p0));
   45374   return __ret;
   45375 }
   45376 #else
   45377 __ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
   45378   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   45379   float64x2_t __ret;
   45380   __ret = __noswap_vcvt_f64_f32(__noswap_vget_high_f32(__rev0));
   45381   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   45382   return __ret;
   45383 }
   45384 #endif
   45385 
   45386 #ifdef __LITTLE_ENDIAN__
   45387 #define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
   45388   uint32_t __s0 = __p0; \
   45389   float32_t __ret; \
   45390   __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
   45391   __ret; \
   45392 })
   45393 #else
   45394 #define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
   45395   uint32_t __s0 = __p0; \
   45396   float32_t __ret; \
   45397   __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
   45398   __ret; \
   45399 })
   45400 #endif
   45401 
   45402 #ifdef __LITTLE_ENDIAN__
   45403 #define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
   45404   int32_t __s0 = __p0; \
   45405   float32_t __ret; \
   45406   __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
   45407   __ret; \
   45408 })
   45409 #else
   45410 #define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
   45411   int32_t __s0 = __p0; \
   45412   float32_t __ret; \
   45413   __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
   45414   __ret; \
   45415 })
   45416 #endif
   45417 
   45418 #ifdef __LITTLE_ENDIAN__
   45419 #define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
   45420   uint64x2_t __s0 = __p0; \
   45421   float64x2_t __ret; \
   45422   __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 51); \
   45423   __ret; \
   45424 })
   45425 #else
   45426 #define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
   45427   uint64x2_t __s0 = __p0; \
   45428   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   45429   float64x2_t __ret; \
   45430   __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 51); \
   45431   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   45432   __ret; \
   45433 })
   45434 #endif
   45435 
   45436 #ifdef __LITTLE_ENDIAN__
   45437 #define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
   45438   int64x2_t __s0 = __p0; \
   45439   float64x2_t __ret; \
   45440   __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 35); \
   45441   __ret; \
   45442 })
   45443 #else
   45444 #define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
   45445   int64x2_t __s0 = __p0; \
   45446   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   45447   float64x2_t __ret; \
   45448   __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 35); \
   45449   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   45450   __ret; \
   45451 })
   45452 #endif
   45453 
   45454 #ifdef __LITTLE_ENDIAN__
   45455 #define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
   45456   uint64x1_t __s0 = __p0; \
   45457   float64x1_t __ret; \
   45458   __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
   45459   __ret; \
   45460 })
   45461 #else
   45462 #define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
   45463   uint64x1_t __s0 = __p0; \
   45464   float64x1_t __ret; \
   45465   __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
   45466   __ret; \
   45467 })
   45468 #endif
   45469 
   45470 #ifdef __LITTLE_ENDIAN__
   45471 #define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
   45472   int64x1_t __s0 = __p0; \
   45473   float64x1_t __ret; \
   45474   __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
   45475   __ret; \
   45476 })
   45477 #else
   45478 #define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
   45479   int64x1_t __s0 = __p0; \
   45480   float64x1_t __ret; \
   45481   __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
   45482   __ret; \
   45483 })
   45484 #endif
   45485 
   45486 #ifdef __LITTLE_ENDIAN__
   45487 #define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
   45488   uint64_t __s0 = __p0; \
   45489   float64_t __ret; \
   45490   __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
   45491   __ret; \
   45492 })
   45493 #else
   45494 #define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
   45495   uint64_t __s0 = __p0; \
   45496   float64_t __ret; \
   45497   __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
   45498   __ret; \
   45499 })
   45500 #endif
   45501 
   45502 #ifdef __LITTLE_ENDIAN__
   45503 #define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
   45504   int64_t __s0 = __p0; \
   45505   float64_t __ret; \
   45506   __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
   45507   __ret; \
   45508 })
   45509 #else
   45510 #define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
   45511   int64_t __s0 = __p0; \
   45512   float64_t __ret; \
   45513   __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
   45514   __ret; \
   45515 })
   45516 #endif
   45517 
   45518 #ifdef __LITTLE_ENDIAN__
   45519 #define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
   45520   float32_t __s0 = __p0; \
   45521   int32_t __ret; \
   45522   __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
   45523   __ret; \
   45524 })
   45525 #else
   45526 #define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
   45527   float32_t __s0 = __p0; \
   45528   int32_t __ret; \
   45529   __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
   45530   __ret; \
   45531 })
   45532 #endif
   45533 
   45534 #ifdef __LITTLE_ENDIAN__
   45535 #define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
   45536   float64x2_t __s0 = __p0; \
   45537   int64x2_t __ret; \
   45538   __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__s0, __p1, 35); \
   45539   __ret; \
   45540 })
   45541 #else
   45542 #define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
   45543   float64x2_t __s0 = __p0; \
   45544   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   45545   int64x2_t __ret; \
   45546   __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__rev0, __p1, 35); \
   45547   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   45548   __ret; \
   45549 })
   45550 #endif
   45551 
   45552 #ifdef __LITTLE_ENDIAN__
   45553 #define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
   45554   float64x1_t __s0 = __p0; \
   45555   int64x1_t __ret; \
   45556   __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
   45557   __ret; \
   45558 })
   45559 #else
   45560 #define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
   45561   float64x1_t __s0 = __p0; \
   45562   int64x1_t __ret; \
   45563   __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
   45564   __ret; \
   45565 })
   45566 #endif
   45567 
   45568 #ifdef __LITTLE_ENDIAN__
   45569 #define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
   45570   float64_t __s0 = __p0; \
   45571   int64_t __ret; \
   45572   __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
   45573   __ret; \
   45574 })
   45575 #else
   45576 #define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
   45577   float64_t __s0 = __p0; \
   45578   int64_t __ret; \
   45579   __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
   45580   __ret; \
   45581 })
   45582 #endif
   45583 
   45584 #ifdef __LITTLE_ENDIAN__
   45585 #define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
   45586   float32_t __s0 = __p0; \
   45587   uint32_t __ret; \
   45588   __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
   45589   __ret; \
   45590 })
   45591 #else
   45592 #define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
   45593   float32_t __s0 = __p0; \
   45594   uint32_t __ret; \
   45595   __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
   45596   __ret; \
   45597 })
   45598 #endif
   45599 
   45600 #ifdef __LITTLE_ENDIAN__
   45601 #define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
   45602   float64x2_t __s0 = __p0; \
   45603   uint64x2_t __ret; \
   45604   __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__s0, __p1, 51); \
   45605   __ret; \
   45606 })
   45607 #else
   45608 #define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
   45609   float64x2_t __s0 = __p0; \
   45610   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   45611   uint64x2_t __ret; \
   45612   __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__rev0, __p1, 51); \
   45613   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   45614   __ret; \
   45615 })
   45616 #endif
   45617 
   45618 #ifdef __LITTLE_ENDIAN__
   45619 #define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
   45620   float64x1_t __s0 = __p0; \
   45621   uint64x1_t __ret; \
   45622   __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
   45623   __ret; \
   45624 })
   45625 #else
   45626 #define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
   45627   float64x1_t __s0 = __p0; \
   45628   uint64x1_t __ret; \
   45629   __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
   45630   __ret; \
   45631 })
   45632 #endif
   45633 
   45634 #ifdef __LITTLE_ENDIAN__
   45635 #define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
   45636   float64_t __s0 = __p0; \
   45637   uint64_t __ret; \
   45638   __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
   45639   __ret; \
   45640 })
   45641 #else
   45642 #define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
   45643   float64_t __s0 = __p0; \
   45644   uint64_t __ret; \
   45645   __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
   45646   __ret; \
   45647 })
   45648 #endif
   45649 
   45650 #ifdef __LITTLE_ENDIAN__
   45651 __ai int32_t vcvts_s32_f32(float32_t __p0) {
   45652   int32_t __ret;
   45653   __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
   45654   return __ret;
   45655 }
   45656 #else
   45657 __ai int32_t vcvts_s32_f32(float32_t __p0) {
   45658   int32_t __ret;
   45659   __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
   45660   return __ret;
   45661 }
   45662 #endif
   45663 
   45664 #ifdef __LITTLE_ENDIAN__
   45665 __ai int64_t vcvtd_s64_f64(float64_t __p0) {
   45666   int64_t __ret;
   45667   __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
   45668   return __ret;
   45669 }
   45670 #else
   45671 __ai int64_t vcvtd_s64_f64(float64_t __p0) {
   45672   int64_t __ret;
   45673   __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
   45674   return __ret;
   45675 }
   45676 #endif
   45677 
   45678 #ifdef __LITTLE_ENDIAN__
   45679 __ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
   45680   int64x2_t __ret;
   45681   __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__p0, 35);
   45682   return __ret;
   45683 }
   45684 #else
   45685 __ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
   45686   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   45687   int64x2_t __ret;
   45688   __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__rev0, 35);
   45689   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   45690   return __ret;
   45691 }
   45692 #endif
   45693 
   45694 #ifdef __LITTLE_ENDIAN__
   45695 __ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
   45696   int64x1_t __ret;
   45697   __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
   45698   return __ret;
   45699 }
   45700 #else
   45701 __ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
   45702   int64x1_t __ret;
   45703   __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
   45704   return __ret;
   45705 }
   45706 #endif
   45707 
   45708 #ifdef __LITTLE_ENDIAN__
   45709 __ai uint32_t vcvts_u32_f32(float32_t __p0) {
   45710   uint32_t __ret;
   45711   __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
   45712   return __ret;
   45713 }
   45714 #else
   45715 __ai uint32_t vcvts_u32_f32(float32_t __p0) {
   45716   uint32_t __ret;
   45717   __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
   45718   return __ret;
   45719 }
   45720 #endif
   45721 
   45722 #ifdef __LITTLE_ENDIAN__
   45723 __ai uint64_t vcvtd_u64_f64(float64_t __p0) {
   45724   uint64_t __ret;
   45725   __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
   45726   return __ret;
   45727 }
   45728 #else
   45729 __ai uint64_t vcvtd_u64_f64(float64_t __p0) {
   45730   uint64_t __ret;
   45731   __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
   45732   return __ret;
   45733 }
   45734 #endif
   45735 
   45736 #ifdef __LITTLE_ENDIAN__
   45737 __ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
   45738   uint64x2_t __ret;
   45739   __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__p0, 51);
   45740   return __ret;
   45741 }
   45742 #else
   45743 __ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
   45744   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   45745   uint64x2_t __ret;
   45746   __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__rev0, 51);
   45747   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   45748   return __ret;
   45749 }
   45750 #endif
   45751 
   45752 #ifdef __LITTLE_ENDIAN__
   45753 __ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
   45754   uint64x1_t __ret;
   45755   __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
   45756   return __ret;
   45757 }
   45758 #else
   45759 __ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
   45760   uint64x1_t __ret;
   45761   __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
   45762   return __ret;
   45763 }
   45764 #endif
   45765 
   45766 #ifdef __LITTLE_ENDIAN__
   45767 __ai int32_t vcvtas_s32_f32(float32_t __p0) {
   45768   int32_t __ret;
   45769   __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
   45770   return __ret;
   45771 }
   45772 #else
   45773 __ai int32_t vcvtas_s32_f32(float32_t __p0) {
   45774   int32_t __ret;
   45775   __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
   45776   return __ret;
   45777 }
   45778 #endif
   45779 
   45780 #ifdef __LITTLE_ENDIAN__
   45781 __ai int64_t vcvtad_s64_f64(float64_t __p0) {
   45782   int64_t __ret;
   45783   __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
   45784   return __ret;
   45785 }
   45786 #else
   45787 __ai int64_t vcvtad_s64_f64(float64_t __p0) {
   45788   int64_t __ret;
   45789   __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
   45790   return __ret;
   45791 }
   45792 #endif
   45793 
   45794 #ifdef __LITTLE_ENDIAN__
   45795 __ai uint32_t vcvtas_u32_f32(float32_t __p0) {
   45796   uint32_t __ret;
   45797   __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
   45798   return __ret;
   45799 }
   45800 #else
   45801 __ai uint32_t vcvtas_u32_f32(float32_t __p0) {
   45802   uint32_t __ret;
   45803   __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
   45804   return __ret;
   45805 }
   45806 #endif
   45807 
   45808 #ifdef __LITTLE_ENDIAN__
   45809 __ai uint64_t vcvtad_u64_f64(float64_t __p0) {
   45810   uint64_t __ret;
   45811   __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
   45812   return __ret;
   45813 }
   45814 #else
   45815 __ai uint64_t vcvtad_u64_f64(float64_t __p0) {
   45816   uint64_t __ret;
   45817   __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
   45818   return __ret;
   45819 }
   45820 #endif
   45821 
   45822 #ifdef __LITTLE_ENDIAN__
   45823 __ai int32_t vcvtms_s32_f32(float32_t __p0) {
   45824   int32_t __ret;
   45825   __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
   45826   return __ret;
   45827 }
   45828 #else
   45829 __ai int32_t vcvtms_s32_f32(float32_t __p0) {
   45830   int32_t __ret;
   45831   __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
   45832   return __ret;
   45833 }
   45834 #endif
   45835 
   45836 #ifdef __LITTLE_ENDIAN__
   45837 __ai int64_t vcvtmd_s64_f64(float64_t __p0) {
   45838   int64_t __ret;
   45839   __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
   45840   return __ret;
   45841 }
   45842 #else
   45843 __ai int64_t vcvtmd_s64_f64(float64_t __p0) {
   45844   int64_t __ret;
   45845   __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
   45846   return __ret;
   45847 }
   45848 #endif
   45849 
   45850 #ifdef __LITTLE_ENDIAN__
   45851 __ai uint32_t vcvtms_u32_f32(float32_t __p0) {
   45852   uint32_t __ret;
   45853   __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
   45854   return __ret;
   45855 }
   45856 #else
   45857 __ai uint32_t vcvtms_u32_f32(float32_t __p0) {
   45858   uint32_t __ret;
   45859   __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
   45860   return __ret;
   45861 }
   45862 #endif
   45863 
   45864 #ifdef __LITTLE_ENDIAN__
   45865 __ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
   45866   uint64_t __ret;
   45867   __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
   45868   return __ret;
   45869 }
   45870 #else
   45871 __ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
   45872   uint64_t __ret;
   45873   __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
   45874   return __ret;
   45875 }
   45876 #endif
   45877 
   45878 #ifdef __LITTLE_ENDIAN__
   45879 __ai int32_t vcvtns_s32_f32(float32_t __p0) {
   45880   int32_t __ret;
   45881   __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
   45882   return __ret;
   45883 }
   45884 #else
   45885 __ai int32_t vcvtns_s32_f32(float32_t __p0) {
   45886   int32_t __ret;
   45887   __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
   45888   return __ret;
   45889 }
   45890 #endif
   45891 
   45892 #ifdef __LITTLE_ENDIAN__
   45893 __ai int64_t vcvtnd_s64_f64(float64_t __p0) {
   45894   int64_t __ret;
   45895   __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
   45896   return __ret;
   45897 }
   45898 #else
   45899 __ai int64_t vcvtnd_s64_f64(float64_t __p0) {
   45900   int64_t __ret;
   45901   __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
   45902   return __ret;
   45903 }
   45904 #endif
   45905 
   45906 #ifdef __LITTLE_ENDIAN__
   45907 __ai uint32_t vcvtns_u32_f32(float32_t __p0) {
   45908   uint32_t __ret;
   45909   __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
   45910   return __ret;
   45911 }
   45912 #else
   45913 __ai uint32_t vcvtns_u32_f32(float32_t __p0) {
   45914   uint32_t __ret;
   45915   __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
   45916   return __ret;
   45917 }
   45918 #endif
   45919 
   45920 #ifdef __LITTLE_ENDIAN__
   45921 __ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
   45922   uint64_t __ret;
   45923   __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
   45924   return __ret;
   45925 }
   45926 #else
   45927 __ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
   45928   uint64_t __ret;
   45929   __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
   45930   return __ret;
   45931 }
   45932 #endif
   45933 
   45934 #ifdef __LITTLE_ENDIAN__
   45935 __ai int32_t vcvtps_s32_f32(float32_t __p0) {
   45936   int32_t __ret;
   45937   __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
   45938   return __ret;
   45939 }
   45940 #else
   45941 __ai int32_t vcvtps_s32_f32(float32_t __p0) {
   45942   int32_t __ret;
   45943   __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
   45944   return __ret;
   45945 }
   45946 #endif
   45947 
   45948 #ifdef __LITTLE_ENDIAN__
   45949 __ai int64_t vcvtpd_s64_f64(float64_t __p0) {
   45950   int64_t __ret;
   45951   __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
   45952   return __ret;
   45953 }
   45954 #else
   45955 __ai int64_t vcvtpd_s64_f64(float64_t __p0) {
   45956   int64_t __ret;
   45957   __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
   45958   return __ret;
   45959 }
   45960 #endif
   45961 
   45962 #ifdef __LITTLE_ENDIAN__
   45963 __ai uint32_t vcvtps_u32_f32(float32_t __p0) {
   45964   uint32_t __ret;
   45965   __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
   45966   return __ret;
   45967 }
   45968 #else
   45969 __ai uint32_t vcvtps_u32_f32(float32_t __p0) {
   45970   uint32_t __ret;
   45971   __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
   45972   return __ret;
   45973 }
   45974 #endif
   45975 
   45976 #ifdef __LITTLE_ENDIAN__
   45977 __ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
   45978   uint64_t __ret;
   45979   __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
   45980   return __ret;
   45981 }
   45982 #else
   45983 __ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
   45984   uint64_t __ret;
   45985   __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
   45986   return __ret;
   45987 }
   45988 #endif
   45989 
   45990 #ifdef __LITTLE_ENDIAN__
   45991 __ai float32_t vcvtxd_f32_f64(float64_t __p0) {
   45992   float32_t __ret;
   45993   __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
   45994   return __ret;
   45995 }
   45996 #else
   45997 __ai float32_t vcvtxd_f32_f64(float64_t __p0) {
   45998   float32_t __ret;
   45999   __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
   46000   return __ret;
   46001 }
   46002 #endif
   46003 
   46004 #ifdef __LITTLE_ENDIAN__
   46005 __ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
   46006   float32x2_t __ret;
   46007   __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
   46008   return __ret;
   46009 }
   46010 #else
   46011 __ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
   46012   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   46013   float32x2_t __ret;
   46014   __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__rev0, 42);
   46015   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   46016   return __ret;
   46017 }
   46018 __ai float32x2_t __noswap_vcvtx_f32_f64(float64x2_t __p0) {
   46019   float32x2_t __ret;
   46020   __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
   46021   return __ret;
   46022 }
   46023 #endif
   46024 
   46025 #ifdef __LITTLE_ENDIAN__
   46026 __ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
   46027   float32x4_t __ret;
   46028   __ret = vcombine_f32(__p0, vcvtx_f32_f64(__p1));
   46029   return __ret;
   46030 }
   46031 #else
   46032 __ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
   46033   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   46034   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   46035   float32x4_t __ret;
   46036   __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvtx_f32_f64(__rev1));
   46037   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   46038   return __ret;
   46039 }
   46040 #endif
   46041 
   46042 #ifdef __LITTLE_ENDIAN__
   46043 __ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
   46044   float64x2_t __ret;
   46045   __ret = __p0 / __p1;
   46046   return __ret;
   46047 }
   46048 #else
   46049 __ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
   46050   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   46051   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   46052   float64x2_t __ret;
   46053   __ret = __rev0 / __rev1;
   46054   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   46055   return __ret;
   46056 }
   46057 #endif
   46058 
   46059 #ifdef __LITTLE_ENDIAN__
   46060 __ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
   46061   float32x4_t __ret;
   46062   __ret = __p0 / __p1;
   46063   return __ret;
   46064 }
   46065 #else
   46066 __ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
   46067   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   46068   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   46069   float32x4_t __ret;
   46070   __ret = __rev0 / __rev1;
   46071   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   46072   return __ret;
   46073 }
   46074 #endif
   46075 
   46076 #ifdef __LITTLE_ENDIAN__
   46077 __ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
   46078   float64x1_t __ret;
   46079   __ret = __p0 / __p1;
   46080   return __ret;
   46081 }
   46082 #else
   46083 __ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
   46084   float64x1_t __ret;
   46085   __ret = __p0 / __p1;
   46086   return __ret;
   46087 }
   46088 #endif
   46089 
   46090 #ifdef __LITTLE_ENDIAN__
   46091 __ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
   46092   float32x2_t __ret;
   46093   __ret = __p0 / __p1;
   46094   return __ret;
   46095 }
   46096 #else
   46097 __ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
   46098   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   46099   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   46100   float32x2_t __ret;
   46101   __ret = __rev0 / __rev1;
   46102   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   46103   return __ret;
   46104 }
   46105 #endif
   46106 
   46107 #ifdef __LITTLE_ENDIAN__
   46108 #define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
   46109   poly8x8_t __s0 = __p0; \
   46110   poly8_t __ret; \
   46111   __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
   46112   __ret; \
   46113 })
   46114 #else
   46115 #define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
   46116   poly8x8_t __s0 = __p0; \
   46117   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46118   poly8_t __ret; \
   46119   __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
   46120   __ret; \
   46121 })
   46122 #endif
   46123 
   46124 #ifdef __LITTLE_ENDIAN__
   46125 #define vduph_lane_p16(__p0, __p1) __extension__ ({ \
   46126   poly16x4_t __s0 = __p0; \
   46127   poly16_t __ret; \
   46128   __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
   46129   __ret; \
   46130 })
   46131 #else
   46132 #define vduph_lane_p16(__p0, __p1) __extension__ ({ \
   46133   poly16x4_t __s0 = __p0; \
   46134   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46135   poly16_t __ret; \
   46136   __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
   46137   __ret; \
   46138 })
   46139 #endif
   46140 
   46141 #ifdef __LITTLE_ENDIAN__
   46142 #define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
   46143   uint8x8_t __s0 = __p0; \
   46144   uint8_t __ret; \
   46145   __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
   46146   __ret; \
   46147 })
   46148 #else
   46149 #define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
   46150   uint8x8_t __s0 = __p0; \
   46151   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46152   uint8_t __ret; \
   46153   __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
   46154   __ret; \
   46155 })
   46156 #endif
   46157 
   46158 #ifdef __LITTLE_ENDIAN__
   46159 #define vdups_lane_u32(__p0, __p1) __extension__ ({ \
   46160   uint32x2_t __s0 = __p0; \
   46161   uint32_t __ret; \
   46162   __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
   46163   __ret; \
   46164 })
   46165 #else
   46166 #define vdups_lane_u32(__p0, __p1) __extension__ ({ \
   46167   uint32x2_t __s0 = __p0; \
   46168   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46169   uint32_t __ret; \
   46170   __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
   46171   __ret; \
   46172 })
   46173 #endif
   46174 
   46175 #ifdef __LITTLE_ENDIAN__
   46176 #define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
   46177   uint64x1_t __s0 = __p0; \
   46178   uint64_t __ret; \
   46179   __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
   46180   __ret; \
   46181 })
   46182 #else
   46183 #define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
   46184   uint64x1_t __s0 = __p0; \
   46185   uint64_t __ret; \
   46186   __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
   46187   __ret; \
   46188 })
   46189 #endif
   46190 
   46191 #ifdef __LITTLE_ENDIAN__
   46192 #define vduph_lane_u16(__p0, __p1) __extension__ ({ \
   46193   uint16x4_t __s0 = __p0; \
   46194   uint16_t __ret; \
   46195   __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
   46196   __ret; \
   46197 })
   46198 #else
   46199 #define vduph_lane_u16(__p0, __p1) __extension__ ({ \
   46200   uint16x4_t __s0 = __p0; \
   46201   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46202   uint16_t __ret; \
   46203   __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
   46204   __ret; \
   46205 })
   46206 #endif
   46207 
   46208 #ifdef __LITTLE_ENDIAN__
   46209 #define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
   46210   int8x8_t __s0 = __p0; \
   46211   int8_t __ret; \
   46212   __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
   46213   __ret; \
   46214 })
   46215 #else
   46216 #define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
   46217   int8x8_t __s0 = __p0; \
   46218   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46219   int8_t __ret; \
   46220   __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
   46221   __ret; \
   46222 })
   46223 #endif
   46224 
   46225 #ifdef __LITTLE_ENDIAN__
   46226 #define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
   46227   float64x1_t __s0 = __p0; \
   46228   float64_t __ret; \
   46229   __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
   46230   __ret; \
   46231 })
   46232 #else
   46233 #define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
   46234   float64x1_t __s0 = __p0; \
   46235   float64_t __ret; \
   46236   __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
   46237   __ret; \
   46238 })
   46239 #endif
   46240 
   46241 #ifdef __LITTLE_ENDIAN__
   46242 #define vdups_lane_f32(__p0, __p1) __extension__ ({ \
   46243   float32x2_t __s0 = __p0; \
   46244   float32_t __ret; \
   46245   __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__s0, __p1); \
   46246   __ret; \
   46247 })
   46248 #else
   46249 #define vdups_lane_f32(__p0, __p1) __extension__ ({ \
   46250   float32x2_t __s0 = __p0; \
   46251   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46252   float32_t __ret; \
   46253   __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__rev0, __p1); \
   46254   __ret; \
   46255 })
   46256 #endif
   46257 
   46258 #ifdef __LITTLE_ENDIAN__
   46259 #define vdups_lane_s32(__p0, __p1) __extension__ ({ \
   46260   int32x2_t __s0 = __p0; \
   46261   int32_t __ret; \
   46262   __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
   46263   __ret; \
   46264 })
   46265 #else
   46266 #define vdups_lane_s32(__p0, __p1) __extension__ ({ \
   46267   int32x2_t __s0 = __p0; \
   46268   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46269   int32_t __ret; \
   46270   __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
   46271   __ret; \
   46272 })
   46273 #endif
   46274 
   46275 #ifdef __LITTLE_ENDIAN__
   46276 #define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
   46277   int64x1_t __s0 = __p0; \
   46278   int64_t __ret; \
   46279   __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
   46280   __ret; \
   46281 })
   46282 #else
   46283 #define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
   46284   int64x1_t __s0 = __p0; \
   46285   int64_t __ret; \
   46286   __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
   46287   __ret; \
   46288 })
   46289 #endif
   46290 
   46291 #ifdef __LITTLE_ENDIAN__
   46292 #define vduph_lane_s16(__p0, __p1) __extension__ ({ \
   46293   int16x4_t __s0 = __p0; \
   46294   int16_t __ret; \
   46295   __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
   46296   __ret; \
   46297 })
   46298 #else
   46299 #define vduph_lane_s16(__p0, __p1) __extension__ ({ \
   46300   int16x4_t __s0 = __p0; \
   46301   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46302   int16_t __ret; \
   46303   __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
   46304   __ret; \
   46305 })
   46306 #endif
   46307 
   46308 #ifdef __LITTLE_ENDIAN__
   46309 #define vdup_lane_p64(__p0, __p1) __extension__ ({ \
   46310   poly64x1_t __s0 = __p0; \
   46311   poly64x1_t __ret; \
   46312   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   46313   __ret; \
   46314 })
   46315 #else
   46316 #define vdup_lane_p64(__p0, __p1) __extension__ ({ \
   46317   poly64x1_t __s0 = __p0; \
   46318   poly64x1_t __ret; \
   46319   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   46320   __ret; \
   46321 })
   46322 #endif
   46323 
   46324 #ifdef __LITTLE_ENDIAN__
   46325 #define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
   46326   poly64x1_t __s0 = __p0; \
   46327   poly64x2_t __ret; \
   46328   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   46329   __ret; \
   46330 })
   46331 #else
   46332 #define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
   46333   poly64x1_t __s0 = __p0; \
   46334   poly64x2_t __ret; \
   46335   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   46336   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   46337   __ret; \
   46338 })
   46339 #endif
   46340 
   46341 #ifdef __LITTLE_ENDIAN__
   46342 #define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
   46343   float64x1_t __s0 = __p0; \
   46344   float64x2_t __ret; \
   46345   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   46346   __ret; \
   46347 })
   46348 #else
   46349 #define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
   46350   float64x1_t __s0 = __p0; \
   46351   float64x2_t __ret; \
   46352   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   46353   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   46354   __ret; \
   46355 })
   46356 #endif
   46357 
   46358 #ifdef __LITTLE_ENDIAN__
   46359 #define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
   46360   float16x4_t __s0 = __p0; \
   46361   float16x8_t __ret; \
   46362   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46363   __ret; \
   46364 })
   46365 #else
   46366 #define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
   46367   float16x4_t __s0 = __p0; \
   46368   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46369   float16x8_t __ret; \
   46370   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46371   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   46372   __ret; \
   46373 })
   46374 #endif
   46375 
   46376 #ifdef __LITTLE_ENDIAN__
   46377 #define vdup_lane_f64(__p0, __p1) __extension__ ({ \
   46378   float64x1_t __s0 = __p0; \
   46379   float64x1_t __ret; \
   46380   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   46381   __ret; \
   46382 })
   46383 #else
   46384 #define vdup_lane_f64(__p0, __p1) __extension__ ({ \
   46385   float64x1_t __s0 = __p0; \
   46386   float64x1_t __ret; \
   46387   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   46388   __ret; \
   46389 })
   46390 #endif
   46391 
   46392 #ifdef __LITTLE_ENDIAN__
   46393 #define vdup_lane_f16(__p0, __p1) __extension__ ({ \
   46394   float16x4_t __s0 = __p0; \
   46395   float16x4_t __ret; \
   46396   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   46397   __ret; \
   46398 })
   46399 #else
   46400 #define vdup_lane_f16(__p0, __p1) __extension__ ({ \
   46401   float16x4_t __s0 = __p0; \
   46402   float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46403   float16x4_t __ret; \
   46404   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   46405   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   46406   __ret; \
   46407 })
   46408 #endif
   46409 
   46410 #ifdef __LITTLE_ENDIAN__
   46411 #define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
   46412   poly8x16_t __s0 = __p0; \
   46413   poly8_t __ret; \
   46414   __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
   46415   __ret; \
   46416 })
   46417 #else
   46418 #define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
   46419   poly8x16_t __s0 = __p0; \
   46420   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46421   poly8_t __ret; \
   46422   __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
   46423   __ret; \
   46424 })
   46425 #endif
   46426 
   46427 #ifdef __LITTLE_ENDIAN__
   46428 #define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
   46429   poly16x8_t __s0 = __p0; \
   46430   poly16_t __ret; \
   46431   __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
   46432   __ret; \
   46433 })
   46434 #else
   46435 #define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
   46436   poly16x8_t __s0 = __p0; \
   46437   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46438   poly16_t __ret; \
   46439   __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
   46440   __ret; \
   46441 })
   46442 #endif
   46443 
   46444 #ifdef __LITTLE_ENDIAN__
   46445 #define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
   46446   uint8x16_t __s0 = __p0; \
   46447   uint8_t __ret; \
   46448   __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
   46449   __ret; \
   46450 })
   46451 #else
   46452 #define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
   46453   uint8x16_t __s0 = __p0; \
   46454   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46455   uint8_t __ret; \
   46456   __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
   46457   __ret; \
   46458 })
   46459 #endif
   46460 
   46461 #ifdef __LITTLE_ENDIAN__
   46462 #define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
   46463   uint32x4_t __s0 = __p0; \
   46464   uint32_t __ret; \
   46465   __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
   46466   __ret; \
   46467 })
   46468 #else
   46469 #define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
   46470   uint32x4_t __s0 = __p0; \
   46471   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46472   uint32_t __ret; \
   46473   __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
   46474   __ret; \
   46475 })
   46476 #endif
   46477 
   46478 #ifdef __LITTLE_ENDIAN__
   46479 #define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
   46480   uint64x2_t __s0 = __p0; \
   46481   uint64_t __ret; \
   46482   __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
   46483   __ret; \
   46484 })
   46485 #else
   46486 #define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
   46487   uint64x2_t __s0 = __p0; \
   46488   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46489   uint64_t __ret; \
   46490   __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
   46491   __ret; \
   46492 })
   46493 #endif
   46494 
   46495 #ifdef __LITTLE_ENDIAN__
   46496 #define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
   46497   uint16x8_t __s0 = __p0; \
   46498   uint16_t __ret; \
   46499   __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
   46500   __ret; \
   46501 })
   46502 #else
   46503 #define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
   46504   uint16x8_t __s0 = __p0; \
   46505   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46506   uint16_t __ret; \
   46507   __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
   46508   __ret; \
   46509 })
   46510 #endif
   46511 
   46512 #ifdef __LITTLE_ENDIAN__
   46513 #define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
   46514   int8x16_t __s0 = __p0; \
   46515   int8_t __ret; \
   46516   __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
   46517   __ret; \
   46518 })
   46519 #else
   46520 #define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
   46521   int8x16_t __s0 = __p0; \
   46522   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46523   int8_t __ret; \
   46524   __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
   46525   __ret; \
   46526 })
   46527 #endif
   46528 
   46529 #ifdef __LITTLE_ENDIAN__
   46530 #define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
   46531   float64x2_t __s0 = __p0; \
   46532   float64_t __ret; \
   46533   __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__s0, __p1); \
   46534   __ret; \
   46535 })
   46536 #else
   46537 #define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
   46538   float64x2_t __s0 = __p0; \
   46539   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46540   float64_t __ret; \
   46541   __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__rev0, __p1); \
   46542   __ret; \
   46543 })
   46544 #endif
   46545 
   46546 #ifdef __LITTLE_ENDIAN__
   46547 #define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
   46548   float32x4_t __s0 = __p0; \
   46549   float32_t __ret; \
   46550   __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__s0, __p1); \
   46551   __ret; \
   46552 })
   46553 #else
   46554 #define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
   46555   float32x4_t __s0 = __p0; \
   46556   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46557   float32_t __ret; \
   46558   __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__rev0, __p1); \
   46559   __ret; \
   46560 })
   46561 #endif
   46562 
   46563 #ifdef __LITTLE_ENDIAN__
   46564 #define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
   46565   int32x4_t __s0 = __p0; \
   46566   int32_t __ret; \
   46567   __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
   46568   __ret; \
   46569 })
   46570 #else
   46571 #define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
   46572   int32x4_t __s0 = __p0; \
   46573   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46574   int32_t __ret; \
   46575   __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
   46576   __ret; \
   46577 })
   46578 #endif
   46579 
   46580 #ifdef __LITTLE_ENDIAN__
   46581 #define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
   46582   int64x2_t __s0 = __p0; \
   46583   int64_t __ret; \
   46584   __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
   46585   __ret; \
   46586 })
   46587 #else
   46588 #define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
   46589   int64x2_t __s0 = __p0; \
   46590   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46591   int64_t __ret; \
   46592   __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
   46593   __ret; \
   46594 })
   46595 #endif
   46596 
   46597 #ifdef __LITTLE_ENDIAN__
   46598 #define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
   46599   int16x8_t __s0 = __p0; \
   46600   int16_t __ret; \
   46601   __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
   46602   __ret; \
   46603 })
   46604 #else
   46605 #define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
   46606   int16x8_t __s0 = __p0; \
   46607   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46608   int16_t __ret; \
   46609   __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
   46610   __ret; \
   46611 })
   46612 #endif
   46613 
   46614 #ifdef __LITTLE_ENDIAN__
   46615 #define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
   46616   poly8x16_t __s0 = __p0; \
   46617   poly8x8_t __ret; \
   46618   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46619   __ret; \
   46620 })
   46621 #else
   46622 #define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
   46623   poly8x16_t __s0 = __p0; \
   46624   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46625   poly8x8_t __ret; \
   46626   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46627   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   46628   __ret; \
   46629 })
   46630 #endif
   46631 
   46632 #ifdef __LITTLE_ENDIAN__
   46633 #define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
   46634   poly64x2_t __s0 = __p0; \
   46635   poly64x1_t __ret; \
   46636   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   46637   __ret; \
   46638 })
   46639 #else
   46640 #define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
   46641   poly64x2_t __s0 = __p0; \
   46642   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46643   poly64x1_t __ret; \
   46644   __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
   46645   __ret; \
   46646 })
   46647 #endif
   46648 
   46649 #ifdef __LITTLE_ENDIAN__
   46650 #define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
   46651   poly16x8_t __s0 = __p0; \
   46652   poly16x4_t __ret; \
   46653   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   46654   __ret; \
   46655 })
   46656 #else
   46657 #define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
   46658   poly16x8_t __s0 = __p0; \
   46659   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46660   poly16x4_t __ret; \
   46661   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   46662   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   46663   __ret; \
   46664 })
   46665 #endif
   46666 
   46667 #ifdef __LITTLE_ENDIAN__
   46668 #define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
   46669   poly8x16_t __s0 = __p0; \
   46670   poly8x16_t __ret; \
   46671   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46672   __ret; \
   46673 })
   46674 #else
   46675 #define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
   46676   poly8x16_t __s0 = __p0; \
   46677   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46678   poly8x16_t __ret; \
   46679   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46680   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46681   __ret; \
   46682 })
   46683 #endif
   46684 
   46685 #ifdef __LITTLE_ENDIAN__
   46686 #define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
   46687   poly64x2_t __s0 = __p0; \
   46688   poly64x2_t __ret; \
   46689   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   46690   __ret; \
   46691 })
   46692 #else
   46693 #define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
   46694   poly64x2_t __s0 = __p0; \
   46695   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46696   poly64x2_t __ret; \
   46697   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   46698   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   46699   __ret; \
   46700 })
   46701 #endif
   46702 
   46703 #ifdef __LITTLE_ENDIAN__
   46704 #define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
   46705   poly16x8_t __s0 = __p0; \
   46706   poly16x8_t __ret; \
   46707   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46708   __ret; \
   46709 })
   46710 #else
   46711 #define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
   46712   poly16x8_t __s0 = __p0; \
   46713   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46714   poly16x8_t __ret; \
   46715   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46716   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   46717   __ret; \
   46718 })
   46719 #endif
   46720 
   46721 #ifdef __LITTLE_ENDIAN__
   46722 #define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
   46723   uint8x16_t __s0 = __p0; \
   46724   uint8x16_t __ret; \
   46725   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46726   __ret; \
   46727 })
   46728 #else
   46729 #define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
   46730   uint8x16_t __s0 = __p0; \
   46731   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46732   uint8x16_t __ret; \
   46733   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46734   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46735   __ret; \
   46736 })
   46737 #endif
   46738 
   46739 #ifdef __LITTLE_ENDIAN__
   46740 #define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
   46741   uint32x4_t __s0 = __p0; \
   46742   uint32x4_t __ret; \
   46743   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   46744   __ret; \
   46745 })
   46746 #else
   46747 #define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
   46748   uint32x4_t __s0 = __p0; \
   46749   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46750   uint32x4_t __ret; \
   46751   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   46752   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   46753   __ret; \
   46754 })
   46755 #endif
   46756 
   46757 #ifdef __LITTLE_ENDIAN__
   46758 #define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
   46759   uint64x2_t __s0 = __p0; \
   46760   uint64x2_t __ret; \
   46761   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   46762   __ret; \
   46763 })
   46764 #else
   46765 #define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
   46766   uint64x2_t __s0 = __p0; \
   46767   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46768   uint64x2_t __ret; \
   46769   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   46770   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   46771   __ret; \
   46772 })
   46773 #endif
   46774 
   46775 #ifdef __LITTLE_ENDIAN__
   46776 #define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
   46777   uint16x8_t __s0 = __p0; \
   46778   uint16x8_t __ret; \
   46779   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46780   __ret; \
   46781 })
   46782 #else
   46783 #define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
   46784   uint16x8_t __s0 = __p0; \
   46785   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46786   uint16x8_t __ret; \
   46787   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46788   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   46789   __ret; \
   46790 })
   46791 #endif
   46792 
   46793 #ifdef __LITTLE_ENDIAN__
   46794 #define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
   46795   int8x16_t __s0 = __p0; \
   46796   int8x16_t __ret; \
   46797   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46798   __ret; \
   46799 })
   46800 #else
   46801 #define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
   46802   int8x16_t __s0 = __p0; \
   46803   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46804   int8x16_t __ret; \
   46805   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46806   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46807   __ret; \
   46808 })
   46809 #endif
   46810 
   46811 #ifdef __LITTLE_ENDIAN__
   46812 #define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
   46813   float64x2_t __s0 = __p0; \
   46814   float64x2_t __ret; \
   46815   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   46816   __ret; \
   46817 })
   46818 #else
   46819 #define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
   46820   float64x2_t __s0 = __p0; \
   46821   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46822   float64x2_t __ret; \
   46823   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   46824   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   46825   __ret; \
   46826 })
   46827 #endif
   46828 
   46829 #ifdef __LITTLE_ENDIAN__
   46830 #define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
   46831   float32x4_t __s0 = __p0; \
   46832   float32x4_t __ret; \
   46833   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   46834   __ret; \
   46835 })
   46836 #else
   46837 #define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
   46838   float32x4_t __s0 = __p0; \
   46839   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46840   float32x4_t __ret; \
   46841   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   46842   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   46843   __ret; \
   46844 })
   46845 #endif
   46846 
   46847 #ifdef __LITTLE_ENDIAN__
   46848 #define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
   46849   float16x8_t __s0 = __p0; \
   46850   float16x8_t __ret; \
   46851   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46852   __ret; \
   46853 })
   46854 #else
   46855 #define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
   46856   float16x8_t __s0 = __p0; \
   46857   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46858   float16x8_t __ret; \
   46859   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46860   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   46861   __ret; \
   46862 })
   46863 #endif
   46864 
   46865 #ifdef __LITTLE_ENDIAN__
   46866 #define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
   46867   int32x4_t __s0 = __p0; \
   46868   int32x4_t __ret; \
   46869   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   46870   __ret; \
   46871 })
   46872 #else
   46873 #define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
   46874   int32x4_t __s0 = __p0; \
   46875   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46876   int32x4_t __ret; \
   46877   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   46878   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   46879   __ret; \
   46880 })
   46881 #endif
   46882 
   46883 #ifdef __LITTLE_ENDIAN__
   46884 #define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
   46885   int64x2_t __s0 = __p0; \
   46886   int64x2_t __ret; \
   46887   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   46888   __ret; \
   46889 })
   46890 #else
   46891 #define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
   46892   int64x2_t __s0 = __p0; \
   46893   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46894   int64x2_t __ret; \
   46895   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   46896   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   46897   __ret; \
   46898 })
   46899 #endif
   46900 
   46901 #ifdef __LITTLE_ENDIAN__
   46902 #define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
   46903   int16x8_t __s0 = __p0; \
   46904   int16x8_t __ret; \
   46905   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46906   __ret; \
   46907 })
   46908 #else
   46909 #define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
   46910   int16x8_t __s0 = __p0; \
   46911   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46912   int16x8_t __ret; \
   46913   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46914   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   46915   __ret; \
   46916 })
   46917 #endif
   46918 
   46919 #ifdef __LITTLE_ENDIAN__
   46920 #define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
   46921   uint8x16_t __s0 = __p0; \
   46922   uint8x8_t __ret; \
   46923   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46924   __ret; \
   46925 })
   46926 #else
   46927 #define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
   46928   uint8x16_t __s0 = __p0; \
   46929   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   46930   uint8x8_t __ret; \
   46931   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46932   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   46933   __ret; \
   46934 })
   46935 #endif
   46936 
   46937 #ifdef __LITTLE_ENDIAN__
   46938 #define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
   46939   uint32x4_t __s0 = __p0; \
   46940   uint32x2_t __ret; \
   46941   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   46942   __ret; \
   46943 })
   46944 #else
   46945 #define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
   46946   uint32x4_t __s0 = __p0; \
   46947   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   46948   uint32x2_t __ret; \
   46949   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   46950   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   46951   __ret; \
   46952 })
   46953 #endif
   46954 
   46955 #ifdef __LITTLE_ENDIAN__
   46956 #define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
   46957   uint64x2_t __s0 = __p0; \
   46958   uint64x1_t __ret; \
   46959   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   46960   __ret; \
   46961 })
   46962 #else
   46963 #define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
   46964   uint64x2_t __s0 = __p0; \
   46965   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   46966   uint64x1_t __ret; \
   46967   __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
   46968   __ret; \
   46969 })
   46970 #endif
   46971 
   46972 #ifdef __LITTLE_ENDIAN__
   46973 #define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
   46974   uint16x8_t __s0 = __p0; \
   46975   uint16x4_t __ret; \
   46976   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   46977   __ret; \
   46978 })
   46979 #else
   46980 #define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
   46981   uint16x8_t __s0 = __p0; \
   46982   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   46983   uint16x4_t __ret; \
   46984   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   46985   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   46986   __ret; \
   46987 })
   46988 #endif
   46989 
   46990 #ifdef __LITTLE_ENDIAN__
   46991 #define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
   46992   int8x16_t __s0 = __p0; \
   46993   int8x8_t __ret; \
   46994   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   46995   __ret; \
   46996 })
   46997 #else
   46998 #define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
   46999   int8x16_t __s0 = __p0; \
   47000   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   47001   int8x8_t __ret; \
   47002   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
   47003   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   47004   __ret; \
   47005 })
   47006 #endif
   47007 
   47008 #ifdef __LITTLE_ENDIAN__
   47009 #define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
   47010   float64x2_t __s0 = __p0; \
   47011   float64x1_t __ret; \
   47012   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   47013   __ret; \
   47014 })
   47015 #else
   47016 #define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
   47017   float64x2_t __s0 = __p0; \
   47018   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   47019   float64x1_t __ret; \
   47020   __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
   47021   __ret; \
   47022 })
   47023 #endif
   47024 
   47025 #ifdef __LITTLE_ENDIAN__
   47026 #define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
   47027   float32x4_t __s0 = __p0; \
   47028   float32x2_t __ret; \
   47029   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   47030   __ret; \
   47031 })
   47032 #else
   47033 #define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
   47034   float32x4_t __s0 = __p0; \
   47035   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   47036   float32x2_t __ret; \
   47037   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   47038   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   47039   __ret; \
   47040 })
   47041 #endif
   47042 
   47043 #ifdef __LITTLE_ENDIAN__
   47044 #define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
   47045   float16x8_t __s0 = __p0; \
   47046   float16x4_t __ret; \
   47047   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   47048   __ret; \
   47049 })
   47050 #else
   47051 #define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
   47052   float16x8_t __s0 = __p0; \
   47053   float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   47054   float16x4_t __ret; \
   47055   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   47056   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   47057   __ret; \
   47058 })
   47059 #endif
   47060 
   47061 #ifdef __LITTLE_ENDIAN__
   47062 #define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
   47063   int32x4_t __s0 = __p0; \
   47064   int32x2_t __ret; \
   47065   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
   47066   __ret; \
   47067 })
   47068 #else
   47069 #define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
   47070   int32x4_t __s0 = __p0; \
   47071   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   47072   int32x2_t __ret; \
   47073   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
   47074   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   47075   __ret; \
   47076 })
   47077 #endif
   47078 
   47079 #ifdef __LITTLE_ENDIAN__
   47080 #define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
   47081   int64x2_t __s0 = __p0; \
   47082   int64x1_t __ret; \
   47083   __ret = __builtin_shufflevector(__s0, __s0, __p1); \
   47084   __ret; \
   47085 })
   47086 #else
   47087 #define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
   47088   int64x2_t __s0 = __p0; \
   47089   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   47090   int64x1_t __ret; \
   47091   __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
   47092   __ret; \
   47093 })
   47094 #endif
   47095 
   47096 #ifdef __LITTLE_ENDIAN__
   47097 #define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
   47098   int16x8_t __s0 = __p0; \
   47099   int16x4_t __ret; \
   47100   __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
   47101   __ret; \
   47102 })
   47103 #else
   47104 #define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
   47105   int16x8_t __s0 = __p0; \
   47106   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   47107   int16x4_t __ret; \
   47108   __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
   47109   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   47110   __ret; \
   47111 })
   47112 #endif
   47113 
   47114 #ifdef __LITTLE_ENDIAN__
   47115 __ai poly64x1_t vdup_n_p64(poly64_t __p0) {
   47116   poly64x1_t __ret;
   47117   __ret = (poly64x1_t) {__p0};
   47118   return __ret;
   47119 }
   47120 #else
   47121 __ai poly64x1_t vdup_n_p64(poly64_t __p0) {
   47122   poly64x1_t __ret;
   47123   __ret = (poly64x1_t) {__p0};
   47124   return __ret;
   47125 }
   47126 #endif
   47127 
   47128 #ifdef __LITTLE_ENDIAN__
   47129 __ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
   47130   poly64x2_t __ret;
   47131   __ret = (poly64x2_t) {__p0, __p0};
   47132   return __ret;
   47133 }
   47134 #else
   47135 __ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
   47136   poly64x2_t __ret;
   47137   __ret = (poly64x2_t) {__p0, __p0};
   47138   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   47139   return __ret;
   47140 }
   47141 #endif
   47142 
   47143 #ifdef __LITTLE_ENDIAN__
   47144 __ai float64x2_t vdupq_n_f64(float64_t __p0) {
   47145   float64x2_t __ret;
   47146   __ret = (float64x2_t) {__p0, __p0};
   47147   return __ret;
   47148 }
   47149 #else
   47150 __ai float64x2_t vdupq_n_f64(float64_t __p0) {
   47151   float64x2_t __ret;
   47152   __ret = (float64x2_t) {__p0, __p0};
   47153   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   47154   return __ret;
   47155 }
   47156 #endif
   47157 
   47158 #ifdef __LITTLE_ENDIAN__
   47159 __ai float64x1_t vdup_n_f64(float64_t __p0) {
   47160   float64x1_t __ret;
   47161   __ret = (float64x1_t) {__p0};
   47162   return __ret;
   47163 }
   47164 #else
   47165 __ai float64x1_t vdup_n_f64(float64_t __p0) {
   47166   float64x1_t __ret;
   47167   __ret = (float64x1_t) {__p0};
   47168   return __ret;
   47169 }
   47170 #endif
   47171 
   47172 #ifdef __LITTLE_ENDIAN__
   47173 #define vext_p64(__p0, __p1, __p2) __extension__ ({ \
   47174   poly64x1_t __s0 = __p0; \
   47175   poly64x1_t __s1 = __p1; \
   47176   poly64x1_t __ret; \
   47177   __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
   47178   __ret; \
   47179 })
   47180 #else
   47181 #define vext_p64(__p0, __p1, __p2) __extension__ ({ \
   47182   poly64x1_t __s0 = __p0; \
   47183   poly64x1_t __s1 = __p1; \
   47184   poly64x1_t __ret; \
   47185   __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
   47186   __ret; \
   47187 })
   47188 #endif
   47189 
   47190 #ifdef __LITTLE_ENDIAN__
   47191 #define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
   47192   poly64x2_t __s0 = __p0; \
   47193   poly64x2_t __s1 = __p1; \
   47194   poly64x2_t __ret; \
   47195   __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
   47196   __ret; \
   47197 })
   47198 #else
   47199 #define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
   47200   poly64x2_t __s0 = __p0; \
   47201   poly64x2_t __s1 = __p1; \
   47202   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   47203   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   47204   poly64x2_t __ret; \
   47205   __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
   47206   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   47207   __ret; \
   47208 })
   47209 #endif
   47210 
   47211 #ifdef __LITTLE_ENDIAN__
   47212 #define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
   47213   float64x2_t __s0 = __p0; \
   47214   float64x2_t __s1 = __p1; \
   47215   float64x2_t __ret; \
   47216   __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 42); \
   47217   __ret; \
   47218 })
   47219 #else
   47220 #define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
   47221   float64x2_t __s0 = __p0; \
   47222   float64x2_t __s1 = __p1; \
   47223   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   47224   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   47225   float64x2_t __ret; \
   47226   __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 42); \
   47227   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   47228   __ret; \
   47229 })
   47230 #endif
   47231 
   47232 #ifdef __LITTLE_ENDIAN__
   47233 #define vext_f64(__p0, __p1, __p2) __extension__ ({ \
   47234   float64x1_t __s0 = __p0; \
   47235   float64x1_t __s1 = __p1; \
   47236   float64x1_t __ret; \
   47237   __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
   47238   __ret; \
   47239 })
   47240 #else
   47241 #define vext_f64(__p0, __p1, __p2) __extension__ ({ \
   47242   float64x1_t __s0 = __p0; \
   47243   float64x1_t __s1 = __p1; \
   47244   float64x1_t __ret; \
   47245   __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
   47246   __ret; \
   47247 })
   47248 #endif
   47249 
   47250 #ifdef __LITTLE_ENDIAN__
   47251 __ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   47252   float64x2_t __ret;
   47253   __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
   47254   return __ret;
   47255 }
   47256 #else
   47257 __ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   47258   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   47259   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   47260   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   47261   float64x2_t __ret;
   47262   __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
   47263   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   47264   return __ret;
   47265 }
   47266 __ai float64x2_t __noswap_vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   47267   float64x2_t __ret;
   47268   __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
   47269   return __ret;
   47270 }
   47271 #endif
   47272 
   47273 #ifdef __LITTLE_ENDIAN__
   47274 __ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   47275   float64x1_t __ret;
   47276   __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
   47277   return __ret;
   47278 }
   47279 #else
   47280 __ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   47281   float64x1_t __ret;
   47282   __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
   47283   return __ret;
   47284 }
   47285 __ai float64x1_t __noswap_vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   47286   float64x1_t __ret;
   47287   __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
   47288   return __ret;
   47289 }
   47290 #endif
   47291 
   47292 #ifdef __LITTLE_ENDIAN__
   47293 #define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47294   float64_t __s0 = __p0; \
   47295   float64_t __s1 = __p1; \
   47296   float64x1_t __s2 = __p2; \
   47297   float64_t __ret; \
   47298   __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
   47299   __ret; \
   47300 })
   47301 #else
   47302 #define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47303   float64_t __s0 = __p0; \
   47304   float64_t __s1 = __p1; \
   47305   float64x1_t __s2 = __p2; \
   47306   float64_t __ret; \
   47307   __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
   47308   __ret; \
   47309 })
   47310 #define __noswap_vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47311   float64_t __s0 = __p0; \
   47312   float64_t __s1 = __p1; \
   47313   float64x1_t __s2 = __p2; \
   47314   float64_t __ret; \
   47315   __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
   47316   __ret; \
   47317 })
   47318 #endif
   47319 
   47320 #ifdef __LITTLE_ENDIAN__
   47321 #define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47322   float32_t __s0 = __p0; \
   47323   float32_t __s1 = __p1; \
   47324   float32x2_t __s2 = __p2; \
   47325   float32_t __ret; \
   47326   __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
   47327   __ret; \
   47328 })
   47329 #else
   47330 #define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47331   float32_t __s0 = __p0; \
   47332   float32_t __s1 = __p1; \
   47333   float32x2_t __s2 = __p2; \
   47334   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   47335   float32_t __ret; \
   47336   __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__rev2, __p3); \
   47337   __ret; \
   47338 })
   47339 #define __noswap_vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47340   float32_t __s0 = __p0; \
   47341   float32_t __s1 = __p1; \
   47342   float32x2_t __s2 = __p2; \
   47343   float32_t __ret; \
   47344   __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
   47345   __ret; \
   47346 })
   47347 #endif
   47348 
   47349 #ifdef __LITTLE_ENDIAN__
   47350 #define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47351   float64x2_t __s0 = __p0; \
   47352   float64x2_t __s1 = __p1; \
   47353   float64x1_t __s2 = __p2; \
   47354   float64x2_t __ret; \
   47355   __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
   47356   __ret; \
   47357 })
   47358 #else
   47359 #define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47360   float64x2_t __s0 = __p0; \
   47361   float64x2_t __s1 = __p1; \
   47362   float64x1_t __s2 = __p2; \
   47363   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   47364   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   47365   float64x2_t __ret; \
   47366   __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__s2, __p3, 42); \
   47367   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   47368   __ret; \
   47369 })
   47370 #define __noswap_vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47371   float64x2_t __s0 = __p0; \
   47372   float64x2_t __s1 = __p1; \
   47373   float64x1_t __s2 = __p2; \
   47374   float64x2_t __ret; \
   47375   __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
   47376   __ret; \
   47377 })
   47378 #endif
   47379 
   47380 #ifdef __LITTLE_ENDIAN__
   47381 #define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47382   float32x4_t __s0 = __p0; \
   47383   float32x4_t __s1 = __p1; \
   47384   float32x2_t __s2 = __p2; \
   47385   float32x4_t __ret; \
   47386   __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
   47387   __ret; \
   47388 })
   47389 #else
   47390 #define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47391   float32x4_t __s0 = __p0; \
   47392   float32x4_t __s1 = __p1; \
   47393   float32x2_t __s2 = __p2; \
   47394   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   47395   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   47396   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   47397   float32x4_t __ret; \
   47398   __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, __p3, 41); \
   47399   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   47400   __ret; \
   47401 })
   47402 #define __noswap_vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47403   float32x4_t __s0 = __p0; \
   47404   float32x4_t __s1 = __p1; \
   47405   float32x2_t __s2 = __p2; \
   47406   float32x4_t __ret; \
   47407   __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
   47408   __ret; \
   47409 })
   47410 #endif
   47411 
   47412 #ifdef __LITTLE_ENDIAN__
   47413 #define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47414   float64x1_t __s0 = __p0; \
   47415   float64x1_t __s1 = __p1; \
   47416   float64x1_t __s2 = __p2; \
   47417   float64x1_t __ret; \
   47418   __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
   47419   __ret; \
   47420 })
   47421 #else
   47422 #define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47423   float64x1_t __s0 = __p0; \
   47424   float64x1_t __s1 = __p1; \
   47425   float64x1_t __s2 = __p2; \
   47426   float64x1_t __ret; \
   47427   __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
   47428   __ret; \
   47429 })
   47430 #define __noswap_vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47431   float64x1_t __s0 = __p0; \
   47432   float64x1_t __s1 = __p1; \
   47433   float64x1_t __s2 = __p2; \
   47434   float64x1_t __ret; \
   47435   __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
   47436   __ret; \
   47437 })
   47438 #endif
   47439 
   47440 #ifdef __LITTLE_ENDIAN__
   47441 #define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47442   float32x2_t __s0 = __p0; \
   47443   float32x2_t __s1 = __p1; \
   47444   float32x2_t __s2 = __p2; \
   47445   float32x2_t __ret; \
   47446   __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
   47447   __ret; \
   47448 })
   47449 #else
   47450 #define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47451   float32x2_t __s0 = __p0; \
   47452   float32x2_t __s1 = __p1; \
   47453   float32x2_t __s2 = __p2; \
   47454   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   47455   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   47456   float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   47457   float32x2_t __ret; \
   47458   __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, __p3, 9); \
   47459   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   47460   __ret; \
   47461 })
   47462 #define __noswap_vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47463   float32x2_t __s0 = __p0; \
   47464   float32x2_t __s1 = __p1; \
   47465   float32x2_t __s2 = __p2; \
   47466   float32x2_t __ret; \
   47467   __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
   47468   __ret; \
   47469 })
   47470 #endif
   47471 
   47472 #ifdef __LITTLE_ENDIAN__
   47473 #define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47474   float64_t __s0 = __p0; \
   47475   float64_t __s1 = __p1; \
   47476   float64x2_t __s2 = __p2; \
   47477   float64_t __ret; \
   47478   __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
   47479   __ret; \
   47480 })
   47481 #else
   47482 #define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47483   float64_t __s0 = __p0; \
   47484   float64_t __s1 = __p1; \
   47485   float64x2_t __s2 = __p2; \
   47486   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   47487   float64_t __ret; \
   47488   __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__rev2, __p3); \
   47489   __ret; \
   47490 })
   47491 #define __noswap_vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47492   float64_t __s0 = __p0; \
   47493   float64_t __s1 = __p1; \
   47494   float64x2_t __s2 = __p2; \
   47495   float64_t __ret; \
   47496   __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
   47497   __ret; \
   47498 })
   47499 #endif
   47500 
   47501 #ifdef __LITTLE_ENDIAN__
   47502 #define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47503   float32_t __s0 = __p0; \
   47504   float32_t __s1 = __p1; \
   47505   float32x4_t __s2 = __p2; \
   47506   float32_t __ret; \
   47507   __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
   47508   __ret; \
   47509 })
   47510 #else
   47511 #define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47512   float32_t __s0 = __p0; \
   47513   float32_t __s1 = __p1; \
   47514   float32x4_t __s2 = __p2; \
   47515   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   47516   float32_t __ret; \
   47517   __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__rev2, __p3); \
   47518   __ret; \
   47519 })
   47520 #define __noswap_vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47521   float32_t __s0 = __p0; \
   47522   float32_t __s1 = __p1; \
   47523   float32x4_t __s2 = __p2; \
   47524   float32_t __ret; \
   47525   __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
   47526   __ret; \
   47527 })
   47528 #endif
   47529 
   47530 #ifdef __LITTLE_ENDIAN__
   47531 #define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47532   float64x2_t __s0 = __p0; \
   47533   float64x2_t __s1 = __p1; \
   47534   float64x2_t __s2 = __p2; \
   47535   float64x2_t __ret; \
   47536   __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
   47537   __ret; \
   47538 })
   47539 #else
   47540 #define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47541   float64x2_t __s0 = __p0; \
   47542   float64x2_t __s1 = __p1; \
   47543   float64x2_t __s2 = __p2; \
   47544   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   47545   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   47546   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   47547   float64x2_t __ret; \
   47548   __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 42); \
   47549   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   47550   __ret; \
   47551 })
   47552 #define __noswap_vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47553   float64x2_t __s0 = __p0; \
   47554   float64x2_t __s1 = __p1; \
   47555   float64x2_t __s2 = __p2; \
   47556   float64x2_t __ret; \
   47557   __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
   47558   __ret; \
   47559 })
   47560 #endif
   47561 
   47562 #ifdef __LITTLE_ENDIAN__
   47563 #define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47564   float32x4_t __s0 = __p0; \
   47565   float32x4_t __s1 = __p1; \
   47566   float32x4_t __s2 = __p2; \
   47567   float32x4_t __ret; \
   47568   __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
   47569   __ret; \
   47570 })
   47571 #else
   47572 #define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47573   float32x4_t __s0 = __p0; \
   47574   float32x4_t __s1 = __p1; \
   47575   float32x4_t __s2 = __p2; \
   47576   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   47577   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   47578   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   47579   float32x4_t __ret; \
   47580   __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 41); \
   47581   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   47582   __ret; \
   47583 })
   47584 #define __noswap_vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47585   float32x4_t __s0 = __p0; \
   47586   float32x4_t __s1 = __p1; \
   47587   float32x4_t __s2 = __p2; \
   47588   float32x4_t __ret; \
   47589   __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
   47590   __ret; \
   47591 })
   47592 #endif
   47593 
   47594 #ifdef __LITTLE_ENDIAN__
   47595 #define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47596   float64x1_t __s0 = __p0; \
   47597   float64x1_t __s1 = __p1; \
   47598   float64x2_t __s2 = __p2; \
   47599   float64x1_t __ret; \
   47600   __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
   47601   __ret; \
   47602 })
   47603 #else
   47604 #define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47605   float64x1_t __s0 = __p0; \
   47606   float64x1_t __s1 = __p1; \
   47607   float64x2_t __s2 = __p2; \
   47608   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   47609   float64x1_t __ret; \
   47610   __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__rev2, __p3, 10); \
   47611   __ret; \
   47612 })
   47613 #define __noswap_vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
   47614   float64x1_t __s0 = __p0; \
   47615   float64x1_t __s1 = __p1; \
   47616   float64x2_t __s2 = __p2; \
   47617   float64x1_t __ret; \
   47618   __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
   47619   __ret; \
   47620 })
   47621 #endif
   47622 
   47623 #ifdef __LITTLE_ENDIAN__
   47624 #define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47625   float32x2_t __s0 = __p0; \
   47626   float32x2_t __s1 = __p1; \
   47627   float32x4_t __s2 = __p2; \
   47628   float32x2_t __ret; \
   47629   __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
   47630   __ret; \
   47631 })
   47632 #else
   47633 #define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47634   float32x2_t __s0 = __p0; \
   47635   float32x2_t __s1 = __p1; \
   47636   float32x4_t __s2 = __p2; \
   47637   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   47638   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   47639   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   47640   float32x2_t __ret; \
   47641   __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x16_t)__rev2, __p3, 9); \
   47642   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   47643   __ret; \
   47644 })
   47645 #define __noswap_vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   47646   float32x2_t __s0 = __p0; \
   47647   float32x2_t __s1 = __p1; \
   47648   float32x4_t __s2 = __p2; \
   47649   float32x2_t __ret; \
   47650   __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
   47651   __ret; \
   47652 })
   47653 #endif
   47654 
   47655 #ifdef __LITTLE_ENDIAN__
   47656 __ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
   47657   float64x2_t __ret;
   47658   __ret = vfmaq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
   47659   return __ret;
   47660 }
   47661 #else
   47662 __ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
   47663   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   47664   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   47665   float64x2_t __ret;
   47666   __ret = __noswap_vfmaq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
   47667   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   47668   return __ret;
   47669 }
   47670 #endif
   47671 
   47672 #ifdef __LITTLE_ENDIAN__
   47673 __ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
   47674   float32x4_t __ret;
   47675   __ret = vfmaq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
   47676   return __ret;
   47677 }
   47678 #else
   47679 __ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
   47680   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   47681   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   47682   float32x4_t __ret;
   47683   __ret = __noswap_vfmaq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
   47684   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   47685   return __ret;
   47686 }
   47687 #endif
   47688 
   47689 #ifdef __LITTLE_ENDIAN__
   47690 __ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
   47691   float32x2_t __ret;
   47692   __ret = vfma_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
   47693   return __ret;
   47694 }
   47695 #else
   47696 __ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
   47697   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   47698   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   47699   float32x2_t __ret;
   47700   __ret = __noswap_vfma_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
   47701   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   47702   return __ret;
   47703 }
   47704 #endif
   47705 
   47706 #ifdef __LITTLE_ENDIAN__
   47707 __ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   47708   float64x2_t __ret;
   47709   __ret = vfmaq_f64(__p0, -__p1, __p2);
   47710   return __ret;
   47711 }
   47712 #else
   47713 __ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   47714   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   47715   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   47716   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   47717   float64x2_t __ret;
   47718   __ret = __noswap_vfmaq_f64(__rev0, -__rev1, __rev2);
   47719   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   47720   return __ret;
   47721 }
   47722 #endif
   47723 
   47724 #ifdef __LITTLE_ENDIAN__
   47725 __ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   47726   float64x1_t __ret;
   47727   __ret = vfma_f64(__p0, -__p1, __p2);
   47728   return __ret;
   47729 }
   47730 #else
   47731 __ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   47732   float64x1_t __ret;
   47733   __ret = __noswap_vfma_f64(__p0, -__p1, __p2);
   47734   return __ret;
   47735 }
   47736 #endif
   47737 
   47738 #ifdef __LITTLE_ENDIAN__
   47739 #define vfmsd_lane_f64(__p0_88, __p1_88, __p2_88, __p3_88) __extension__ ({ \
   47740   float64_t __s0_88 = __p0_88; \
   47741   float64_t __s1_88 = __p1_88; \
   47742   float64x1_t __s2_88 = __p2_88; \
   47743   float64_t __ret_88; \
   47744   __ret_88 = vfmad_lane_f64(__s0_88, -__s1_88, __s2_88, __p3_88); \
   47745   __ret_88; \
   47746 })
   47747 #else
   47748 #define vfmsd_lane_f64(__p0_89, __p1_89, __p2_89, __p3_89) __extension__ ({ \
   47749   float64_t __s0_89 = __p0_89; \
   47750   float64_t __s1_89 = __p1_89; \
   47751   float64x1_t __s2_89 = __p2_89; \
   47752   float64_t __ret_89; \
   47753   __ret_89 = __noswap_vfmad_lane_f64(__s0_89, -__s1_89, __s2_89, __p3_89); \
   47754   __ret_89; \
   47755 })
   47756 #endif
   47757 
   47758 #ifdef __LITTLE_ENDIAN__
   47759 #define vfmss_lane_f32(__p0_90, __p1_90, __p2_90, __p3_90) __extension__ ({ \
   47760   float32_t __s0_90 = __p0_90; \
   47761   float32_t __s1_90 = __p1_90; \
   47762   float32x2_t __s2_90 = __p2_90; \
   47763   float32_t __ret_90; \
   47764   __ret_90 = vfmas_lane_f32(__s0_90, -__s1_90, __s2_90, __p3_90); \
   47765   __ret_90; \
   47766 })
   47767 #else
   47768 #define vfmss_lane_f32(__p0_91, __p1_91, __p2_91, __p3_91) __extension__ ({ \
   47769   float32_t __s0_91 = __p0_91; \
   47770   float32_t __s1_91 = __p1_91; \
   47771   float32x2_t __s2_91 = __p2_91; \
   47772   float32x2_t __rev2_91;  __rev2_91 = __builtin_shufflevector(__s2_91, __s2_91, 1, 0); \
   47773   float32_t __ret_91; \
   47774   __ret_91 = __noswap_vfmas_lane_f32(__s0_91, -__s1_91, __rev2_91, __p3_91); \
   47775   __ret_91; \
   47776 })
   47777 #endif
   47778 
   47779 #ifdef __LITTLE_ENDIAN__
   47780 #define vfmsq_lane_f64(__p0_92, __p1_92, __p2_92, __p3_92) __extension__ ({ \
   47781   float64x2_t __s0_92 = __p0_92; \
   47782   float64x2_t __s1_92 = __p1_92; \
   47783   float64x1_t __s2_92 = __p2_92; \
   47784   float64x2_t __ret_92; \
   47785   __ret_92 = vfmaq_lane_f64(__s0_92, -__s1_92, __s2_92, __p3_92); \
   47786   __ret_92; \
   47787 })
   47788 #else
   47789 #define vfmsq_lane_f64(__p0_93, __p1_93, __p2_93, __p3_93) __extension__ ({ \
   47790   float64x2_t __s0_93 = __p0_93; \
   47791   float64x2_t __s1_93 = __p1_93; \
   47792   float64x1_t __s2_93 = __p2_93; \
   47793   float64x2_t __rev0_93;  __rev0_93 = __builtin_shufflevector(__s0_93, __s0_93, 1, 0); \
   47794   float64x2_t __rev1_93;  __rev1_93 = __builtin_shufflevector(__s1_93, __s1_93, 1, 0); \
   47795   float64x2_t __ret_93; \
   47796   __ret_93 = __noswap_vfmaq_lane_f64(__rev0_93, -__rev1_93, __s2_93, __p3_93); \
   47797   __ret_93 = __builtin_shufflevector(__ret_93, __ret_93, 1, 0); \
   47798   __ret_93; \
   47799 })
   47800 #endif
   47801 
   47802 #ifdef __LITTLE_ENDIAN__
   47803 #define vfmsq_lane_f32(__p0_94, __p1_94, __p2_94, __p3_94) __extension__ ({ \
   47804   float32x4_t __s0_94 = __p0_94; \
   47805   float32x4_t __s1_94 = __p1_94; \
   47806   float32x2_t __s2_94 = __p2_94; \
   47807   float32x4_t __ret_94; \
   47808   __ret_94 = vfmaq_lane_f32(__s0_94, -__s1_94, __s2_94, __p3_94); \
   47809   __ret_94; \
   47810 })
   47811 #else
   47812 #define vfmsq_lane_f32(__p0_95, __p1_95, __p2_95, __p3_95) __extension__ ({ \
   47813   float32x4_t __s0_95 = __p0_95; \
   47814   float32x4_t __s1_95 = __p1_95; \
   47815   float32x2_t __s2_95 = __p2_95; \
   47816   float32x4_t __rev0_95;  __rev0_95 = __builtin_shufflevector(__s0_95, __s0_95, 3, 2, 1, 0); \
   47817   float32x4_t __rev1_95;  __rev1_95 = __builtin_shufflevector(__s1_95, __s1_95, 3, 2, 1, 0); \
   47818   float32x2_t __rev2_95;  __rev2_95 = __builtin_shufflevector(__s2_95, __s2_95, 1, 0); \
   47819   float32x4_t __ret_95; \
   47820   __ret_95 = __noswap_vfmaq_lane_f32(__rev0_95, -__rev1_95, __rev2_95, __p3_95); \
   47821   __ret_95 = __builtin_shufflevector(__ret_95, __ret_95, 3, 2, 1, 0); \
   47822   __ret_95; \
   47823 })
   47824 #endif
   47825 
   47826 #ifdef __LITTLE_ENDIAN__
   47827 #define vfms_lane_f64(__p0_96, __p1_96, __p2_96, __p3_96) __extension__ ({ \
   47828   float64x1_t __s0_96 = __p0_96; \
   47829   float64x1_t __s1_96 = __p1_96; \
   47830   float64x1_t __s2_96 = __p2_96; \
   47831   float64x1_t __ret_96; \
   47832   __ret_96 = vfma_lane_f64(__s0_96, -__s1_96, __s2_96, __p3_96); \
   47833   __ret_96; \
   47834 })
   47835 #else
   47836 #define vfms_lane_f64(__p0_97, __p1_97, __p2_97, __p3_97) __extension__ ({ \
   47837   float64x1_t __s0_97 = __p0_97; \
   47838   float64x1_t __s1_97 = __p1_97; \
   47839   float64x1_t __s2_97 = __p2_97; \
   47840   float64x1_t __ret_97; \
   47841   __ret_97 = __noswap_vfma_lane_f64(__s0_97, -__s1_97, __s2_97, __p3_97); \
   47842   __ret_97; \
   47843 })
   47844 #endif
   47845 
   47846 #ifdef __LITTLE_ENDIAN__
   47847 #define vfms_lane_f32(__p0_98, __p1_98, __p2_98, __p3_98) __extension__ ({ \
   47848   float32x2_t __s0_98 = __p0_98; \
   47849   float32x2_t __s1_98 = __p1_98; \
   47850   float32x2_t __s2_98 = __p2_98; \
   47851   float32x2_t __ret_98; \
   47852   __ret_98 = vfma_lane_f32(__s0_98, -__s1_98, __s2_98, __p3_98); \
   47853   __ret_98; \
   47854 })
   47855 #else
   47856 #define vfms_lane_f32(__p0_99, __p1_99, __p2_99, __p3_99) __extension__ ({ \
   47857   float32x2_t __s0_99 = __p0_99; \
   47858   float32x2_t __s1_99 = __p1_99; \
   47859   float32x2_t __s2_99 = __p2_99; \
   47860   float32x2_t __rev0_99;  __rev0_99 = __builtin_shufflevector(__s0_99, __s0_99, 1, 0); \
   47861   float32x2_t __rev1_99;  __rev1_99 = __builtin_shufflevector(__s1_99, __s1_99, 1, 0); \
   47862   float32x2_t __rev2_99;  __rev2_99 = __builtin_shufflevector(__s2_99, __s2_99, 1, 0); \
   47863   float32x2_t __ret_99; \
   47864   __ret_99 = __noswap_vfma_lane_f32(__rev0_99, -__rev1_99, __rev2_99, __p3_99); \
   47865   __ret_99 = __builtin_shufflevector(__ret_99, __ret_99, 1, 0); \
   47866   __ret_99; \
   47867 })
   47868 #endif
   47869 
   47870 #ifdef __LITTLE_ENDIAN__
   47871 #define vfmsd_laneq_f64(__p0_100, __p1_100, __p2_100, __p3_100) __extension__ ({ \
   47872   float64_t __s0_100 = __p0_100; \
   47873   float64_t __s1_100 = __p1_100; \
   47874   float64x2_t __s2_100 = __p2_100; \
   47875   float64_t __ret_100; \
   47876   __ret_100 = vfmad_laneq_f64(__s0_100, -__s1_100, __s2_100, __p3_100); \
   47877   __ret_100; \
   47878 })
   47879 #else
   47880 #define vfmsd_laneq_f64(__p0_101, __p1_101, __p2_101, __p3_101) __extension__ ({ \
   47881   float64_t __s0_101 = __p0_101; \
   47882   float64_t __s1_101 = __p1_101; \
   47883   float64x2_t __s2_101 = __p2_101; \
   47884   float64x2_t __rev2_101;  __rev2_101 = __builtin_shufflevector(__s2_101, __s2_101, 1, 0); \
   47885   float64_t __ret_101; \
   47886   __ret_101 = __noswap_vfmad_laneq_f64(__s0_101, -__s1_101, __rev2_101, __p3_101); \
   47887   __ret_101; \
   47888 })
   47889 #endif
   47890 
   47891 #ifdef __LITTLE_ENDIAN__
   47892 #define vfmss_laneq_f32(__p0_102, __p1_102, __p2_102, __p3_102) __extension__ ({ \
   47893   float32_t __s0_102 = __p0_102; \
   47894   float32_t __s1_102 = __p1_102; \
   47895   float32x4_t __s2_102 = __p2_102; \
   47896   float32_t __ret_102; \
   47897   __ret_102 = vfmas_laneq_f32(__s0_102, -__s1_102, __s2_102, __p3_102); \
   47898   __ret_102; \
   47899 })
   47900 #else
   47901 #define vfmss_laneq_f32(__p0_103, __p1_103, __p2_103, __p3_103) __extension__ ({ \
   47902   float32_t __s0_103 = __p0_103; \
   47903   float32_t __s1_103 = __p1_103; \
   47904   float32x4_t __s2_103 = __p2_103; \
   47905   float32x4_t __rev2_103;  __rev2_103 = __builtin_shufflevector(__s2_103, __s2_103, 3, 2, 1, 0); \
   47906   float32_t __ret_103; \
   47907   __ret_103 = __noswap_vfmas_laneq_f32(__s0_103, -__s1_103, __rev2_103, __p3_103); \
   47908   __ret_103; \
   47909 })
   47910 #endif
   47911 
   47912 #ifdef __LITTLE_ENDIAN__
   47913 #define vfmsq_laneq_f64(__p0_104, __p1_104, __p2_104, __p3_104) __extension__ ({ \
   47914   float64x2_t __s0_104 = __p0_104; \
   47915   float64x2_t __s1_104 = __p1_104; \
   47916   float64x2_t __s2_104 = __p2_104; \
   47917   float64x2_t __ret_104; \
   47918   __ret_104 = vfmaq_laneq_f64(__s0_104, -__s1_104, __s2_104, __p3_104); \
   47919   __ret_104; \
   47920 })
   47921 #else
   47922 #define vfmsq_laneq_f64(__p0_105, __p1_105, __p2_105, __p3_105) __extension__ ({ \
   47923   float64x2_t __s0_105 = __p0_105; \
   47924   float64x2_t __s1_105 = __p1_105; \
   47925   float64x2_t __s2_105 = __p2_105; \
   47926   float64x2_t __rev0_105;  __rev0_105 = __builtin_shufflevector(__s0_105, __s0_105, 1, 0); \
   47927   float64x2_t __rev1_105;  __rev1_105 = __builtin_shufflevector(__s1_105, __s1_105, 1, 0); \
   47928   float64x2_t __rev2_105;  __rev2_105 = __builtin_shufflevector(__s2_105, __s2_105, 1, 0); \
   47929   float64x2_t __ret_105; \
   47930   __ret_105 = __noswap_vfmaq_laneq_f64(__rev0_105, -__rev1_105, __rev2_105, __p3_105); \
   47931   __ret_105 = __builtin_shufflevector(__ret_105, __ret_105, 1, 0); \
   47932   __ret_105; \
   47933 })
   47934 #endif
   47935 
   47936 #ifdef __LITTLE_ENDIAN__
   47937 #define vfmsq_laneq_f32(__p0_106, __p1_106, __p2_106, __p3_106) __extension__ ({ \
   47938   float32x4_t __s0_106 = __p0_106; \
   47939   float32x4_t __s1_106 = __p1_106; \
   47940   float32x4_t __s2_106 = __p2_106; \
   47941   float32x4_t __ret_106; \
   47942   __ret_106 = vfmaq_laneq_f32(__s0_106, -__s1_106, __s2_106, __p3_106); \
   47943   __ret_106; \
   47944 })
   47945 #else
   47946 #define vfmsq_laneq_f32(__p0_107, __p1_107, __p2_107, __p3_107) __extension__ ({ \
   47947   float32x4_t __s0_107 = __p0_107; \
   47948   float32x4_t __s1_107 = __p1_107; \
   47949   float32x4_t __s2_107 = __p2_107; \
   47950   float32x4_t __rev0_107;  __rev0_107 = __builtin_shufflevector(__s0_107, __s0_107, 3, 2, 1, 0); \
   47951   float32x4_t __rev1_107;  __rev1_107 = __builtin_shufflevector(__s1_107, __s1_107, 3, 2, 1, 0); \
   47952   float32x4_t __rev2_107;  __rev2_107 = __builtin_shufflevector(__s2_107, __s2_107, 3, 2, 1, 0); \
   47953   float32x4_t __ret_107; \
   47954   __ret_107 = __noswap_vfmaq_laneq_f32(__rev0_107, -__rev1_107, __rev2_107, __p3_107); \
   47955   __ret_107 = __builtin_shufflevector(__ret_107, __ret_107, 3, 2, 1, 0); \
   47956   __ret_107; \
   47957 })
   47958 #endif
   47959 
   47960 #ifdef __LITTLE_ENDIAN__
   47961 #define vfms_laneq_f64(__p0_108, __p1_108, __p2_108, __p3_108) __extension__ ({ \
   47962   float64x1_t __s0_108 = __p0_108; \
   47963   float64x1_t __s1_108 = __p1_108; \
   47964   float64x2_t __s2_108 = __p2_108; \
   47965   float64x1_t __ret_108; \
   47966   __ret_108 = vfma_laneq_f64(__s0_108, -__s1_108, __s2_108, __p3_108); \
   47967   __ret_108; \
   47968 })
   47969 #else
   47970 #define vfms_laneq_f64(__p0_109, __p1_109, __p2_109, __p3_109) __extension__ ({ \
   47971   float64x1_t __s0_109 = __p0_109; \
   47972   float64x1_t __s1_109 = __p1_109; \
   47973   float64x2_t __s2_109 = __p2_109; \
   47974   float64x2_t __rev2_109;  __rev2_109 = __builtin_shufflevector(__s2_109, __s2_109, 1, 0); \
   47975   float64x1_t __ret_109; \
   47976   __ret_109 = __noswap_vfma_laneq_f64(__s0_109, -__s1_109, __rev2_109, __p3_109); \
   47977   __ret_109; \
   47978 })
   47979 #endif
   47980 
   47981 #ifdef __LITTLE_ENDIAN__
   47982 #define vfms_laneq_f32(__p0_110, __p1_110, __p2_110, __p3_110) __extension__ ({ \
   47983   float32x2_t __s0_110 = __p0_110; \
   47984   float32x2_t __s1_110 = __p1_110; \
   47985   float32x4_t __s2_110 = __p2_110; \
   47986   float32x2_t __ret_110; \
   47987   __ret_110 = vfma_laneq_f32(__s0_110, -__s1_110, __s2_110, __p3_110); \
   47988   __ret_110; \
   47989 })
   47990 #else
   47991 #define vfms_laneq_f32(__p0_111, __p1_111, __p2_111, __p3_111) __extension__ ({ \
   47992   float32x2_t __s0_111 = __p0_111; \
   47993   float32x2_t __s1_111 = __p1_111; \
   47994   float32x4_t __s2_111 = __p2_111; \
   47995   float32x2_t __rev0_111;  __rev0_111 = __builtin_shufflevector(__s0_111, __s0_111, 1, 0); \
   47996   float32x2_t __rev1_111;  __rev1_111 = __builtin_shufflevector(__s1_111, __s1_111, 1, 0); \
   47997   float32x4_t __rev2_111;  __rev2_111 = __builtin_shufflevector(__s2_111, __s2_111, 3, 2, 1, 0); \
   47998   float32x2_t __ret_111; \
   47999   __ret_111 = __noswap_vfma_laneq_f32(__rev0_111, -__rev1_111, __rev2_111, __p3_111); \
   48000   __ret_111 = __builtin_shufflevector(__ret_111, __ret_111, 1, 0); \
   48001   __ret_111; \
   48002 })
   48003 #endif
   48004 
   48005 #ifdef __LITTLE_ENDIAN__
   48006 __ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
   48007   float64x2_t __ret;
   48008   __ret = vfmaq_f64(__p0, -__p1, (float64x2_t) {__p2, __p2});
   48009   return __ret;
   48010 }
   48011 #else
   48012 __ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
   48013   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   48014   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   48015   float64x2_t __ret;
   48016   __ret = __noswap_vfmaq_f64(__rev0, -__rev1, (float64x2_t) {__p2, __p2});
   48017   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   48018   return __ret;
   48019 }
   48020 #endif
   48021 
   48022 #ifdef __LITTLE_ENDIAN__
   48023 __ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
   48024   float32x4_t __ret;
   48025   __ret = vfmaq_f32(__p0, -__p1, (float32x4_t) {__p2, __p2, __p2, __p2});
   48026   return __ret;
   48027 }
   48028 #else
   48029 __ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
   48030   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   48031   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   48032   float32x4_t __ret;
   48033   __ret = __noswap_vfmaq_f32(__rev0, -__rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
   48034   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   48035   return __ret;
   48036 }
   48037 #endif
   48038 
   48039 #ifdef __LITTLE_ENDIAN__
   48040 __ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
   48041   float32x2_t __ret;
   48042   __ret = vfma_f32(__p0, -__p1, (float32x2_t) {__p2, __p2});
   48043   return __ret;
   48044 }
   48045 #else
   48046 __ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
   48047   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   48048   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   48049   float32x2_t __ret;
   48050   __ret = __noswap_vfma_f32(__rev0, -__rev1, (float32x2_t) {__p2, __p2});
   48051   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   48052   return __ret;
   48053 }
   48054 #endif
   48055 
   48056 #ifdef __LITTLE_ENDIAN__
   48057 __ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
   48058   poly64x1_t __ret;
   48059   __ret = __builtin_shufflevector(__p0, __p0, 1);
   48060   return __ret;
   48061 }
   48062 #else
   48063 __ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
   48064   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   48065   poly64x1_t __ret;
   48066   __ret = __builtin_shufflevector(__rev0, __rev0, 1);
   48067   return __ret;
   48068 }
   48069 __ai poly64x1_t __noswap_vget_high_p64(poly64x2_t __p0) {
   48070   poly64x1_t __ret;
   48071   __ret = __builtin_shufflevector(__p0, __p0, 1);
   48072   return __ret;
   48073 }
   48074 #endif
   48075 
   48076 #ifdef __LITTLE_ENDIAN__
   48077 __ai float64x1_t vget_high_f64(float64x2_t __p0) {
   48078   float64x1_t __ret;
   48079   __ret = __builtin_shufflevector(__p0, __p0, 1);
   48080   return __ret;
   48081 }
   48082 #else
   48083 __ai float64x1_t vget_high_f64(float64x2_t __p0) {
   48084   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   48085   float64x1_t __ret;
   48086   __ret = __builtin_shufflevector(__rev0, __rev0, 1);
   48087   return __ret;
   48088 }
   48089 #endif
   48090 
   48091 #ifdef __LITTLE_ENDIAN__
   48092 #define vget_lane_p64(__p0, __p1) __extension__ ({ \
   48093   poly64x1_t __s0 = __p0; \
   48094   poly64_t __ret; \
   48095   __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
   48096   __ret; \
   48097 })
   48098 #else
   48099 #define vget_lane_p64(__p0, __p1) __extension__ ({ \
   48100   poly64x1_t __s0 = __p0; \
   48101   poly64_t __ret; \
   48102   __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
   48103   __ret; \
   48104 })
   48105 #define __noswap_vget_lane_p64(__p0, __p1) __extension__ ({ \
   48106   poly64x1_t __s0 = __p0; \
   48107   poly64_t __ret; \
   48108   __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
   48109   __ret; \
   48110 })
   48111 #endif
   48112 
   48113 #ifdef __LITTLE_ENDIAN__
   48114 #define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
   48115   poly64x2_t __s0 = __p0; \
   48116   poly64_t __ret; \
   48117   __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
   48118   __ret; \
   48119 })
   48120 #else
   48121 #define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
   48122   poly64x2_t __s0 = __p0; \
   48123   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   48124   poly64_t __ret; \
   48125   __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
   48126   __ret; \
   48127 })
   48128 #define __noswap_vgetq_lane_p64(__p0, __p1) __extension__ ({ \
   48129   poly64x2_t __s0 = __p0; \
   48130   poly64_t __ret; \
   48131   __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
   48132   __ret; \
   48133 })
   48134 #endif
   48135 
   48136 #ifdef __LITTLE_ENDIAN__
   48137 #define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
   48138   float64x2_t __s0 = __p0; \
   48139   float64_t __ret; \
   48140   __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
   48141   __ret; \
   48142 })
   48143 #else
   48144 #define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
   48145   float64x2_t __s0 = __p0; \
   48146   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   48147   float64_t __ret; \
   48148   __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__rev0, __p1); \
   48149   __ret; \
   48150 })
   48151 #define __noswap_vgetq_lane_f64(__p0, __p1) __extension__ ({ \
   48152   float64x2_t __s0 = __p0; \
   48153   float64_t __ret; \
   48154   __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
   48155   __ret; \
   48156 })
   48157 #endif
   48158 
   48159 #ifdef __LITTLE_ENDIAN__
   48160 #define vget_lane_f64(__p0, __p1) __extension__ ({ \
   48161   float64x1_t __s0 = __p0; \
   48162   float64_t __ret; \
   48163   __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
   48164   __ret; \
   48165 })
   48166 #else
   48167 #define vget_lane_f64(__p0, __p1) __extension__ ({ \
   48168   float64x1_t __s0 = __p0; \
   48169   float64_t __ret; \
   48170   __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
   48171   __ret; \
   48172 })
   48173 #define __noswap_vget_lane_f64(__p0, __p1) __extension__ ({ \
   48174   float64x1_t __s0 = __p0; \
   48175   float64_t __ret; \
   48176   __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
   48177   __ret; \
   48178 })
   48179 #endif
   48180 
   48181 #ifdef __LITTLE_ENDIAN__
   48182 __ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
   48183   poly64x1_t __ret;
   48184   __ret = __builtin_shufflevector(__p0, __p0, 0);
   48185   return __ret;
   48186 }
   48187 #else
   48188 __ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
   48189   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   48190   poly64x1_t __ret;
   48191   __ret = __builtin_shufflevector(__rev0, __rev0, 0);
   48192   return __ret;
   48193 }
   48194 #endif
   48195 
   48196 #ifdef __LITTLE_ENDIAN__
   48197 __ai float64x1_t vget_low_f64(float64x2_t __p0) {
   48198   float64x1_t __ret;
   48199   __ret = __builtin_shufflevector(__p0, __p0, 0);
   48200   return __ret;
   48201 }
   48202 #else
   48203 __ai float64x1_t vget_low_f64(float64x2_t __p0) {
   48204   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   48205   float64x1_t __ret;
   48206   __ret = __builtin_shufflevector(__rev0, __rev0, 0);
   48207   return __ret;
   48208 }
   48209 #endif
   48210 
   48211 #ifdef __LITTLE_ENDIAN__
   48212 #define vld1_p64(__p0) __extension__ ({ \
   48213   poly64x1_t __ret; \
   48214   __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
   48215   __ret; \
   48216 })
   48217 #else
   48218 #define vld1_p64(__p0) __extension__ ({ \
   48219   poly64x1_t __ret; \
   48220   __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
   48221   __ret; \
   48222 })
   48223 #endif
   48224 
   48225 #ifdef __LITTLE_ENDIAN__
   48226 #define vld1q_p64(__p0) __extension__ ({ \
   48227   poly64x2_t __ret; \
   48228   __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
   48229   __ret; \
   48230 })
   48231 #else
   48232 #define vld1q_p64(__p0) __extension__ ({ \
   48233   poly64x2_t __ret; \
   48234   __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
   48235   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   48236   __ret; \
   48237 })
   48238 #endif
   48239 
   48240 #ifdef __LITTLE_ENDIAN__
   48241 #define vld1q_f64(__p0) __extension__ ({ \
   48242   float64x2_t __ret; \
   48243   __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
   48244   __ret; \
   48245 })
   48246 #else
   48247 #define vld1q_f64(__p0) __extension__ ({ \
   48248   float64x2_t __ret; \
   48249   __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
   48250   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   48251   __ret; \
   48252 })
   48253 #endif
   48254 
   48255 #ifdef __LITTLE_ENDIAN__
   48256 #define vld1_f64(__p0) __extension__ ({ \
   48257   float64x1_t __ret; \
   48258   __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
   48259   __ret; \
   48260 })
   48261 #else
   48262 #define vld1_f64(__p0) __extension__ ({ \
   48263   float64x1_t __ret; \
   48264   __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
   48265   __ret; \
   48266 })
   48267 #endif
   48268 
   48269 #ifdef __LITTLE_ENDIAN__
   48270 #define vld1_dup_p64(__p0) __extension__ ({ \
   48271   poly64x1_t __ret; \
   48272   __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
   48273   __ret; \
   48274 })
   48275 #else
   48276 #define vld1_dup_p64(__p0) __extension__ ({ \
   48277   poly64x1_t __ret; \
   48278   __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
   48279   __ret; \
   48280 })
   48281 #endif
   48282 
   48283 #ifdef __LITTLE_ENDIAN__
   48284 #define vld1q_dup_p64(__p0) __extension__ ({ \
   48285   poly64x2_t __ret; \
   48286   __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
   48287   __ret; \
   48288 })
   48289 #else
   48290 #define vld1q_dup_p64(__p0) __extension__ ({ \
   48291   poly64x2_t __ret; \
   48292   __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
   48293   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   48294   __ret; \
   48295 })
   48296 #endif
   48297 
   48298 #ifdef __LITTLE_ENDIAN__
   48299 #define vld1q_dup_f64(__p0) __extension__ ({ \
   48300   float64x2_t __ret; \
   48301   __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
   48302   __ret; \
   48303 })
   48304 #else
   48305 #define vld1q_dup_f64(__p0) __extension__ ({ \
   48306   float64x2_t __ret; \
   48307   __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
   48308   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   48309   __ret; \
   48310 })
   48311 #endif
   48312 
   48313 #ifdef __LITTLE_ENDIAN__
   48314 #define vld1_dup_f64(__p0) __extension__ ({ \
   48315   float64x1_t __ret; \
   48316   __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
   48317   __ret; \
   48318 })
   48319 #else
   48320 #define vld1_dup_f64(__p0) __extension__ ({ \
   48321   float64x1_t __ret; \
   48322   __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
   48323   __ret; \
   48324 })
   48325 #endif
   48326 
   48327 #ifdef __LITTLE_ENDIAN__
   48328 #define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   48329   poly64x1_t __s1 = __p1; \
   48330   poly64x1_t __ret; \
   48331   __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
   48332   __ret; \
   48333 })
   48334 #else
   48335 #define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   48336   poly64x1_t __s1 = __p1; \
   48337   poly64x1_t __ret; \
   48338   __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
   48339   __ret; \
   48340 })
   48341 #endif
   48342 
   48343 #ifdef __LITTLE_ENDIAN__
   48344 #define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   48345   poly64x2_t __s1 = __p1; \
   48346   poly64x2_t __ret; \
   48347   __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
   48348   __ret; \
   48349 })
   48350 #else
   48351 #define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   48352   poly64x2_t __s1 = __p1; \
   48353   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   48354   poly64x2_t __ret; \
   48355   __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
   48356   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   48357   __ret; \
   48358 })
   48359 #endif
   48360 
   48361 #ifdef __LITTLE_ENDIAN__
   48362 #define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   48363   float64x2_t __s1 = __p1; \
   48364   float64x2_t __ret; \
   48365   __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
   48366   __ret; \
   48367 })
   48368 #else
   48369 #define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   48370   float64x2_t __s1 = __p1; \
   48371   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   48372   float64x2_t __ret; \
   48373   __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
   48374   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   48375   __ret; \
   48376 })
   48377 #endif
   48378 
   48379 #ifdef __LITTLE_ENDIAN__
   48380 #define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   48381   float64x1_t __s1 = __p1; \
   48382   float64x1_t __ret; \
   48383   __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
   48384   __ret; \
   48385 })
   48386 #else
   48387 #define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   48388   float64x1_t __s1 = __p1; \
   48389   float64x1_t __ret; \
   48390   __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
   48391   __ret; \
   48392 })
   48393 #endif
   48394 
   48395 #ifdef __LITTLE_ENDIAN__
   48396 #define vld1_p8_x2(__p0) __extension__ ({ \
   48397   poly8x8x2_t __ret; \
   48398   __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
   48399   __ret; \
   48400 })
   48401 #else
   48402 #define vld1_p8_x2(__p0) __extension__ ({ \
   48403   poly8x8x2_t __ret; \
   48404   __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
   48405  \
   48406   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   48407   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   48408   __ret; \
   48409 })
   48410 #endif
   48411 
   48412 #ifdef __LITTLE_ENDIAN__
   48413 #define vld1_p64_x2(__p0) __extension__ ({ \
   48414   poly64x1x2_t __ret; \
   48415   __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
   48416   __ret; \
   48417 })
   48418 #else
   48419 #define vld1_p64_x2(__p0) __extension__ ({ \
   48420   poly64x1x2_t __ret; \
   48421   __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
   48422   __ret; \
   48423 })
   48424 #endif
   48425 
   48426 #ifdef __LITTLE_ENDIAN__
   48427 #define vld1_p16_x2(__p0) __extension__ ({ \
   48428   poly16x4x2_t __ret; \
   48429   __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
   48430   __ret; \
   48431 })
   48432 #else
   48433 #define vld1_p16_x2(__p0) __extension__ ({ \
   48434   poly16x4x2_t __ret; \
   48435   __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
   48436  \
   48437   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   48438   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   48439   __ret; \
   48440 })
   48441 #endif
   48442 
   48443 #ifdef __LITTLE_ENDIAN__
   48444 #define vld1q_p8_x2(__p0) __extension__ ({ \
   48445   poly8x16x2_t __ret; \
   48446   __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
   48447   __ret; \
   48448 })
   48449 #else
   48450 #define vld1q_p8_x2(__p0) __extension__ ({ \
   48451   poly8x16x2_t __ret; \
   48452   __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
   48453  \
   48454   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48455   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48456   __ret; \
   48457 })
   48458 #endif
   48459 
   48460 #ifdef __LITTLE_ENDIAN__
   48461 #define vld1q_p64_x2(__p0) __extension__ ({ \
   48462   poly64x2x2_t __ret; \
   48463   __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
   48464   __ret; \
   48465 })
   48466 #else
   48467 #define vld1q_p64_x2(__p0) __extension__ ({ \
   48468   poly64x2x2_t __ret; \
   48469   __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
   48470  \
   48471   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   48472   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   48473   __ret; \
   48474 })
   48475 #endif
   48476 
   48477 #ifdef __LITTLE_ENDIAN__
   48478 #define vld1q_p16_x2(__p0) __extension__ ({ \
   48479   poly16x8x2_t __ret; \
   48480   __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
   48481   __ret; \
   48482 })
   48483 #else
   48484 #define vld1q_p16_x2(__p0) __extension__ ({ \
   48485   poly16x8x2_t __ret; \
   48486   __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
   48487  \
   48488   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   48489   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   48490   __ret; \
   48491 })
   48492 #endif
   48493 
   48494 #ifdef __LITTLE_ENDIAN__
   48495 #define vld1q_u8_x2(__p0) __extension__ ({ \
   48496   uint8x16x2_t __ret; \
   48497   __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
   48498   __ret; \
   48499 })
   48500 #else
   48501 #define vld1q_u8_x2(__p0) __extension__ ({ \
   48502   uint8x16x2_t __ret; \
   48503   __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
   48504  \
   48505   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48506   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48507   __ret; \
   48508 })
   48509 #endif
   48510 
   48511 #ifdef __LITTLE_ENDIAN__
   48512 #define vld1q_u32_x2(__p0) __extension__ ({ \
   48513   uint32x4x2_t __ret; \
   48514   __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
   48515   __ret; \
   48516 })
   48517 #else
   48518 #define vld1q_u32_x2(__p0) __extension__ ({ \
   48519   uint32x4x2_t __ret; \
   48520   __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
   48521  \
   48522   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   48523   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   48524   __ret; \
   48525 })
   48526 #endif
   48527 
   48528 #ifdef __LITTLE_ENDIAN__
   48529 #define vld1q_u64_x2(__p0) __extension__ ({ \
   48530   uint64x2x2_t __ret; \
   48531   __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
   48532   __ret; \
   48533 })
   48534 #else
   48535 #define vld1q_u64_x2(__p0) __extension__ ({ \
   48536   uint64x2x2_t __ret; \
   48537   __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
   48538  \
   48539   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   48540   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   48541   __ret; \
   48542 })
   48543 #endif
   48544 
   48545 #ifdef __LITTLE_ENDIAN__
   48546 #define vld1q_u16_x2(__p0) __extension__ ({ \
   48547   uint16x8x2_t __ret; \
   48548   __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
   48549   __ret; \
   48550 })
   48551 #else
   48552 #define vld1q_u16_x2(__p0) __extension__ ({ \
   48553   uint16x8x2_t __ret; \
   48554   __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
   48555  \
   48556   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   48557   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   48558   __ret; \
   48559 })
   48560 #endif
   48561 
   48562 #ifdef __LITTLE_ENDIAN__
   48563 #define vld1q_s8_x2(__p0) __extension__ ({ \
   48564   int8x16x2_t __ret; \
   48565   __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
   48566   __ret; \
   48567 })
   48568 #else
   48569 #define vld1q_s8_x2(__p0) __extension__ ({ \
   48570   int8x16x2_t __ret; \
   48571   __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
   48572  \
   48573   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48574   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48575   __ret; \
   48576 })
   48577 #endif
   48578 
   48579 #ifdef __LITTLE_ENDIAN__
   48580 #define vld1q_f64_x2(__p0) __extension__ ({ \
   48581   float64x2x2_t __ret; \
   48582   __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
   48583   __ret; \
   48584 })
   48585 #else
   48586 #define vld1q_f64_x2(__p0) __extension__ ({ \
   48587   float64x2x2_t __ret; \
   48588   __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
   48589  \
   48590   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   48591   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   48592   __ret; \
   48593 })
   48594 #endif
   48595 
   48596 #ifdef __LITTLE_ENDIAN__
   48597 #define vld1q_f32_x2(__p0) __extension__ ({ \
   48598   float32x4x2_t __ret; \
   48599   __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
   48600   __ret; \
   48601 })
   48602 #else
   48603 #define vld1q_f32_x2(__p0) __extension__ ({ \
   48604   float32x4x2_t __ret; \
   48605   __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
   48606  \
   48607   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   48608   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   48609   __ret; \
   48610 })
   48611 #endif
   48612 
   48613 #ifdef __LITTLE_ENDIAN__
   48614 #define vld1q_f16_x2(__p0) __extension__ ({ \
   48615   float16x8x2_t __ret; \
   48616   __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
   48617   __ret; \
   48618 })
   48619 #else
   48620 #define vld1q_f16_x2(__p0) __extension__ ({ \
   48621   float16x8x2_t __ret; \
   48622   __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
   48623  \
   48624   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   48625   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   48626   __ret; \
   48627 })
   48628 #endif
   48629 
   48630 #ifdef __LITTLE_ENDIAN__
   48631 #define vld1q_s32_x2(__p0) __extension__ ({ \
   48632   int32x4x2_t __ret; \
   48633   __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
   48634   __ret; \
   48635 })
   48636 #else
   48637 #define vld1q_s32_x2(__p0) __extension__ ({ \
   48638   int32x4x2_t __ret; \
   48639   __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
   48640  \
   48641   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   48642   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   48643   __ret; \
   48644 })
   48645 #endif
   48646 
   48647 #ifdef __LITTLE_ENDIAN__
   48648 #define vld1q_s64_x2(__p0) __extension__ ({ \
   48649   int64x2x2_t __ret; \
   48650   __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
   48651   __ret; \
   48652 })
   48653 #else
   48654 #define vld1q_s64_x2(__p0) __extension__ ({ \
   48655   int64x2x2_t __ret; \
   48656   __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
   48657  \
   48658   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   48659   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   48660   __ret; \
   48661 })
   48662 #endif
   48663 
   48664 #ifdef __LITTLE_ENDIAN__
   48665 #define vld1q_s16_x2(__p0) __extension__ ({ \
   48666   int16x8x2_t __ret; \
   48667   __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
   48668   __ret; \
   48669 })
   48670 #else
   48671 #define vld1q_s16_x2(__p0) __extension__ ({ \
   48672   int16x8x2_t __ret; \
   48673   __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
   48674  \
   48675   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   48676   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   48677   __ret; \
   48678 })
   48679 #endif
   48680 
   48681 #ifdef __LITTLE_ENDIAN__
   48682 #define vld1_u8_x2(__p0) __extension__ ({ \
   48683   uint8x8x2_t __ret; \
   48684   __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
   48685   __ret; \
   48686 })
   48687 #else
   48688 #define vld1_u8_x2(__p0) __extension__ ({ \
   48689   uint8x8x2_t __ret; \
   48690   __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
   48691  \
   48692   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   48693   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   48694   __ret; \
   48695 })
   48696 #endif
   48697 
   48698 #ifdef __LITTLE_ENDIAN__
   48699 #define vld1_u32_x2(__p0) __extension__ ({ \
   48700   uint32x2x2_t __ret; \
   48701   __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
   48702   __ret; \
   48703 })
   48704 #else
   48705 #define vld1_u32_x2(__p0) __extension__ ({ \
   48706   uint32x2x2_t __ret; \
   48707   __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
   48708  \
   48709   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   48710   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   48711   __ret; \
   48712 })
   48713 #endif
   48714 
   48715 #ifdef __LITTLE_ENDIAN__
   48716 #define vld1_u64_x2(__p0) __extension__ ({ \
   48717   uint64x1x2_t __ret; \
   48718   __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
   48719   __ret; \
   48720 })
   48721 #else
   48722 #define vld1_u64_x2(__p0) __extension__ ({ \
   48723   uint64x1x2_t __ret; \
   48724   __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
   48725   __ret; \
   48726 })
   48727 #endif
   48728 
   48729 #ifdef __LITTLE_ENDIAN__
   48730 #define vld1_u16_x2(__p0) __extension__ ({ \
   48731   uint16x4x2_t __ret; \
   48732   __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
   48733   __ret; \
   48734 })
   48735 #else
   48736 #define vld1_u16_x2(__p0) __extension__ ({ \
   48737   uint16x4x2_t __ret; \
   48738   __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
   48739  \
   48740   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   48741   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   48742   __ret; \
   48743 })
   48744 #endif
   48745 
   48746 #ifdef __LITTLE_ENDIAN__
   48747 #define vld1_s8_x2(__p0) __extension__ ({ \
   48748   int8x8x2_t __ret; \
   48749   __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
   48750   __ret; \
   48751 })
   48752 #else
   48753 #define vld1_s8_x2(__p0) __extension__ ({ \
   48754   int8x8x2_t __ret; \
   48755   __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
   48756  \
   48757   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   48758   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   48759   __ret; \
   48760 })
   48761 #endif
   48762 
   48763 #ifdef __LITTLE_ENDIAN__
   48764 #define vld1_f64_x2(__p0) __extension__ ({ \
   48765   float64x1x2_t __ret; \
   48766   __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
   48767   __ret; \
   48768 })
   48769 #else
   48770 #define vld1_f64_x2(__p0) __extension__ ({ \
   48771   float64x1x2_t __ret; \
   48772   __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
   48773   __ret; \
   48774 })
   48775 #endif
   48776 
   48777 #ifdef __LITTLE_ENDIAN__
   48778 #define vld1_f32_x2(__p0) __extension__ ({ \
   48779   float32x2x2_t __ret; \
   48780   __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
   48781   __ret; \
   48782 })
   48783 #else
   48784 #define vld1_f32_x2(__p0) __extension__ ({ \
   48785   float32x2x2_t __ret; \
   48786   __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
   48787  \
   48788   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   48789   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   48790   __ret; \
   48791 })
   48792 #endif
   48793 
   48794 #ifdef __LITTLE_ENDIAN__
   48795 #define vld1_f16_x2(__p0) __extension__ ({ \
   48796   float16x4x2_t __ret; \
   48797   __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
   48798   __ret; \
   48799 })
   48800 #else
   48801 #define vld1_f16_x2(__p0) __extension__ ({ \
   48802   float16x4x2_t __ret; \
   48803   __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
   48804  \
   48805   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   48806   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   48807   __ret; \
   48808 })
   48809 #endif
   48810 
   48811 #ifdef __LITTLE_ENDIAN__
   48812 #define vld1_s32_x2(__p0) __extension__ ({ \
   48813   int32x2x2_t __ret; \
   48814   __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
   48815   __ret; \
   48816 })
   48817 #else
   48818 #define vld1_s32_x2(__p0) __extension__ ({ \
   48819   int32x2x2_t __ret; \
   48820   __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
   48821  \
   48822   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   48823   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   48824   __ret; \
   48825 })
   48826 #endif
   48827 
   48828 #ifdef __LITTLE_ENDIAN__
   48829 #define vld1_s64_x2(__p0) __extension__ ({ \
   48830   int64x1x2_t __ret; \
   48831   __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
   48832   __ret; \
   48833 })
   48834 #else
   48835 #define vld1_s64_x2(__p0) __extension__ ({ \
   48836   int64x1x2_t __ret; \
   48837   __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
   48838   __ret; \
   48839 })
   48840 #endif
   48841 
   48842 #ifdef __LITTLE_ENDIAN__
   48843 #define vld1_s16_x2(__p0) __extension__ ({ \
   48844   int16x4x2_t __ret; \
   48845   __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
   48846   __ret; \
   48847 })
   48848 #else
   48849 #define vld1_s16_x2(__p0) __extension__ ({ \
   48850   int16x4x2_t __ret; \
   48851   __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
   48852  \
   48853   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   48854   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   48855   __ret; \
   48856 })
   48857 #endif
   48858 
   48859 #ifdef __LITTLE_ENDIAN__
   48860 #define vld1_p8_x3(__p0) __extension__ ({ \
   48861   poly8x8x3_t __ret; \
   48862   __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
   48863   __ret; \
   48864 })
   48865 #else
   48866 #define vld1_p8_x3(__p0) __extension__ ({ \
   48867   poly8x8x3_t __ret; \
   48868   __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
   48869  \
   48870   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   48871   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   48872   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   48873   __ret; \
   48874 })
   48875 #endif
   48876 
   48877 #ifdef __LITTLE_ENDIAN__
   48878 #define vld1_p64_x3(__p0) __extension__ ({ \
   48879   poly64x1x3_t __ret; \
   48880   __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
   48881   __ret; \
   48882 })
   48883 #else
   48884 #define vld1_p64_x3(__p0) __extension__ ({ \
   48885   poly64x1x3_t __ret; \
   48886   __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
   48887   __ret; \
   48888 })
   48889 #endif
   48890 
   48891 #ifdef __LITTLE_ENDIAN__
   48892 #define vld1_p16_x3(__p0) __extension__ ({ \
   48893   poly16x4x3_t __ret; \
   48894   __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
   48895   __ret; \
   48896 })
   48897 #else
   48898 #define vld1_p16_x3(__p0) __extension__ ({ \
   48899   poly16x4x3_t __ret; \
   48900   __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
   48901  \
   48902   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   48903   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   48904   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   48905   __ret; \
   48906 })
   48907 #endif
   48908 
   48909 #ifdef __LITTLE_ENDIAN__
   48910 #define vld1q_p8_x3(__p0) __extension__ ({ \
   48911   poly8x16x3_t __ret; \
   48912   __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
   48913   __ret; \
   48914 })
   48915 #else
   48916 #define vld1q_p8_x3(__p0) __extension__ ({ \
   48917   poly8x16x3_t __ret; \
   48918   __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
   48919  \
   48920   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48921   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48922   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48923   __ret; \
   48924 })
   48925 #endif
   48926 
   48927 #ifdef __LITTLE_ENDIAN__
   48928 #define vld1q_p64_x3(__p0) __extension__ ({ \
   48929   poly64x2x3_t __ret; \
   48930   __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
   48931   __ret; \
   48932 })
   48933 #else
   48934 #define vld1q_p64_x3(__p0) __extension__ ({ \
   48935   poly64x2x3_t __ret; \
   48936   __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
   48937  \
   48938   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   48939   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   48940   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   48941   __ret; \
   48942 })
   48943 #endif
   48944 
   48945 #ifdef __LITTLE_ENDIAN__
   48946 #define vld1q_p16_x3(__p0) __extension__ ({ \
   48947   poly16x8x3_t __ret; \
   48948   __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
   48949   __ret; \
   48950 })
   48951 #else
   48952 #define vld1q_p16_x3(__p0) __extension__ ({ \
   48953   poly16x8x3_t __ret; \
   48954   __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
   48955  \
   48956   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   48957   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   48958   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   48959   __ret; \
   48960 })
   48961 #endif
   48962 
   48963 #ifdef __LITTLE_ENDIAN__
   48964 #define vld1q_u8_x3(__p0) __extension__ ({ \
   48965   uint8x16x3_t __ret; \
   48966   __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
   48967   __ret; \
   48968 })
   48969 #else
   48970 #define vld1q_u8_x3(__p0) __extension__ ({ \
   48971   uint8x16x3_t __ret; \
   48972   __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
   48973  \
   48974   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48975   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48976   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   48977   __ret; \
   48978 })
   48979 #endif
   48980 
   48981 #ifdef __LITTLE_ENDIAN__
   48982 #define vld1q_u32_x3(__p0) __extension__ ({ \
   48983   uint32x4x3_t __ret; \
   48984   __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
   48985   __ret; \
   48986 })
   48987 #else
   48988 #define vld1q_u32_x3(__p0) __extension__ ({ \
   48989   uint32x4x3_t __ret; \
   48990   __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
   48991  \
   48992   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   48993   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   48994   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   48995   __ret; \
   48996 })
   48997 #endif
   48998 
   48999 #ifdef __LITTLE_ENDIAN__
   49000 #define vld1q_u64_x3(__p0) __extension__ ({ \
   49001   uint64x2x3_t __ret; \
   49002   __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
   49003   __ret; \
   49004 })
   49005 #else
   49006 #define vld1q_u64_x3(__p0) __extension__ ({ \
   49007   uint64x2x3_t __ret; \
   49008   __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
   49009  \
   49010   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49011   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49012   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49013   __ret; \
   49014 })
   49015 #endif
   49016 
   49017 #ifdef __LITTLE_ENDIAN__
   49018 #define vld1q_u16_x3(__p0) __extension__ ({ \
   49019   uint16x8x3_t __ret; \
   49020   __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
   49021   __ret; \
   49022 })
   49023 #else
   49024 #define vld1q_u16_x3(__p0) __extension__ ({ \
   49025   uint16x8x3_t __ret; \
   49026   __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
   49027  \
   49028   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49029   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49030   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49031   __ret; \
   49032 })
   49033 #endif
   49034 
   49035 #ifdef __LITTLE_ENDIAN__
   49036 #define vld1q_s8_x3(__p0) __extension__ ({ \
   49037   int8x16x3_t __ret; \
   49038   __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
   49039   __ret; \
   49040 })
   49041 #else
   49042 #define vld1q_s8_x3(__p0) __extension__ ({ \
   49043   int8x16x3_t __ret; \
   49044   __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
   49045  \
   49046   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49047   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49048   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49049   __ret; \
   49050 })
   49051 #endif
   49052 
   49053 #ifdef __LITTLE_ENDIAN__
   49054 #define vld1q_f64_x3(__p0) __extension__ ({ \
   49055   float64x2x3_t __ret; \
   49056   __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
   49057   __ret; \
   49058 })
   49059 #else
   49060 #define vld1q_f64_x3(__p0) __extension__ ({ \
   49061   float64x2x3_t __ret; \
   49062   __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
   49063  \
   49064   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49065   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49066   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49067   __ret; \
   49068 })
   49069 #endif
   49070 
   49071 #ifdef __LITTLE_ENDIAN__
   49072 #define vld1q_f32_x3(__p0) __extension__ ({ \
   49073   float32x4x3_t __ret; \
   49074   __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
   49075   __ret; \
   49076 })
   49077 #else
   49078 #define vld1q_f32_x3(__p0) __extension__ ({ \
   49079   float32x4x3_t __ret; \
   49080   __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
   49081  \
   49082   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49083   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49084   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49085   __ret; \
   49086 })
   49087 #endif
   49088 
   49089 #ifdef __LITTLE_ENDIAN__
   49090 #define vld1q_f16_x3(__p0) __extension__ ({ \
   49091   float16x8x3_t __ret; \
   49092   __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
   49093   __ret; \
   49094 })
   49095 #else
   49096 #define vld1q_f16_x3(__p0) __extension__ ({ \
   49097   float16x8x3_t __ret; \
   49098   __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
   49099  \
   49100   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49101   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49102   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49103   __ret; \
   49104 })
   49105 #endif
   49106 
   49107 #ifdef __LITTLE_ENDIAN__
   49108 #define vld1q_s32_x3(__p0) __extension__ ({ \
   49109   int32x4x3_t __ret; \
   49110   __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
   49111   __ret; \
   49112 })
   49113 #else
   49114 #define vld1q_s32_x3(__p0) __extension__ ({ \
   49115   int32x4x3_t __ret; \
   49116   __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
   49117  \
   49118   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49119   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49120   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49121   __ret; \
   49122 })
   49123 #endif
   49124 
   49125 #ifdef __LITTLE_ENDIAN__
   49126 #define vld1q_s64_x3(__p0) __extension__ ({ \
   49127   int64x2x3_t __ret; \
   49128   __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
   49129   __ret; \
   49130 })
   49131 #else
   49132 #define vld1q_s64_x3(__p0) __extension__ ({ \
   49133   int64x2x3_t __ret; \
   49134   __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
   49135  \
   49136   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49137   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49138   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49139   __ret; \
   49140 })
   49141 #endif
   49142 
   49143 #ifdef __LITTLE_ENDIAN__
   49144 #define vld1q_s16_x3(__p0) __extension__ ({ \
   49145   int16x8x3_t __ret; \
   49146   __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
   49147   __ret; \
   49148 })
   49149 #else
   49150 #define vld1q_s16_x3(__p0) __extension__ ({ \
   49151   int16x8x3_t __ret; \
   49152   __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
   49153  \
   49154   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49155   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49156   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49157   __ret; \
   49158 })
   49159 #endif
   49160 
   49161 #ifdef __LITTLE_ENDIAN__
   49162 #define vld1_u8_x3(__p0) __extension__ ({ \
   49163   uint8x8x3_t __ret; \
   49164   __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
   49165   __ret; \
   49166 })
   49167 #else
   49168 #define vld1_u8_x3(__p0) __extension__ ({ \
   49169   uint8x8x3_t __ret; \
   49170   __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
   49171  \
   49172   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49173   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49174   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49175   __ret; \
   49176 })
   49177 #endif
   49178 
   49179 #ifdef __LITTLE_ENDIAN__
   49180 #define vld1_u32_x3(__p0) __extension__ ({ \
   49181   uint32x2x3_t __ret; \
   49182   __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
   49183   __ret; \
   49184 })
   49185 #else
   49186 #define vld1_u32_x3(__p0) __extension__ ({ \
   49187   uint32x2x3_t __ret; \
   49188   __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
   49189  \
   49190   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49191   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49192   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49193   __ret; \
   49194 })
   49195 #endif
   49196 
   49197 #ifdef __LITTLE_ENDIAN__
   49198 #define vld1_u64_x3(__p0) __extension__ ({ \
   49199   uint64x1x3_t __ret; \
   49200   __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
   49201   __ret; \
   49202 })
   49203 #else
   49204 #define vld1_u64_x3(__p0) __extension__ ({ \
   49205   uint64x1x3_t __ret; \
   49206   __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
   49207   __ret; \
   49208 })
   49209 #endif
   49210 
   49211 #ifdef __LITTLE_ENDIAN__
   49212 #define vld1_u16_x3(__p0) __extension__ ({ \
   49213   uint16x4x3_t __ret; \
   49214   __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
   49215   __ret; \
   49216 })
   49217 #else
   49218 #define vld1_u16_x3(__p0) __extension__ ({ \
   49219   uint16x4x3_t __ret; \
   49220   __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
   49221  \
   49222   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49223   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49224   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49225   __ret; \
   49226 })
   49227 #endif
   49228 
   49229 #ifdef __LITTLE_ENDIAN__
   49230 #define vld1_s8_x3(__p0) __extension__ ({ \
   49231   int8x8x3_t __ret; \
   49232   __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
   49233   __ret; \
   49234 })
   49235 #else
   49236 #define vld1_s8_x3(__p0) __extension__ ({ \
   49237   int8x8x3_t __ret; \
   49238   __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
   49239  \
   49240   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49241   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49242   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49243   __ret; \
   49244 })
   49245 #endif
   49246 
   49247 #ifdef __LITTLE_ENDIAN__
   49248 #define vld1_f64_x3(__p0) __extension__ ({ \
   49249   float64x1x3_t __ret; \
   49250   __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
   49251   __ret; \
   49252 })
   49253 #else
   49254 #define vld1_f64_x3(__p0) __extension__ ({ \
   49255   float64x1x3_t __ret; \
   49256   __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
   49257   __ret; \
   49258 })
   49259 #endif
   49260 
   49261 #ifdef __LITTLE_ENDIAN__
   49262 #define vld1_f32_x3(__p0) __extension__ ({ \
   49263   float32x2x3_t __ret; \
   49264   __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
   49265   __ret; \
   49266 })
   49267 #else
   49268 #define vld1_f32_x3(__p0) __extension__ ({ \
   49269   float32x2x3_t __ret; \
   49270   __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
   49271  \
   49272   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49273   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49274   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49275   __ret; \
   49276 })
   49277 #endif
   49278 
   49279 #ifdef __LITTLE_ENDIAN__
   49280 #define vld1_f16_x3(__p0) __extension__ ({ \
   49281   float16x4x3_t __ret; \
   49282   __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
   49283   __ret; \
   49284 })
   49285 #else
   49286 #define vld1_f16_x3(__p0) __extension__ ({ \
   49287   float16x4x3_t __ret; \
   49288   __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
   49289  \
   49290   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49291   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49292   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49293   __ret; \
   49294 })
   49295 #endif
   49296 
   49297 #ifdef __LITTLE_ENDIAN__
   49298 #define vld1_s32_x3(__p0) __extension__ ({ \
   49299   int32x2x3_t __ret; \
   49300   __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
   49301   __ret; \
   49302 })
   49303 #else
   49304 #define vld1_s32_x3(__p0) __extension__ ({ \
   49305   int32x2x3_t __ret; \
   49306   __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
   49307  \
   49308   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49309   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49310   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49311   __ret; \
   49312 })
   49313 #endif
   49314 
   49315 #ifdef __LITTLE_ENDIAN__
   49316 #define vld1_s64_x3(__p0) __extension__ ({ \
   49317   int64x1x3_t __ret; \
   49318   __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
   49319   __ret; \
   49320 })
   49321 #else
   49322 #define vld1_s64_x3(__p0) __extension__ ({ \
   49323   int64x1x3_t __ret; \
   49324   __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
   49325   __ret; \
   49326 })
   49327 #endif
   49328 
   49329 #ifdef __LITTLE_ENDIAN__
   49330 #define vld1_s16_x3(__p0) __extension__ ({ \
   49331   int16x4x3_t __ret; \
   49332   __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
   49333   __ret; \
   49334 })
   49335 #else
   49336 #define vld1_s16_x3(__p0) __extension__ ({ \
   49337   int16x4x3_t __ret; \
   49338   __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
   49339  \
   49340   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49341   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49342   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49343   __ret; \
   49344 })
   49345 #endif
   49346 
   49347 #ifdef __LITTLE_ENDIAN__
   49348 #define vld1_p8_x4(__p0) __extension__ ({ \
   49349   poly8x8x4_t __ret; \
   49350   __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
   49351   __ret; \
   49352 })
   49353 #else
   49354 #define vld1_p8_x4(__p0) __extension__ ({ \
   49355   poly8x8x4_t __ret; \
   49356   __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
   49357  \
   49358   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49359   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49360   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49361   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   49362   __ret; \
   49363 })
   49364 #endif
   49365 
   49366 #ifdef __LITTLE_ENDIAN__
   49367 #define vld1_p64_x4(__p0) __extension__ ({ \
   49368   poly64x1x4_t __ret; \
   49369   __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
   49370   __ret; \
   49371 })
   49372 #else
   49373 #define vld1_p64_x4(__p0) __extension__ ({ \
   49374   poly64x1x4_t __ret; \
   49375   __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
   49376   __ret; \
   49377 })
   49378 #endif
   49379 
   49380 #ifdef __LITTLE_ENDIAN__
   49381 #define vld1_p16_x4(__p0) __extension__ ({ \
   49382   poly16x4x4_t __ret; \
   49383   __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
   49384   __ret; \
   49385 })
   49386 #else
   49387 #define vld1_p16_x4(__p0) __extension__ ({ \
   49388   poly16x4x4_t __ret; \
   49389   __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
   49390  \
   49391   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49392   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49393   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49394   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   49395   __ret; \
   49396 })
   49397 #endif
   49398 
   49399 #ifdef __LITTLE_ENDIAN__
   49400 #define vld1q_p8_x4(__p0) __extension__ ({ \
   49401   poly8x16x4_t __ret; \
   49402   __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
   49403   __ret; \
   49404 })
   49405 #else
   49406 #define vld1q_p8_x4(__p0) __extension__ ({ \
   49407   poly8x16x4_t __ret; \
   49408   __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
   49409  \
   49410   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49411   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49412   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49413   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49414   __ret; \
   49415 })
   49416 #endif
   49417 
   49418 #ifdef __LITTLE_ENDIAN__
   49419 #define vld1q_p64_x4(__p0) __extension__ ({ \
   49420   poly64x2x4_t __ret; \
   49421   __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
   49422   __ret; \
   49423 })
   49424 #else
   49425 #define vld1q_p64_x4(__p0) __extension__ ({ \
   49426   poly64x2x4_t __ret; \
   49427   __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
   49428  \
   49429   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49430   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49431   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49432   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   49433   __ret; \
   49434 })
   49435 #endif
   49436 
   49437 #ifdef __LITTLE_ENDIAN__
   49438 #define vld1q_p16_x4(__p0) __extension__ ({ \
   49439   poly16x8x4_t __ret; \
   49440   __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
   49441   __ret; \
   49442 })
   49443 #else
   49444 #define vld1q_p16_x4(__p0) __extension__ ({ \
   49445   poly16x8x4_t __ret; \
   49446   __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
   49447  \
   49448   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49449   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49450   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49451   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   49452   __ret; \
   49453 })
   49454 #endif
   49455 
   49456 #ifdef __LITTLE_ENDIAN__
   49457 #define vld1q_u8_x4(__p0) __extension__ ({ \
   49458   uint8x16x4_t __ret; \
   49459   __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
   49460   __ret; \
   49461 })
   49462 #else
   49463 #define vld1q_u8_x4(__p0) __extension__ ({ \
   49464   uint8x16x4_t __ret; \
   49465   __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
   49466  \
   49467   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49468   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49469   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49470   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49471   __ret; \
   49472 })
   49473 #endif
   49474 
   49475 #ifdef __LITTLE_ENDIAN__
   49476 #define vld1q_u32_x4(__p0) __extension__ ({ \
   49477   uint32x4x4_t __ret; \
   49478   __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
   49479   __ret; \
   49480 })
   49481 #else
   49482 #define vld1q_u32_x4(__p0) __extension__ ({ \
   49483   uint32x4x4_t __ret; \
   49484   __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
   49485  \
   49486   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49487   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49488   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49489   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   49490   __ret; \
   49491 })
   49492 #endif
   49493 
   49494 #ifdef __LITTLE_ENDIAN__
   49495 #define vld1q_u64_x4(__p0) __extension__ ({ \
   49496   uint64x2x4_t __ret; \
   49497   __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
   49498   __ret; \
   49499 })
   49500 #else
   49501 #define vld1q_u64_x4(__p0) __extension__ ({ \
   49502   uint64x2x4_t __ret; \
   49503   __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
   49504  \
   49505   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49506   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49507   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49508   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   49509   __ret; \
   49510 })
   49511 #endif
   49512 
   49513 #ifdef __LITTLE_ENDIAN__
   49514 #define vld1q_u16_x4(__p0) __extension__ ({ \
   49515   uint16x8x4_t __ret; \
   49516   __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
   49517   __ret; \
   49518 })
   49519 #else
   49520 #define vld1q_u16_x4(__p0) __extension__ ({ \
   49521   uint16x8x4_t __ret; \
   49522   __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
   49523  \
   49524   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49525   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49526   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49527   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   49528   __ret; \
   49529 })
   49530 #endif
   49531 
   49532 #ifdef __LITTLE_ENDIAN__
   49533 #define vld1q_s8_x4(__p0) __extension__ ({ \
   49534   int8x16x4_t __ret; \
   49535   __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
   49536   __ret; \
   49537 })
   49538 #else
   49539 #define vld1q_s8_x4(__p0) __extension__ ({ \
   49540   int8x16x4_t __ret; \
   49541   __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
   49542  \
   49543   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49544   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49545   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49546   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49547   __ret; \
   49548 })
   49549 #endif
   49550 
   49551 #ifdef __LITTLE_ENDIAN__
   49552 #define vld1q_f64_x4(__p0) __extension__ ({ \
   49553   float64x2x4_t __ret; \
   49554   __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
   49555   __ret; \
   49556 })
   49557 #else
   49558 #define vld1q_f64_x4(__p0) __extension__ ({ \
   49559   float64x2x4_t __ret; \
   49560   __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
   49561  \
   49562   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49563   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49564   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49565   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   49566   __ret; \
   49567 })
   49568 #endif
   49569 
   49570 #ifdef __LITTLE_ENDIAN__
   49571 #define vld1q_f32_x4(__p0) __extension__ ({ \
   49572   float32x4x4_t __ret; \
   49573   __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
   49574   __ret; \
   49575 })
   49576 #else
   49577 #define vld1q_f32_x4(__p0) __extension__ ({ \
   49578   float32x4x4_t __ret; \
   49579   __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
   49580  \
   49581   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49582   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49583   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49584   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   49585   __ret; \
   49586 })
   49587 #endif
   49588 
   49589 #ifdef __LITTLE_ENDIAN__
   49590 #define vld1q_f16_x4(__p0) __extension__ ({ \
   49591   float16x8x4_t __ret; \
   49592   __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
   49593   __ret; \
   49594 })
   49595 #else
   49596 #define vld1q_f16_x4(__p0) __extension__ ({ \
   49597   float16x8x4_t __ret; \
   49598   __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
   49599  \
   49600   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49601   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49602   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49603   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   49604   __ret; \
   49605 })
   49606 #endif
   49607 
   49608 #ifdef __LITTLE_ENDIAN__
   49609 #define vld1q_s32_x4(__p0) __extension__ ({ \
   49610   int32x4x4_t __ret; \
   49611   __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
   49612   __ret; \
   49613 })
   49614 #else
   49615 #define vld1q_s32_x4(__p0) __extension__ ({ \
   49616   int32x4x4_t __ret; \
   49617   __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
   49618  \
   49619   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49620   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49621   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49622   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   49623   __ret; \
   49624 })
   49625 #endif
   49626 
   49627 #ifdef __LITTLE_ENDIAN__
   49628 #define vld1q_s64_x4(__p0) __extension__ ({ \
   49629   int64x2x4_t __ret; \
   49630   __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
   49631   __ret; \
   49632 })
   49633 #else
   49634 #define vld1q_s64_x4(__p0) __extension__ ({ \
   49635   int64x2x4_t __ret; \
   49636   __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
   49637  \
   49638   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49639   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49640   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49641   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   49642   __ret; \
   49643 })
   49644 #endif
   49645 
   49646 #ifdef __LITTLE_ENDIAN__
   49647 #define vld1q_s16_x4(__p0) __extension__ ({ \
   49648   int16x8x4_t __ret; \
   49649   __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
   49650   __ret; \
   49651 })
   49652 #else
   49653 #define vld1q_s16_x4(__p0) __extension__ ({ \
   49654   int16x8x4_t __ret; \
   49655   __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
   49656  \
   49657   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49658   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49659   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49660   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   49661   __ret; \
   49662 })
   49663 #endif
   49664 
   49665 #ifdef __LITTLE_ENDIAN__
   49666 #define vld1_u8_x4(__p0) __extension__ ({ \
   49667   uint8x8x4_t __ret; \
   49668   __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
   49669   __ret; \
   49670 })
   49671 #else
   49672 #define vld1_u8_x4(__p0) __extension__ ({ \
   49673   uint8x8x4_t __ret; \
   49674   __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
   49675  \
   49676   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49677   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49678   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49679   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   49680   __ret; \
   49681 })
   49682 #endif
   49683 
   49684 #ifdef __LITTLE_ENDIAN__
   49685 #define vld1_u32_x4(__p0) __extension__ ({ \
   49686   uint32x2x4_t __ret; \
   49687   __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
   49688   __ret; \
   49689 })
   49690 #else
   49691 #define vld1_u32_x4(__p0) __extension__ ({ \
   49692   uint32x2x4_t __ret; \
   49693   __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
   49694  \
   49695   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49696   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49697   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49698   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   49699   __ret; \
   49700 })
   49701 #endif
   49702 
   49703 #ifdef __LITTLE_ENDIAN__
   49704 #define vld1_u64_x4(__p0) __extension__ ({ \
   49705   uint64x1x4_t __ret; \
   49706   __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
   49707   __ret; \
   49708 })
   49709 #else
   49710 #define vld1_u64_x4(__p0) __extension__ ({ \
   49711   uint64x1x4_t __ret; \
   49712   __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
   49713   __ret; \
   49714 })
   49715 #endif
   49716 
   49717 #ifdef __LITTLE_ENDIAN__
   49718 #define vld1_u16_x4(__p0) __extension__ ({ \
   49719   uint16x4x4_t __ret; \
   49720   __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
   49721   __ret; \
   49722 })
   49723 #else
   49724 #define vld1_u16_x4(__p0) __extension__ ({ \
   49725   uint16x4x4_t __ret; \
   49726   __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
   49727  \
   49728   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49729   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49730   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49731   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   49732   __ret; \
   49733 })
   49734 #endif
   49735 
   49736 #ifdef __LITTLE_ENDIAN__
   49737 #define vld1_s8_x4(__p0) __extension__ ({ \
   49738   int8x8x4_t __ret; \
   49739   __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
   49740   __ret; \
   49741 })
   49742 #else
   49743 #define vld1_s8_x4(__p0) __extension__ ({ \
   49744   int8x8x4_t __ret; \
   49745   __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
   49746  \
   49747   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   49748   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   49749   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   49750   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   49751   __ret; \
   49752 })
   49753 #endif
   49754 
   49755 #ifdef __LITTLE_ENDIAN__
   49756 #define vld1_f64_x4(__p0) __extension__ ({ \
   49757   float64x1x4_t __ret; \
   49758   __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
   49759   __ret; \
   49760 })
   49761 #else
   49762 #define vld1_f64_x4(__p0) __extension__ ({ \
   49763   float64x1x4_t __ret; \
   49764   __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
   49765   __ret; \
   49766 })
   49767 #endif
   49768 
   49769 #ifdef __LITTLE_ENDIAN__
   49770 #define vld1_f32_x4(__p0) __extension__ ({ \
   49771   float32x2x4_t __ret; \
   49772   __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
   49773   __ret; \
   49774 })
   49775 #else
   49776 #define vld1_f32_x4(__p0) __extension__ ({ \
   49777   float32x2x4_t __ret; \
   49778   __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
   49779  \
   49780   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49781   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49782   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49783   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   49784   __ret; \
   49785 })
   49786 #endif
   49787 
   49788 #ifdef __LITTLE_ENDIAN__
   49789 #define vld1_f16_x4(__p0) __extension__ ({ \
   49790   float16x4x4_t __ret; \
   49791   __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
   49792   __ret; \
   49793 })
   49794 #else
   49795 #define vld1_f16_x4(__p0) __extension__ ({ \
   49796   float16x4x4_t __ret; \
   49797   __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
   49798  \
   49799   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49800   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49801   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49802   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   49803   __ret; \
   49804 })
   49805 #endif
   49806 
   49807 #ifdef __LITTLE_ENDIAN__
   49808 #define vld1_s32_x4(__p0) __extension__ ({ \
   49809   int32x2x4_t __ret; \
   49810   __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
   49811   __ret; \
   49812 })
   49813 #else
   49814 #define vld1_s32_x4(__p0) __extension__ ({ \
   49815   int32x2x4_t __ret; \
   49816   __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
   49817  \
   49818   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49819   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49820   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   49821   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   49822   __ret; \
   49823 })
   49824 #endif
   49825 
   49826 #ifdef __LITTLE_ENDIAN__
   49827 #define vld1_s64_x4(__p0) __extension__ ({ \
   49828   int64x1x4_t __ret; \
   49829   __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
   49830   __ret; \
   49831 })
   49832 #else
   49833 #define vld1_s64_x4(__p0) __extension__ ({ \
   49834   int64x1x4_t __ret; \
   49835   __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
   49836   __ret; \
   49837 })
   49838 #endif
   49839 
   49840 #ifdef __LITTLE_ENDIAN__
   49841 #define vld1_s16_x4(__p0) __extension__ ({ \
   49842   int16x4x4_t __ret; \
   49843   __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
   49844   __ret; \
   49845 })
   49846 #else
   49847 #define vld1_s16_x4(__p0) __extension__ ({ \
   49848   int16x4x4_t __ret; \
   49849   __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
   49850  \
   49851   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   49852   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   49853   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   49854   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   49855   __ret; \
   49856 })
   49857 #endif
   49858 
   49859 #ifdef __LITTLE_ENDIAN__
   49860 #define vld2_p64(__p0) __extension__ ({ \
   49861   poly64x1x2_t __ret; \
   49862   __builtin_neon_vld2_v(&__ret, __p0, 6); \
   49863   __ret; \
   49864 })
   49865 #else
   49866 #define vld2_p64(__p0) __extension__ ({ \
   49867   poly64x1x2_t __ret; \
   49868   __builtin_neon_vld2_v(&__ret, __p0, 6); \
   49869   __ret; \
   49870 })
   49871 #endif
   49872 
   49873 #ifdef __LITTLE_ENDIAN__
   49874 #define vld2q_p64(__p0) __extension__ ({ \
   49875   poly64x2x2_t __ret; \
   49876   __builtin_neon_vld2q_v(&__ret, __p0, 38); \
   49877   __ret; \
   49878 })
   49879 #else
   49880 #define vld2q_p64(__p0) __extension__ ({ \
   49881   poly64x2x2_t __ret; \
   49882   __builtin_neon_vld2q_v(&__ret, __p0, 38); \
   49883  \
   49884   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49885   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49886   __ret; \
   49887 })
   49888 #endif
   49889 
   49890 #ifdef __LITTLE_ENDIAN__
   49891 #define vld2q_u64(__p0) __extension__ ({ \
   49892   uint64x2x2_t __ret; \
   49893   __builtin_neon_vld2q_v(&__ret, __p0, 51); \
   49894   __ret; \
   49895 })
   49896 #else
   49897 #define vld2q_u64(__p0) __extension__ ({ \
   49898   uint64x2x2_t __ret; \
   49899   __builtin_neon_vld2q_v(&__ret, __p0, 51); \
   49900  \
   49901   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49902   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49903   __ret; \
   49904 })
   49905 #endif
   49906 
   49907 #ifdef __LITTLE_ENDIAN__
   49908 #define vld2q_f64(__p0) __extension__ ({ \
   49909   float64x2x2_t __ret; \
   49910   __builtin_neon_vld2q_v(&__ret, __p0, 42); \
   49911   __ret; \
   49912 })
   49913 #else
   49914 #define vld2q_f64(__p0) __extension__ ({ \
   49915   float64x2x2_t __ret; \
   49916   __builtin_neon_vld2q_v(&__ret, __p0, 42); \
   49917  \
   49918   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49919   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49920   __ret; \
   49921 })
   49922 #endif
   49923 
   49924 #ifdef __LITTLE_ENDIAN__
   49925 #define vld2q_s64(__p0) __extension__ ({ \
   49926   int64x2x2_t __ret; \
   49927   __builtin_neon_vld2q_v(&__ret, __p0, 35); \
   49928   __ret; \
   49929 })
   49930 #else
   49931 #define vld2q_s64(__p0) __extension__ ({ \
   49932   int64x2x2_t __ret; \
   49933   __builtin_neon_vld2q_v(&__ret, __p0, 35); \
   49934  \
   49935   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49936   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49937   __ret; \
   49938 })
   49939 #endif
   49940 
   49941 #ifdef __LITTLE_ENDIAN__
   49942 #define vld2_f64(__p0) __extension__ ({ \
   49943   float64x1x2_t __ret; \
   49944   __builtin_neon_vld2_v(&__ret, __p0, 10); \
   49945   __ret; \
   49946 })
   49947 #else
   49948 #define vld2_f64(__p0) __extension__ ({ \
   49949   float64x1x2_t __ret; \
   49950   __builtin_neon_vld2_v(&__ret, __p0, 10); \
   49951   __ret; \
   49952 })
   49953 #endif
   49954 
   49955 #ifdef __LITTLE_ENDIAN__
   49956 #define vld2_dup_p64(__p0) __extension__ ({ \
   49957   poly64x1x2_t __ret; \
   49958   __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
   49959   __ret; \
   49960 })
   49961 #else
   49962 #define vld2_dup_p64(__p0) __extension__ ({ \
   49963   poly64x1x2_t __ret; \
   49964   __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
   49965   __ret; \
   49966 })
   49967 #endif
   49968 
   49969 #ifdef __LITTLE_ENDIAN__
   49970 #define vld2q_dup_p8(__p0) __extension__ ({ \
   49971   poly8x16x2_t __ret; \
   49972   __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
   49973   __ret; \
   49974 })
   49975 #else
   49976 #define vld2q_dup_p8(__p0) __extension__ ({ \
   49977   poly8x16x2_t __ret; \
   49978   __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
   49979  \
   49980   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49981   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   49982   __ret; \
   49983 })
   49984 #endif
   49985 
   49986 #ifdef __LITTLE_ENDIAN__
   49987 #define vld2q_dup_p64(__p0) __extension__ ({ \
   49988   poly64x2x2_t __ret; \
   49989   __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
   49990   __ret; \
   49991 })
   49992 #else
   49993 #define vld2q_dup_p64(__p0) __extension__ ({ \
   49994   poly64x2x2_t __ret; \
   49995   __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
   49996  \
   49997   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   49998   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   49999   __ret; \
   50000 })
   50001 #endif
   50002 
   50003 #ifdef __LITTLE_ENDIAN__
   50004 #define vld2q_dup_p16(__p0) __extension__ ({ \
   50005   poly16x8x2_t __ret; \
   50006   __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
   50007   __ret; \
   50008 })
   50009 #else
   50010 #define vld2q_dup_p16(__p0) __extension__ ({ \
   50011   poly16x8x2_t __ret; \
   50012   __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
   50013  \
   50014   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   50015   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   50016   __ret; \
   50017 })
   50018 #endif
   50019 
   50020 #ifdef __LITTLE_ENDIAN__
   50021 #define vld2q_dup_u8(__p0) __extension__ ({ \
   50022   uint8x16x2_t __ret; \
   50023   __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
   50024   __ret; \
   50025 })
   50026 #else
   50027 #define vld2q_dup_u8(__p0) __extension__ ({ \
   50028   uint8x16x2_t __ret; \
   50029   __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
   50030  \
   50031   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50032   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50033   __ret; \
   50034 })
   50035 #endif
   50036 
   50037 #ifdef __LITTLE_ENDIAN__
   50038 #define vld2q_dup_u32(__p0) __extension__ ({ \
   50039   uint32x4x2_t __ret; \
   50040   __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
   50041   __ret; \
   50042 })
   50043 #else
   50044 #define vld2q_dup_u32(__p0) __extension__ ({ \
   50045   uint32x4x2_t __ret; \
   50046   __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
   50047  \
   50048   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   50049   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   50050   __ret; \
   50051 })
   50052 #endif
   50053 
   50054 #ifdef __LITTLE_ENDIAN__
   50055 #define vld2q_dup_u64(__p0) __extension__ ({ \
   50056   uint64x2x2_t __ret; \
   50057   __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
   50058   __ret; \
   50059 })
   50060 #else
   50061 #define vld2q_dup_u64(__p0) __extension__ ({ \
   50062   uint64x2x2_t __ret; \
   50063   __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
   50064  \
   50065   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50066   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50067   __ret; \
   50068 })
   50069 #endif
   50070 
   50071 #ifdef __LITTLE_ENDIAN__
   50072 #define vld2q_dup_u16(__p0) __extension__ ({ \
   50073   uint16x8x2_t __ret; \
   50074   __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
   50075   __ret; \
   50076 })
   50077 #else
   50078 #define vld2q_dup_u16(__p0) __extension__ ({ \
   50079   uint16x8x2_t __ret; \
   50080   __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
   50081  \
   50082   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   50083   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   50084   __ret; \
   50085 })
   50086 #endif
   50087 
   50088 #ifdef __LITTLE_ENDIAN__
   50089 #define vld2q_dup_s8(__p0) __extension__ ({ \
   50090   int8x16x2_t __ret; \
   50091   __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
   50092   __ret; \
   50093 })
   50094 #else
   50095 #define vld2q_dup_s8(__p0) __extension__ ({ \
   50096   int8x16x2_t __ret; \
   50097   __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
   50098  \
   50099   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50100   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50101   __ret; \
   50102 })
   50103 #endif
   50104 
   50105 #ifdef __LITTLE_ENDIAN__
   50106 #define vld2q_dup_f64(__p0) __extension__ ({ \
   50107   float64x2x2_t __ret; \
   50108   __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
   50109   __ret; \
   50110 })
   50111 #else
   50112 #define vld2q_dup_f64(__p0) __extension__ ({ \
   50113   float64x2x2_t __ret; \
   50114   __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
   50115  \
   50116   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50117   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50118   __ret; \
   50119 })
   50120 #endif
   50121 
   50122 #ifdef __LITTLE_ENDIAN__
   50123 #define vld2q_dup_f32(__p0) __extension__ ({ \
   50124   float32x4x2_t __ret; \
   50125   __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
   50126   __ret; \
   50127 })
   50128 #else
   50129 #define vld2q_dup_f32(__p0) __extension__ ({ \
   50130   float32x4x2_t __ret; \
   50131   __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
   50132  \
   50133   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   50134   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   50135   __ret; \
   50136 })
   50137 #endif
   50138 
   50139 #ifdef __LITTLE_ENDIAN__
   50140 #define vld2q_dup_f16(__p0) __extension__ ({ \
   50141   float16x8x2_t __ret; \
   50142   __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
   50143   __ret; \
   50144 })
   50145 #else
   50146 #define vld2q_dup_f16(__p0) __extension__ ({ \
   50147   float16x8x2_t __ret; \
   50148   __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
   50149  \
   50150   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   50151   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   50152   __ret; \
   50153 })
   50154 #endif
   50155 
   50156 #ifdef __LITTLE_ENDIAN__
   50157 #define vld2q_dup_s32(__p0) __extension__ ({ \
   50158   int32x4x2_t __ret; \
   50159   __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
   50160   __ret; \
   50161 })
   50162 #else
   50163 #define vld2q_dup_s32(__p0) __extension__ ({ \
   50164   int32x4x2_t __ret; \
   50165   __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
   50166  \
   50167   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   50168   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   50169   __ret; \
   50170 })
   50171 #endif
   50172 
   50173 #ifdef __LITTLE_ENDIAN__
   50174 #define vld2q_dup_s64(__p0) __extension__ ({ \
   50175   int64x2x2_t __ret; \
   50176   __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
   50177   __ret; \
   50178 })
   50179 #else
   50180 #define vld2q_dup_s64(__p0) __extension__ ({ \
   50181   int64x2x2_t __ret; \
   50182   __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
   50183  \
   50184   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50185   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50186   __ret; \
   50187 })
   50188 #endif
   50189 
   50190 #ifdef __LITTLE_ENDIAN__
   50191 #define vld2q_dup_s16(__p0) __extension__ ({ \
   50192   int16x8x2_t __ret; \
   50193   __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
   50194   __ret; \
   50195 })
   50196 #else
   50197 #define vld2q_dup_s16(__p0) __extension__ ({ \
   50198   int16x8x2_t __ret; \
   50199   __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
   50200  \
   50201   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   50202   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   50203   __ret; \
   50204 })
   50205 #endif
   50206 
   50207 #ifdef __LITTLE_ENDIAN__
   50208 #define vld2_dup_f64(__p0) __extension__ ({ \
   50209   float64x1x2_t __ret; \
   50210   __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
   50211   __ret; \
   50212 })
   50213 #else
   50214 #define vld2_dup_f64(__p0) __extension__ ({ \
   50215   float64x1x2_t __ret; \
   50216   __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
   50217   __ret; \
   50218 })
   50219 #endif
   50220 
   50221 #ifdef __LITTLE_ENDIAN__
   50222 #define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   50223   poly64x1x2_t __s1 = __p1; \
   50224   poly64x1x2_t __ret; \
   50225   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
   50226   __ret; \
   50227 })
   50228 #else
   50229 #define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   50230   poly64x1x2_t __s1 = __p1; \
   50231   poly64x1x2_t __ret; \
   50232   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
   50233   __ret; \
   50234 })
   50235 #endif
   50236 
   50237 #ifdef __LITTLE_ENDIAN__
   50238 #define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   50239   poly8x16x2_t __s1 = __p1; \
   50240   poly8x16x2_t __ret; \
   50241   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
   50242   __ret; \
   50243 })
   50244 #else
   50245 #define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   50246   poly8x16x2_t __s1 = __p1; \
   50247   poly8x16x2_t __rev1; \
   50248   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50249   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50250   poly8x16x2_t __ret; \
   50251   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
   50252  \
   50253   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50254   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50255   __ret; \
   50256 })
   50257 #endif
   50258 
   50259 #ifdef __LITTLE_ENDIAN__
   50260 #define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   50261   poly64x2x2_t __s1 = __p1; \
   50262   poly64x2x2_t __ret; \
   50263   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
   50264   __ret; \
   50265 })
   50266 #else
   50267 #define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   50268   poly64x2x2_t __s1 = __p1; \
   50269   poly64x2x2_t __rev1; \
   50270   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   50271   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   50272   poly64x2x2_t __ret; \
   50273   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
   50274  \
   50275   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50276   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50277   __ret; \
   50278 })
   50279 #endif
   50280 
   50281 #ifdef __LITTLE_ENDIAN__
   50282 #define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   50283   uint8x16x2_t __s1 = __p1; \
   50284   uint8x16x2_t __ret; \
   50285   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
   50286   __ret; \
   50287 })
   50288 #else
   50289 #define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   50290   uint8x16x2_t __s1 = __p1; \
   50291   uint8x16x2_t __rev1; \
   50292   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50293   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50294   uint8x16x2_t __ret; \
   50295   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
   50296  \
   50297   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50298   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50299   __ret; \
   50300 })
   50301 #endif
   50302 
   50303 #ifdef __LITTLE_ENDIAN__
   50304 #define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   50305   uint64x2x2_t __s1 = __p1; \
   50306   uint64x2x2_t __ret; \
   50307   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
   50308   __ret; \
   50309 })
   50310 #else
   50311 #define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   50312   uint64x2x2_t __s1 = __p1; \
   50313   uint64x2x2_t __rev1; \
   50314   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   50315   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   50316   uint64x2x2_t __ret; \
   50317   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
   50318  \
   50319   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50320   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50321   __ret; \
   50322 })
   50323 #endif
   50324 
   50325 #ifdef __LITTLE_ENDIAN__
   50326 #define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   50327   int8x16x2_t __s1 = __p1; \
   50328   int8x16x2_t __ret; \
   50329   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
   50330   __ret; \
   50331 })
   50332 #else
   50333 #define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   50334   int8x16x2_t __s1 = __p1; \
   50335   int8x16x2_t __rev1; \
   50336   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50337   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50338   int8x16x2_t __ret; \
   50339   __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
   50340  \
   50341   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50342   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50343   __ret; \
   50344 })
   50345 #endif
   50346 
   50347 #ifdef __LITTLE_ENDIAN__
   50348 #define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   50349   float64x2x2_t __s1 = __p1; \
   50350   float64x2x2_t __ret; \
   50351   __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 42); \
   50352   __ret; \
   50353 })
   50354 #else
   50355 #define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   50356   float64x2x2_t __s1 = __p1; \
   50357   float64x2x2_t __rev1; \
   50358   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   50359   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   50360   float64x2x2_t __ret; \
   50361   __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
   50362  \
   50363   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50364   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50365   __ret; \
   50366 })
   50367 #endif
   50368 
   50369 #ifdef __LITTLE_ENDIAN__
   50370 #define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   50371   int64x2x2_t __s1 = __p1; \
   50372   int64x2x2_t __ret; \
   50373   __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 35); \
   50374   __ret; \
   50375 })
   50376 #else
   50377 #define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   50378   int64x2x2_t __s1 = __p1; \
   50379   int64x2x2_t __rev1; \
   50380   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   50381   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   50382   int64x2x2_t __ret; \
   50383   __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
   50384  \
   50385   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50386   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50387   __ret; \
   50388 })
   50389 #endif
   50390 
   50391 #ifdef __LITTLE_ENDIAN__
   50392 #define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   50393   uint64x1x2_t __s1 = __p1; \
   50394   uint64x1x2_t __ret; \
   50395   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
   50396   __ret; \
   50397 })
   50398 #else
   50399 #define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   50400   uint64x1x2_t __s1 = __p1; \
   50401   uint64x1x2_t __ret; \
   50402   __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
   50403   __ret; \
   50404 })
   50405 #endif
   50406 
   50407 #ifdef __LITTLE_ENDIAN__
   50408 #define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   50409   float64x1x2_t __s1 = __p1; \
   50410   float64x1x2_t __ret; \
   50411   __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
   50412   __ret; \
   50413 })
   50414 #else
   50415 #define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   50416   float64x1x2_t __s1 = __p1; \
   50417   float64x1x2_t __ret; \
   50418   __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
   50419   __ret; \
   50420 })
   50421 #endif
   50422 
   50423 #ifdef __LITTLE_ENDIAN__
   50424 #define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   50425   int64x1x2_t __s1 = __p1; \
   50426   int64x1x2_t __ret; \
   50427   __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
   50428   __ret; \
   50429 })
   50430 #else
   50431 #define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   50432   int64x1x2_t __s1 = __p1; \
   50433   int64x1x2_t __ret; \
   50434   __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
   50435   __ret; \
   50436 })
   50437 #endif
   50438 
   50439 #ifdef __LITTLE_ENDIAN__
   50440 #define vld3_p64(__p0) __extension__ ({ \
   50441   poly64x1x3_t __ret; \
   50442   __builtin_neon_vld3_v(&__ret, __p0, 6); \
   50443   __ret; \
   50444 })
   50445 #else
   50446 #define vld3_p64(__p0) __extension__ ({ \
   50447   poly64x1x3_t __ret; \
   50448   __builtin_neon_vld3_v(&__ret, __p0, 6); \
   50449   __ret; \
   50450 })
   50451 #endif
   50452 
   50453 #ifdef __LITTLE_ENDIAN__
   50454 #define vld3q_p64(__p0) __extension__ ({ \
   50455   poly64x2x3_t __ret; \
   50456   __builtin_neon_vld3q_v(&__ret, __p0, 38); \
   50457   __ret; \
   50458 })
   50459 #else
   50460 #define vld3q_p64(__p0) __extension__ ({ \
   50461   poly64x2x3_t __ret; \
   50462   __builtin_neon_vld3q_v(&__ret, __p0, 38); \
   50463  \
   50464   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50465   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50466   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50467   __ret; \
   50468 })
   50469 #endif
   50470 
   50471 #ifdef __LITTLE_ENDIAN__
   50472 #define vld3q_u64(__p0) __extension__ ({ \
   50473   uint64x2x3_t __ret; \
   50474   __builtin_neon_vld3q_v(&__ret, __p0, 51); \
   50475   __ret; \
   50476 })
   50477 #else
   50478 #define vld3q_u64(__p0) __extension__ ({ \
   50479   uint64x2x3_t __ret; \
   50480   __builtin_neon_vld3q_v(&__ret, __p0, 51); \
   50481  \
   50482   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50483   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50484   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50485   __ret; \
   50486 })
   50487 #endif
   50488 
   50489 #ifdef __LITTLE_ENDIAN__
   50490 #define vld3q_f64(__p0) __extension__ ({ \
   50491   float64x2x3_t __ret; \
   50492   __builtin_neon_vld3q_v(&__ret, __p0, 42); \
   50493   __ret; \
   50494 })
   50495 #else
   50496 #define vld3q_f64(__p0) __extension__ ({ \
   50497   float64x2x3_t __ret; \
   50498   __builtin_neon_vld3q_v(&__ret, __p0, 42); \
   50499  \
   50500   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50501   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50502   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50503   __ret; \
   50504 })
   50505 #endif
   50506 
   50507 #ifdef __LITTLE_ENDIAN__
   50508 #define vld3q_s64(__p0) __extension__ ({ \
   50509   int64x2x3_t __ret; \
   50510   __builtin_neon_vld3q_v(&__ret, __p0, 35); \
   50511   __ret; \
   50512 })
   50513 #else
   50514 #define vld3q_s64(__p0) __extension__ ({ \
   50515   int64x2x3_t __ret; \
   50516   __builtin_neon_vld3q_v(&__ret, __p0, 35); \
   50517  \
   50518   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50519   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50520   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50521   __ret; \
   50522 })
   50523 #endif
   50524 
   50525 #ifdef __LITTLE_ENDIAN__
   50526 #define vld3_f64(__p0) __extension__ ({ \
   50527   float64x1x3_t __ret; \
   50528   __builtin_neon_vld3_v(&__ret, __p0, 10); \
   50529   __ret; \
   50530 })
   50531 #else
   50532 #define vld3_f64(__p0) __extension__ ({ \
   50533   float64x1x3_t __ret; \
   50534   __builtin_neon_vld3_v(&__ret, __p0, 10); \
   50535   __ret; \
   50536 })
   50537 #endif
   50538 
   50539 #ifdef __LITTLE_ENDIAN__
   50540 #define vld3_dup_p64(__p0) __extension__ ({ \
   50541   poly64x1x3_t __ret; \
   50542   __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
   50543   __ret; \
   50544 })
   50545 #else
   50546 #define vld3_dup_p64(__p0) __extension__ ({ \
   50547   poly64x1x3_t __ret; \
   50548   __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
   50549   __ret; \
   50550 })
   50551 #endif
   50552 
   50553 #ifdef __LITTLE_ENDIAN__
   50554 #define vld3q_dup_p8(__p0) __extension__ ({ \
   50555   poly8x16x3_t __ret; \
   50556   __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
   50557   __ret; \
   50558 })
   50559 #else
   50560 #define vld3q_dup_p8(__p0) __extension__ ({ \
   50561   poly8x16x3_t __ret; \
   50562   __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
   50563  \
   50564   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50565   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50566   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50567   __ret; \
   50568 })
   50569 #endif
   50570 
   50571 #ifdef __LITTLE_ENDIAN__
   50572 #define vld3q_dup_p64(__p0) __extension__ ({ \
   50573   poly64x2x3_t __ret; \
   50574   __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
   50575   __ret; \
   50576 })
   50577 #else
   50578 #define vld3q_dup_p64(__p0) __extension__ ({ \
   50579   poly64x2x3_t __ret; \
   50580   __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
   50581  \
   50582   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50583   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50584   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50585   __ret; \
   50586 })
   50587 #endif
   50588 
   50589 #ifdef __LITTLE_ENDIAN__
   50590 #define vld3q_dup_p16(__p0) __extension__ ({ \
   50591   poly16x8x3_t __ret; \
   50592   __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
   50593   __ret; \
   50594 })
   50595 #else
   50596 #define vld3q_dup_p16(__p0) __extension__ ({ \
   50597   poly16x8x3_t __ret; \
   50598   __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
   50599  \
   50600   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   50601   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   50602   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   50603   __ret; \
   50604 })
   50605 #endif
   50606 
   50607 #ifdef __LITTLE_ENDIAN__
   50608 #define vld3q_dup_u8(__p0) __extension__ ({ \
   50609   uint8x16x3_t __ret; \
   50610   __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
   50611   __ret; \
   50612 })
   50613 #else
   50614 #define vld3q_dup_u8(__p0) __extension__ ({ \
   50615   uint8x16x3_t __ret; \
   50616   __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
   50617  \
   50618   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50619   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50620   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50621   __ret; \
   50622 })
   50623 #endif
   50624 
   50625 #ifdef __LITTLE_ENDIAN__
   50626 #define vld3q_dup_u32(__p0) __extension__ ({ \
   50627   uint32x4x3_t __ret; \
   50628   __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
   50629   __ret; \
   50630 })
   50631 #else
   50632 #define vld3q_dup_u32(__p0) __extension__ ({ \
   50633   uint32x4x3_t __ret; \
   50634   __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
   50635  \
   50636   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   50637   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   50638   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   50639   __ret; \
   50640 })
   50641 #endif
   50642 
   50643 #ifdef __LITTLE_ENDIAN__
   50644 #define vld3q_dup_u64(__p0) __extension__ ({ \
   50645   uint64x2x3_t __ret; \
   50646   __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
   50647   __ret; \
   50648 })
   50649 #else
   50650 #define vld3q_dup_u64(__p0) __extension__ ({ \
   50651   uint64x2x3_t __ret; \
   50652   __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
   50653  \
   50654   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50655   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50656   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50657   __ret; \
   50658 })
   50659 #endif
   50660 
   50661 #ifdef __LITTLE_ENDIAN__
   50662 #define vld3q_dup_u16(__p0) __extension__ ({ \
   50663   uint16x8x3_t __ret; \
   50664   __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
   50665   __ret; \
   50666 })
   50667 #else
   50668 #define vld3q_dup_u16(__p0) __extension__ ({ \
   50669   uint16x8x3_t __ret; \
   50670   __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
   50671  \
   50672   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   50673   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   50674   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   50675   __ret; \
   50676 })
   50677 #endif
   50678 
   50679 #ifdef __LITTLE_ENDIAN__
   50680 #define vld3q_dup_s8(__p0) __extension__ ({ \
   50681   int8x16x3_t __ret; \
   50682   __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
   50683   __ret; \
   50684 })
   50685 #else
   50686 #define vld3q_dup_s8(__p0) __extension__ ({ \
   50687   int8x16x3_t __ret; \
   50688   __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
   50689  \
   50690   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50691   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50692   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50693   __ret; \
   50694 })
   50695 #endif
   50696 
   50697 #ifdef __LITTLE_ENDIAN__
   50698 #define vld3q_dup_f64(__p0) __extension__ ({ \
   50699   float64x2x3_t __ret; \
   50700   __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
   50701   __ret; \
   50702 })
   50703 #else
   50704 #define vld3q_dup_f64(__p0) __extension__ ({ \
   50705   float64x2x3_t __ret; \
   50706   __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
   50707  \
   50708   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50709   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50710   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50711   __ret; \
   50712 })
   50713 #endif
   50714 
   50715 #ifdef __LITTLE_ENDIAN__
   50716 #define vld3q_dup_f32(__p0) __extension__ ({ \
   50717   float32x4x3_t __ret; \
   50718   __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
   50719   __ret; \
   50720 })
   50721 #else
   50722 #define vld3q_dup_f32(__p0) __extension__ ({ \
   50723   float32x4x3_t __ret; \
   50724   __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
   50725  \
   50726   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   50727   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   50728   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   50729   __ret; \
   50730 })
   50731 #endif
   50732 
   50733 #ifdef __LITTLE_ENDIAN__
   50734 #define vld3q_dup_f16(__p0) __extension__ ({ \
   50735   float16x8x3_t __ret; \
   50736   __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
   50737   __ret; \
   50738 })
   50739 #else
   50740 #define vld3q_dup_f16(__p0) __extension__ ({ \
   50741   float16x8x3_t __ret; \
   50742   __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
   50743  \
   50744   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   50745   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   50746   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   50747   __ret; \
   50748 })
   50749 #endif
   50750 
   50751 #ifdef __LITTLE_ENDIAN__
   50752 #define vld3q_dup_s32(__p0) __extension__ ({ \
   50753   int32x4x3_t __ret; \
   50754   __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
   50755   __ret; \
   50756 })
   50757 #else
   50758 #define vld3q_dup_s32(__p0) __extension__ ({ \
   50759   int32x4x3_t __ret; \
   50760   __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
   50761  \
   50762   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   50763   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   50764   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   50765   __ret; \
   50766 })
   50767 #endif
   50768 
   50769 #ifdef __LITTLE_ENDIAN__
   50770 #define vld3q_dup_s64(__p0) __extension__ ({ \
   50771   int64x2x3_t __ret; \
   50772   __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
   50773   __ret; \
   50774 })
   50775 #else
   50776 #define vld3q_dup_s64(__p0) __extension__ ({ \
   50777   int64x2x3_t __ret; \
   50778   __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
   50779  \
   50780   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50781   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50782   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50783   __ret; \
   50784 })
   50785 #endif
   50786 
   50787 #ifdef __LITTLE_ENDIAN__
   50788 #define vld3q_dup_s16(__p0) __extension__ ({ \
   50789   int16x8x3_t __ret; \
   50790   __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
   50791   __ret; \
   50792 })
   50793 #else
   50794 #define vld3q_dup_s16(__p0) __extension__ ({ \
   50795   int16x8x3_t __ret; \
   50796   __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
   50797  \
   50798   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   50799   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   50800   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   50801   __ret; \
   50802 })
   50803 #endif
   50804 
   50805 #ifdef __LITTLE_ENDIAN__
   50806 #define vld3_dup_f64(__p0) __extension__ ({ \
   50807   float64x1x3_t __ret; \
   50808   __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
   50809   __ret; \
   50810 })
   50811 #else
   50812 #define vld3_dup_f64(__p0) __extension__ ({ \
   50813   float64x1x3_t __ret; \
   50814   __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
   50815   __ret; \
   50816 })
   50817 #endif
   50818 
   50819 #ifdef __LITTLE_ENDIAN__
   50820 #define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   50821   poly64x1x3_t __s1 = __p1; \
   50822   poly64x1x3_t __ret; \
   50823   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
   50824   __ret; \
   50825 })
   50826 #else
   50827 #define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   50828   poly64x1x3_t __s1 = __p1; \
   50829   poly64x1x3_t __ret; \
   50830   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
   50831   __ret; \
   50832 })
   50833 #endif
   50834 
   50835 #ifdef __LITTLE_ENDIAN__
   50836 #define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   50837   poly8x16x3_t __s1 = __p1; \
   50838   poly8x16x3_t __ret; \
   50839   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
   50840   __ret; \
   50841 })
   50842 #else
   50843 #define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   50844   poly8x16x3_t __s1 = __p1; \
   50845   poly8x16x3_t __rev1; \
   50846   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50847   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50848   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50849   poly8x16x3_t __ret; \
   50850   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
   50851  \
   50852   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50853   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50854   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50855   __ret; \
   50856 })
   50857 #endif
   50858 
   50859 #ifdef __LITTLE_ENDIAN__
   50860 #define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   50861   poly64x2x3_t __s1 = __p1; \
   50862   poly64x2x3_t __ret; \
   50863   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
   50864   __ret; \
   50865 })
   50866 #else
   50867 #define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   50868   poly64x2x3_t __s1 = __p1; \
   50869   poly64x2x3_t __rev1; \
   50870   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   50871   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   50872   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   50873   poly64x2x3_t __ret; \
   50874   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
   50875  \
   50876   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50877   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50878   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50879   __ret; \
   50880 })
   50881 #endif
   50882 
   50883 #ifdef __LITTLE_ENDIAN__
   50884 #define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   50885   uint8x16x3_t __s1 = __p1; \
   50886   uint8x16x3_t __ret; \
   50887   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
   50888   __ret; \
   50889 })
   50890 #else
   50891 #define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   50892   uint8x16x3_t __s1 = __p1; \
   50893   uint8x16x3_t __rev1; \
   50894   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50895   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50896   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50897   uint8x16x3_t __ret; \
   50898   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
   50899  \
   50900   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50901   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50902   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50903   __ret; \
   50904 })
   50905 #endif
   50906 
   50907 #ifdef __LITTLE_ENDIAN__
   50908 #define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   50909   uint64x2x3_t __s1 = __p1; \
   50910   uint64x2x3_t __ret; \
   50911   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
   50912   __ret; \
   50913 })
   50914 #else
   50915 #define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   50916   uint64x2x3_t __s1 = __p1; \
   50917   uint64x2x3_t __rev1; \
   50918   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   50919   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   50920   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   50921   uint64x2x3_t __ret; \
   50922   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
   50923  \
   50924   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50925   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50926   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50927   __ret; \
   50928 })
   50929 #endif
   50930 
   50931 #ifdef __LITTLE_ENDIAN__
   50932 #define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   50933   int8x16x3_t __s1 = __p1; \
   50934   int8x16x3_t __ret; \
   50935   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
   50936   __ret; \
   50937 })
   50938 #else
   50939 #define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   50940   int8x16x3_t __s1 = __p1; \
   50941   int8x16x3_t __rev1; \
   50942   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50943   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50944   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50945   int8x16x3_t __ret; \
   50946   __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
   50947  \
   50948   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50949   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50950   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   50951   __ret; \
   50952 })
   50953 #endif
   50954 
   50955 #ifdef __LITTLE_ENDIAN__
   50956 #define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   50957   float64x2x3_t __s1 = __p1; \
   50958   float64x2x3_t __ret; \
   50959   __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
   50960   __ret; \
   50961 })
   50962 #else
   50963 #define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   50964   float64x2x3_t __s1 = __p1; \
   50965   float64x2x3_t __rev1; \
   50966   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   50967   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   50968   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   50969   float64x2x3_t __ret; \
   50970   __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
   50971  \
   50972   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50973   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50974   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50975   __ret; \
   50976 })
   50977 #endif
   50978 
   50979 #ifdef __LITTLE_ENDIAN__
   50980 #define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   50981   int64x2x3_t __s1 = __p1; \
   50982   int64x2x3_t __ret; \
   50983   __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
   50984   __ret; \
   50985 })
   50986 #else
   50987 #define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   50988   int64x2x3_t __s1 = __p1; \
   50989   int64x2x3_t __rev1; \
   50990   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   50991   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   50992   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   50993   int64x2x3_t __ret; \
   50994   __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
   50995  \
   50996   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   50997   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   50998   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   50999   __ret; \
   51000 })
   51001 #endif
   51002 
   51003 #ifdef __LITTLE_ENDIAN__
   51004 #define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   51005   uint64x1x3_t __s1 = __p1; \
   51006   uint64x1x3_t __ret; \
   51007   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
   51008   __ret; \
   51009 })
   51010 #else
   51011 #define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   51012   uint64x1x3_t __s1 = __p1; \
   51013   uint64x1x3_t __ret; \
   51014   __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
   51015   __ret; \
   51016 })
   51017 #endif
   51018 
   51019 #ifdef __LITTLE_ENDIAN__
   51020 #define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   51021   float64x1x3_t __s1 = __p1; \
   51022   float64x1x3_t __ret; \
   51023   __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
   51024   __ret; \
   51025 })
   51026 #else
   51027 #define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   51028   float64x1x3_t __s1 = __p1; \
   51029   float64x1x3_t __ret; \
   51030   __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
   51031   __ret; \
   51032 })
   51033 #endif
   51034 
   51035 #ifdef __LITTLE_ENDIAN__
   51036 #define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   51037   int64x1x3_t __s1 = __p1; \
   51038   int64x1x3_t __ret; \
   51039   __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
   51040   __ret; \
   51041 })
   51042 #else
   51043 #define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   51044   int64x1x3_t __s1 = __p1; \
   51045   int64x1x3_t __ret; \
   51046   __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
   51047   __ret; \
   51048 })
   51049 #endif
   51050 
   51051 #ifdef __LITTLE_ENDIAN__
   51052 #define vld4_p64(__p0) __extension__ ({ \
   51053   poly64x1x4_t __ret; \
   51054   __builtin_neon_vld4_v(&__ret, __p0, 6); \
   51055   __ret; \
   51056 })
   51057 #else
   51058 #define vld4_p64(__p0) __extension__ ({ \
   51059   poly64x1x4_t __ret; \
   51060   __builtin_neon_vld4_v(&__ret, __p0, 6); \
   51061   __ret; \
   51062 })
   51063 #endif
   51064 
   51065 #ifdef __LITTLE_ENDIAN__
   51066 #define vld4q_p64(__p0) __extension__ ({ \
   51067   poly64x2x4_t __ret; \
   51068   __builtin_neon_vld4q_v(&__ret, __p0, 38); \
   51069   __ret; \
   51070 })
   51071 #else
   51072 #define vld4q_p64(__p0) __extension__ ({ \
   51073   poly64x2x4_t __ret; \
   51074   __builtin_neon_vld4q_v(&__ret, __p0, 38); \
   51075  \
   51076   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51077   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51078   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51079   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51080   __ret; \
   51081 })
   51082 #endif
   51083 
   51084 #ifdef __LITTLE_ENDIAN__
   51085 #define vld4q_u64(__p0) __extension__ ({ \
   51086   uint64x2x4_t __ret; \
   51087   __builtin_neon_vld4q_v(&__ret, __p0, 51); \
   51088   __ret; \
   51089 })
   51090 #else
   51091 #define vld4q_u64(__p0) __extension__ ({ \
   51092   uint64x2x4_t __ret; \
   51093   __builtin_neon_vld4q_v(&__ret, __p0, 51); \
   51094  \
   51095   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51096   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51097   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51098   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51099   __ret; \
   51100 })
   51101 #endif
   51102 
   51103 #ifdef __LITTLE_ENDIAN__
   51104 #define vld4q_f64(__p0) __extension__ ({ \
   51105   float64x2x4_t __ret; \
   51106   __builtin_neon_vld4q_v(&__ret, __p0, 42); \
   51107   __ret; \
   51108 })
   51109 #else
   51110 #define vld4q_f64(__p0) __extension__ ({ \
   51111   float64x2x4_t __ret; \
   51112   __builtin_neon_vld4q_v(&__ret, __p0, 42); \
   51113  \
   51114   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51115   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51116   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51117   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51118   __ret; \
   51119 })
   51120 #endif
   51121 
   51122 #ifdef __LITTLE_ENDIAN__
   51123 #define vld4q_s64(__p0) __extension__ ({ \
   51124   int64x2x4_t __ret; \
   51125   __builtin_neon_vld4q_v(&__ret, __p0, 35); \
   51126   __ret; \
   51127 })
   51128 #else
   51129 #define vld4q_s64(__p0) __extension__ ({ \
   51130   int64x2x4_t __ret; \
   51131   __builtin_neon_vld4q_v(&__ret, __p0, 35); \
   51132  \
   51133   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51134   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51135   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51136   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51137   __ret; \
   51138 })
   51139 #endif
   51140 
   51141 #ifdef __LITTLE_ENDIAN__
   51142 #define vld4_f64(__p0) __extension__ ({ \
   51143   float64x1x4_t __ret; \
   51144   __builtin_neon_vld4_v(&__ret, __p0, 10); \
   51145   __ret; \
   51146 })
   51147 #else
   51148 #define vld4_f64(__p0) __extension__ ({ \
   51149   float64x1x4_t __ret; \
   51150   __builtin_neon_vld4_v(&__ret, __p0, 10); \
   51151   __ret; \
   51152 })
   51153 #endif
   51154 
   51155 #ifdef __LITTLE_ENDIAN__
   51156 #define vld4_dup_p64(__p0) __extension__ ({ \
   51157   poly64x1x4_t __ret; \
   51158   __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
   51159   __ret; \
   51160 })
   51161 #else
   51162 #define vld4_dup_p64(__p0) __extension__ ({ \
   51163   poly64x1x4_t __ret; \
   51164   __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
   51165   __ret; \
   51166 })
   51167 #endif
   51168 
   51169 #ifdef __LITTLE_ENDIAN__
   51170 #define vld4q_dup_p8(__p0) __extension__ ({ \
   51171   poly8x16x4_t __ret; \
   51172   __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
   51173   __ret; \
   51174 })
   51175 #else
   51176 #define vld4q_dup_p8(__p0) __extension__ ({ \
   51177   poly8x16x4_t __ret; \
   51178   __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
   51179  \
   51180   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51181   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51182   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51183   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51184   __ret; \
   51185 })
   51186 #endif
   51187 
   51188 #ifdef __LITTLE_ENDIAN__
   51189 #define vld4q_dup_p64(__p0) __extension__ ({ \
   51190   poly64x2x4_t __ret; \
   51191   __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
   51192   __ret; \
   51193 })
   51194 #else
   51195 #define vld4q_dup_p64(__p0) __extension__ ({ \
   51196   poly64x2x4_t __ret; \
   51197   __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
   51198  \
   51199   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51200   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51201   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51202   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51203   __ret; \
   51204 })
   51205 #endif
   51206 
   51207 #ifdef __LITTLE_ENDIAN__
   51208 #define vld4q_dup_p16(__p0) __extension__ ({ \
   51209   poly16x8x4_t __ret; \
   51210   __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
   51211   __ret; \
   51212 })
   51213 #else
   51214 #define vld4q_dup_p16(__p0) __extension__ ({ \
   51215   poly16x8x4_t __ret; \
   51216   __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
   51217  \
   51218   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   51219   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   51220   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   51221   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   51222   __ret; \
   51223 })
   51224 #endif
   51225 
   51226 #ifdef __LITTLE_ENDIAN__
   51227 #define vld4q_dup_u8(__p0) __extension__ ({ \
   51228   uint8x16x4_t __ret; \
   51229   __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
   51230   __ret; \
   51231 })
   51232 #else
   51233 #define vld4q_dup_u8(__p0) __extension__ ({ \
   51234   uint8x16x4_t __ret; \
   51235   __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
   51236  \
   51237   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51238   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51239   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51240   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51241   __ret; \
   51242 })
   51243 #endif
   51244 
   51245 #ifdef __LITTLE_ENDIAN__
   51246 #define vld4q_dup_u32(__p0) __extension__ ({ \
   51247   uint32x4x4_t __ret; \
   51248   __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
   51249   __ret; \
   51250 })
   51251 #else
   51252 #define vld4q_dup_u32(__p0) __extension__ ({ \
   51253   uint32x4x4_t __ret; \
   51254   __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
   51255  \
   51256   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   51257   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   51258   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   51259   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   51260   __ret; \
   51261 })
   51262 #endif
   51263 
   51264 #ifdef __LITTLE_ENDIAN__
   51265 #define vld4q_dup_u64(__p0) __extension__ ({ \
   51266   uint64x2x4_t __ret; \
   51267   __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
   51268   __ret; \
   51269 })
   51270 #else
   51271 #define vld4q_dup_u64(__p0) __extension__ ({ \
   51272   uint64x2x4_t __ret; \
   51273   __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
   51274  \
   51275   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51276   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51277   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51278   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51279   __ret; \
   51280 })
   51281 #endif
   51282 
   51283 #ifdef __LITTLE_ENDIAN__
   51284 #define vld4q_dup_u16(__p0) __extension__ ({ \
   51285   uint16x8x4_t __ret; \
   51286   __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
   51287   __ret; \
   51288 })
   51289 #else
   51290 #define vld4q_dup_u16(__p0) __extension__ ({ \
   51291   uint16x8x4_t __ret; \
   51292   __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
   51293  \
   51294   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   51295   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   51296   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   51297   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   51298   __ret; \
   51299 })
   51300 #endif
   51301 
   51302 #ifdef __LITTLE_ENDIAN__
   51303 #define vld4q_dup_s8(__p0) __extension__ ({ \
   51304   int8x16x4_t __ret; \
   51305   __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
   51306   __ret; \
   51307 })
   51308 #else
   51309 #define vld4q_dup_s8(__p0) __extension__ ({ \
   51310   int8x16x4_t __ret; \
   51311   __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
   51312  \
   51313   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51314   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51315   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51316   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51317   __ret; \
   51318 })
   51319 #endif
   51320 
   51321 #ifdef __LITTLE_ENDIAN__
   51322 #define vld4q_dup_f64(__p0) __extension__ ({ \
   51323   float64x2x4_t __ret; \
   51324   __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
   51325   __ret; \
   51326 })
   51327 #else
   51328 #define vld4q_dup_f64(__p0) __extension__ ({ \
   51329   float64x2x4_t __ret; \
   51330   __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
   51331  \
   51332   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51333   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51334   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51335   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51336   __ret; \
   51337 })
   51338 #endif
   51339 
   51340 #ifdef __LITTLE_ENDIAN__
   51341 #define vld4q_dup_f32(__p0) __extension__ ({ \
   51342   float32x4x4_t __ret; \
   51343   __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
   51344   __ret; \
   51345 })
   51346 #else
   51347 #define vld4q_dup_f32(__p0) __extension__ ({ \
   51348   float32x4x4_t __ret; \
   51349   __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
   51350  \
   51351   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   51352   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   51353   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   51354   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   51355   __ret; \
   51356 })
   51357 #endif
   51358 
   51359 #ifdef __LITTLE_ENDIAN__
   51360 #define vld4q_dup_f16(__p0) __extension__ ({ \
   51361   float16x8x4_t __ret; \
   51362   __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
   51363   __ret; \
   51364 })
   51365 #else
   51366 #define vld4q_dup_f16(__p0) __extension__ ({ \
   51367   float16x8x4_t __ret; \
   51368   __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
   51369  \
   51370   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   51371   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   51372   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   51373   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   51374   __ret; \
   51375 })
   51376 #endif
   51377 
   51378 #ifdef __LITTLE_ENDIAN__
   51379 #define vld4q_dup_s32(__p0) __extension__ ({ \
   51380   int32x4x4_t __ret; \
   51381   __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
   51382   __ret; \
   51383 })
   51384 #else
   51385 #define vld4q_dup_s32(__p0) __extension__ ({ \
   51386   int32x4x4_t __ret; \
   51387   __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
   51388  \
   51389   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
   51390   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
   51391   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
   51392   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
   51393   __ret; \
   51394 })
   51395 #endif
   51396 
   51397 #ifdef __LITTLE_ENDIAN__
   51398 #define vld4q_dup_s64(__p0) __extension__ ({ \
   51399   int64x2x4_t __ret; \
   51400   __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
   51401   __ret; \
   51402 })
   51403 #else
   51404 #define vld4q_dup_s64(__p0) __extension__ ({ \
   51405   int64x2x4_t __ret; \
   51406   __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
   51407  \
   51408   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51409   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51410   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51411   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51412   __ret; \
   51413 })
   51414 #endif
   51415 
   51416 #ifdef __LITTLE_ENDIAN__
   51417 #define vld4q_dup_s16(__p0) __extension__ ({ \
   51418   int16x8x4_t __ret; \
   51419   __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
   51420   __ret; \
   51421 })
   51422 #else
   51423 #define vld4q_dup_s16(__p0) __extension__ ({ \
   51424   int16x8x4_t __ret; \
   51425   __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
   51426  \
   51427   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   51428   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   51429   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   51430   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   51431   __ret; \
   51432 })
   51433 #endif
   51434 
   51435 #ifdef __LITTLE_ENDIAN__
   51436 #define vld4_dup_f64(__p0) __extension__ ({ \
   51437   float64x1x4_t __ret; \
   51438   __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
   51439   __ret; \
   51440 })
   51441 #else
   51442 #define vld4_dup_f64(__p0) __extension__ ({ \
   51443   float64x1x4_t __ret; \
   51444   __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
   51445   __ret; \
   51446 })
   51447 #endif
   51448 
   51449 #ifdef __LITTLE_ENDIAN__
   51450 #define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   51451   poly64x1x4_t __s1 = __p1; \
   51452   poly64x1x4_t __ret; \
   51453   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
   51454   __ret; \
   51455 })
   51456 #else
   51457 #define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   51458   poly64x1x4_t __s1 = __p1; \
   51459   poly64x1x4_t __ret; \
   51460   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
   51461   __ret; \
   51462 })
   51463 #endif
   51464 
   51465 #ifdef __LITTLE_ENDIAN__
   51466 #define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   51467   poly8x16x4_t __s1 = __p1; \
   51468   poly8x16x4_t __ret; \
   51469   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
   51470   __ret; \
   51471 })
   51472 #else
   51473 #define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   51474   poly8x16x4_t __s1 = __p1; \
   51475   poly8x16x4_t __rev1; \
   51476   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51477   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51478   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51479   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51480   poly8x16x4_t __ret; \
   51481   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
   51482  \
   51483   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51484   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51485   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51486   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51487   __ret; \
   51488 })
   51489 #endif
   51490 
   51491 #ifdef __LITTLE_ENDIAN__
   51492 #define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   51493   poly64x2x4_t __s1 = __p1; \
   51494   poly64x2x4_t __ret; \
   51495   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
   51496   __ret; \
   51497 })
   51498 #else
   51499 #define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   51500   poly64x2x4_t __s1 = __p1; \
   51501   poly64x2x4_t __rev1; \
   51502   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   51503   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   51504   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   51505   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   51506   poly64x2x4_t __ret; \
   51507   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
   51508  \
   51509   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51510   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51511   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51512   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51513   __ret; \
   51514 })
   51515 #endif
   51516 
   51517 #ifdef __LITTLE_ENDIAN__
   51518 #define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   51519   uint8x16x4_t __s1 = __p1; \
   51520   uint8x16x4_t __ret; \
   51521   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
   51522   __ret; \
   51523 })
   51524 #else
   51525 #define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   51526   uint8x16x4_t __s1 = __p1; \
   51527   uint8x16x4_t __rev1; \
   51528   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51529   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51530   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51531   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51532   uint8x16x4_t __ret; \
   51533   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
   51534  \
   51535   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51536   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51537   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51538   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51539   __ret; \
   51540 })
   51541 #endif
   51542 
   51543 #ifdef __LITTLE_ENDIAN__
   51544 #define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   51545   uint64x2x4_t __s1 = __p1; \
   51546   uint64x2x4_t __ret; \
   51547   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
   51548   __ret; \
   51549 })
   51550 #else
   51551 #define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   51552   uint64x2x4_t __s1 = __p1; \
   51553   uint64x2x4_t __rev1; \
   51554   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   51555   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   51556   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   51557   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   51558   uint64x2x4_t __ret; \
   51559   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
   51560  \
   51561   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51562   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51563   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51564   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51565   __ret; \
   51566 })
   51567 #endif
   51568 
   51569 #ifdef __LITTLE_ENDIAN__
   51570 #define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   51571   int8x16x4_t __s1 = __p1; \
   51572   int8x16x4_t __ret; \
   51573   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
   51574   __ret; \
   51575 })
   51576 #else
   51577 #define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   51578   int8x16x4_t __s1 = __p1; \
   51579   int8x16x4_t __rev1; \
   51580   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51581   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51582   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51583   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51584   int8x16x4_t __ret; \
   51585   __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
   51586  \
   51587   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51588   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51589   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51590   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   51591   __ret; \
   51592 })
   51593 #endif
   51594 
   51595 #ifdef __LITTLE_ENDIAN__
   51596 #define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   51597   float64x2x4_t __s1 = __p1; \
   51598   float64x2x4_t __ret; \
   51599   __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
   51600   __ret; \
   51601 })
   51602 #else
   51603 #define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   51604   float64x2x4_t __s1 = __p1; \
   51605   float64x2x4_t __rev1; \
   51606   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   51607   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   51608   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   51609   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   51610   float64x2x4_t __ret; \
   51611   __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
   51612  \
   51613   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51614   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51615   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51616   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51617   __ret; \
   51618 })
   51619 #endif
   51620 
   51621 #ifdef __LITTLE_ENDIAN__
   51622 #define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   51623   int64x2x4_t __s1 = __p1; \
   51624   int64x2x4_t __ret; \
   51625   __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
   51626   __ret; \
   51627 })
   51628 #else
   51629 #define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   51630   int64x2x4_t __s1 = __p1; \
   51631   int64x2x4_t __rev1; \
   51632   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   51633   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   51634   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   51635   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   51636   int64x2x4_t __ret; \
   51637   __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
   51638  \
   51639   __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
   51640   __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
   51641   __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
   51642   __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
   51643   __ret; \
   51644 })
   51645 #endif
   51646 
   51647 #ifdef __LITTLE_ENDIAN__
   51648 #define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   51649   uint64x1x4_t __s1 = __p1; \
   51650   uint64x1x4_t __ret; \
   51651   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
   51652   __ret; \
   51653 })
   51654 #else
   51655 #define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   51656   uint64x1x4_t __s1 = __p1; \
   51657   uint64x1x4_t __ret; \
   51658   __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
   51659   __ret; \
   51660 })
   51661 #endif
   51662 
   51663 #ifdef __LITTLE_ENDIAN__
   51664 #define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   51665   float64x1x4_t __s1 = __p1; \
   51666   float64x1x4_t __ret; \
   51667   __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
   51668   __ret; \
   51669 })
   51670 #else
   51671 #define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   51672   float64x1x4_t __s1 = __p1; \
   51673   float64x1x4_t __ret; \
   51674   __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
   51675   __ret; \
   51676 })
   51677 #endif
   51678 
   51679 #ifdef __LITTLE_ENDIAN__
   51680 #define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   51681   int64x1x4_t __s1 = __p1; \
   51682   int64x1x4_t __ret; \
   51683   __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
   51684   __ret; \
   51685 })
   51686 #else
   51687 #define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   51688   int64x1x4_t __s1 = __p1; \
   51689   int64x1x4_t __ret; \
   51690   __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
   51691   __ret; \
   51692 })
   51693 #endif
   51694 
   51695 #ifdef __LITTLE_ENDIAN__
   51696 #define vldrq_p128(__p0) __extension__ ({ \
   51697   poly128_t __ret; \
   51698   __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
   51699   __ret; \
   51700 })
   51701 #else
   51702 #define vldrq_p128(__p0) __extension__ ({ \
   51703   poly128_t __ret; \
   51704   __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
   51705   __ret; \
   51706 })
   51707 #endif
   51708 
   51709 #ifdef __LITTLE_ENDIAN__
   51710 __ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
   51711   float64x2_t __ret;
   51712   __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   51713   return __ret;
   51714 }
   51715 #else
   51716 __ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
   51717   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   51718   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   51719   float64x2_t __ret;
   51720   __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   51721   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   51722   return __ret;
   51723 }
   51724 #endif
   51725 
   51726 #ifdef __LITTLE_ENDIAN__
   51727 __ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
   51728   float64x1_t __ret;
   51729   __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   51730   return __ret;
   51731 }
   51732 #else
   51733 __ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
   51734   float64x1_t __ret;
   51735   __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   51736   return __ret;
   51737 }
   51738 #endif
   51739 
   51740 #ifdef __LITTLE_ENDIAN__
   51741 __ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
   51742   float64_t __ret;
   51743   __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__p0);
   51744   return __ret;
   51745 }
   51746 #else
   51747 __ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
   51748   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   51749   float64_t __ret;
   51750   __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__rev0);
   51751   return __ret;
   51752 }
   51753 #endif
   51754 
   51755 #ifdef __LITTLE_ENDIAN__
   51756 __ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
   51757   float32_t __ret;
   51758   __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__p0);
   51759   return __ret;
   51760 }
   51761 #else
   51762 __ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
   51763   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   51764   float32_t __ret;
   51765   __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__rev0);
   51766   return __ret;
   51767 }
   51768 #endif
   51769 
   51770 #ifdef __LITTLE_ENDIAN__
   51771 __ai float32_t vmaxnmv_f32(float32x2_t __p0) {
   51772   float32_t __ret;
   51773   __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__p0);
   51774   return __ret;
   51775 }
   51776 #else
   51777 __ai float32_t vmaxnmv_f32(float32x2_t __p0) {
   51778   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   51779   float32_t __ret;
   51780   __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__rev0);
   51781   return __ret;
   51782 }
   51783 #endif
   51784 
   51785 #ifdef __LITTLE_ENDIAN__
   51786 __ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
   51787   uint8_t __ret;
   51788   __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__p0);
   51789   return __ret;
   51790 }
   51791 #else
   51792 __ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
   51793   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   51794   uint8_t __ret;
   51795   __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__rev0);
   51796   return __ret;
   51797 }
   51798 #endif
   51799 
   51800 #ifdef __LITTLE_ENDIAN__
   51801 __ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
   51802   uint32_t __ret;
   51803   __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__p0);
   51804   return __ret;
   51805 }
   51806 #else
   51807 __ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
   51808   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   51809   uint32_t __ret;
   51810   __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__rev0);
   51811   return __ret;
   51812 }
   51813 #endif
   51814 
   51815 #ifdef __LITTLE_ENDIAN__
   51816 __ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
   51817   uint16_t __ret;
   51818   __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__p0);
   51819   return __ret;
   51820 }
   51821 #else
   51822 __ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
   51823   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   51824   uint16_t __ret;
   51825   __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__rev0);
   51826   return __ret;
   51827 }
   51828 #endif
   51829 
   51830 #ifdef __LITTLE_ENDIAN__
   51831 __ai int8_t vmaxvq_s8(int8x16_t __p0) {
   51832   int8_t __ret;
   51833   __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__p0);
   51834   return __ret;
   51835 }
   51836 #else
   51837 __ai int8_t vmaxvq_s8(int8x16_t __p0) {
   51838   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   51839   int8_t __ret;
   51840   __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__rev0);
   51841   return __ret;
   51842 }
   51843 #endif
   51844 
   51845 #ifdef __LITTLE_ENDIAN__
   51846 __ai float64_t vmaxvq_f64(float64x2_t __p0) {
   51847   float64_t __ret;
   51848   __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__p0);
   51849   return __ret;
   51850 }
   51851 #else
   51852 __ai float64_t vmaxvq_f64(float64x2_t __p0) {
   51853   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   51854   float64_t __ret;
   51855   __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__rev0);
   51856   return __ret;
   51857 }
   51858 #endif
   51859 
   51860 #ifdef __LITTLE_ENDIAN__
   51861 __ai float32_t vmaxvq_f32(float32x4_t __p0) {
   51862   float32_t __ret;
   51863   __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__p0);
   51864   return __ret;
   51865 }
   51866 #else
   51867 __ai float32_t vmaxvq_f32(float32x4_t __p0) {
   51868   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   51869   float32_t __ret;
   51870   __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__rev0);
   51871   return __ret;
   51872 }
   51873 #endif
   51874 
   51875 #ifdef __LITTLE_ENDIAN__
   51876 __ai int32_t vmaxvq_s32(int32x4_t __p0) {
   51877   int32_t __ret;
   51878   __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__p0);
   51879   return __ret;
   51880 }
   51881 #else
   51882 __ai int32_t vmaxvq_s32(int32x4_t __p0) {
   51883   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   51884   int32_t __ret;
   51885   __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__rev0);
   51886   return __ret;
   51887 }
   51888 #endif
   51889 
   51890 #ifdef __LITTLE_ENDIAN__
   51891 __ai int16_t vmaxvq_s16(int16x8_t __p0) {
   51892   int16_t __ret;
   51893   __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__p0);
   51894   return __ret;
   51895 }
   51896 #else
   51897 __ai int16_t vmaxvq_s16(int16x8_t __p0) {
   51898   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   51899   int16_t __ret;
   51900   __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__rev0);
   51901   return __ret;
   51902 }
   51903 #endif
   51904 
   51905 #ifdef __LITTLE_ENDIAN__
   51906 __ai uint8_t vmaxv_u8(uint8x8_t __p0) {
   51907   uint8_t __ret;
   51908   __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__p0);
   51909   return __ret;
   51910 }
   51911 #else
   51912 __ai uint8_t vmaxv_u8(uint8x8_t __p0) {
   51913   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   51914   uint8_t __ret;
   51915   __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__rev0);
   51916   return __ret;
   51917 }
   51918 #endif
   51919 
   51920 #ifdef __LITTLE_ENDIAN__
   51921 __ai uint32_t vmaxv_u32(uint32x2_t __p0) {
   51922   uint32_t __ret;
   51923   __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__p0);
   51924   return __ret;
   51925 }
   51926 #else
   51927 __ai uint32_t vmaxv_u32(uint32x2_t __p0) {
   51928   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   51929   uint32_t __ret;
   51930   __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__rev0);
   51931   return __ret;
   51932 }
   51933 #endif
   51934 
   51935 #ifdef __LITTLE_ENDIAN__
   51936 __ai uint16_t vmaxv_u16(uint16x4_t __p0) {
   51937   uint16_t __ret;
   51938   __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__p0);
   51939   return __ret;
   51940 }
   51941 #else
   51942 __ai uint16_t vmaxv_u16(uint16x4_t __p0) {
   51943   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   51944   uint16_t __ret;
   51945   __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__rev0);
   51946   return __ret;
   51947 }
   51948 #endif
   51949 
   51950 #ifdef __LITTLE_ENDIAN__
   51951 __ai int8_t vmaxv_s8(int8x8_t __p0) {
   51952   int8_t __ret;
   51953   __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__p0);
   51954   return __ret;
   51955 }
   51956 #else
   51957 __ai int8_t vmaxv_s8(int8x8_t __p0) {
   51958   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   51959   int8_t __ret;
   51960   __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__rev0);
   51961   return __ret;
   51962 }
   51963 #endif
   51964 
   51965 #ifdef __LITTLE_ENDIAN__
   51966 __ai float32_t vmaxv_f32(float32x2_t __p0) {
   51967   float32_t __ret;
   51968   __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__p0);
   51969   return __ret;
   51970 }
   51971 #else
   51972 __ai float32_t vmaxv_f32(float32x2_t __p0) {
   51973   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   51974   float32_t __ret;
   51975   __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__rev0);
   51976   return __ret;
   51977 }
   51978 #endif
   51979 
   51980 #ifdef __LITTLE_ENDIAN__
   51981 __ai int32_t vmaxv_s32(int32x2_t __p0) {
   51982   int32_t __ret;
   51983   __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__p0);
   51984   return __ret;
   51985 }
   51986 #else
   51987 __ai int32_t vmaxv_s32(int32x2_t __p0) {
   51988   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   51989   int32_t __ret;
   51990   __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__rev0);
   51991   return __ret;
   51992 }
   51993 #endif
   51994 
   51995 #ifdef __LITTLE_ENDIAN__
   51996 __ai int16_t vmaxv_s16(int16x4_t __p0) {
   51997   int16_t __ret;
   51998   __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__p0);
   51999   return __ret;
   52000 }
   52001 #else
   52002 __ai int16_t vmaxv_s16(int16x4_t __p0) {
   52003   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   52004   int16_t __ret;
   52005   __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__rev0);
   52006   return __ret;
   52007 }
   52008 #endif
   52009 
   52010 #ifdef __LITTLE_ENDIAN__
   52011 __ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
   52012   float64x2_t __ret;
   52013   __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   52014   return __ret;
   52015 }
   52016 #else
   52017 __ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
   52018   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52019   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   52020   float64x2_t __ret;
   52021   __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   52022   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   52023   return __ret;
   52024 }
   52025 #endif
   52026 
   52027 #ifdef __LITTLE_ENDIAN__
   52028 __ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
   52029   float64x1_t __ret;
   52030   __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   52031   return __ret;
   52032 }
   52033 #else
   52034 __ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
   52035   float64x1_t __ret;
   52036   __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   52037   return __ret;
   52038 }
   52039 #endif
   52040 
   52041 #ifdef __LITTLE_ENDIAN__
   52042 __ai float64_t vminnmvq_f64(float64x2_t __p0) {
   52043   float64_t __ret;
   52044   __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__p0);
   52045   return __ret;
   52046 }
   52047 #else
   52048 __ai float64_t vminnmvq_f64(float64x2_t __p0) {
   52049   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52050   float64_t __ret;
   52051   __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__rev0);
   52052   return __ret;
   52053 }
   52054 #endif
   52055 
   52056 #ifdef __LITTLE_ENDIAN__
   52057 __ai float32_t vminnmvq_f32(float32x4_t __p0) {
   52058   float32_t __ret;
   52059   __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__p0);
   52060   return __ret;
   52061 }
   52062 #else
   52063 __ai float32_t vminnmvq_f32(float32x4_t __p0) {
   52064   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   52065   float32_t __ret;
   52066   __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__rev0);
   52067   return __ret;
   52068 }
   52069 #endif
   52070 
   52071 #ifdef __LITTLE_ENDIAN__
   52072 __ai float32_t vminnmv_f32(float32x2_t __p0) {
   52073   float32_t __ret;
   52074   __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__p0);
   52075   return __ret;
   52076 }
   52077 #else
   52078 __ai float32_t vminnmv_f32(float32x2_t __p0) {
   52079   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52080   float32_t __ret;
   52081   __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__rev0);
   52082   return __ret;
   52083 }
   52084 #endif
   52085 
   52086 #ifdef __LITTLE_ENDIAN__
   52087 __ai uint8_t vminvq_u8(uint8x16_t __p0) {
   52088   uint8_t __ret;
   52089   __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__p0);
   52090   return __ret;
   52091 }
   52092 #else
   52093 __ai uint8_t vminvq_u8(uint8x16_t __p0) {
   52094   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   52095   uint8_t __ret;
   52096   __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__rev0);
   52097   return __ret;
   52098 }
   52099 #endif
   52100 
   52101 #ifdef __LITTLE_ENDIAN__
   52102 __ai uint32_t vminvq_u32(uint32x4_t __p0) {
   52103   uint32_t __ret;
   52104   __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__p0);
   52105   return __ret;
   52106 }
   52107 #else
   52108 __ai uint32_t vminvq_u32(uint32x4_t __p0) {
   52109   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   52110   uint32_t __ret;
   52111   __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__rev0);
   52112   return __ret;
   52113 }
   52114 #endif
   52115 
   52116 #ifdef __LITTLE_ENDIAN__
   52117 __ai uint16_t vminvq_u16(uint16x8_t __p0) {
   52118   uint16_t __ret;
   52119   __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__p0);
   52120   return __ret;
   52121 }
   52122 #else
   52123 __ai uint16_t vminvq_u16(uint16x8_t __p0) {
   52124   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   52125   uint16_t __ret;
   52126   __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__rev0);
   52127   return __ret;
   52128 }
   52129 #endif
   52130 
   52131 #ifdef __LITTLE_ENDIAN__
   52132 __ai int8_t vminvq_s8(int8x16_t __p0) {
   52133   int8_t __ret;
   52134   __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__p0);
   52135   return __ret;
   52136 }
   52137 #else
   52138 __ai int8_t vminvq_s8(int8x16_t __p0) {
   52139   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   52140   int8_t __ret;
   52141   __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__rev0);
   52142   return __ret;
   52143 }
   52144 #endif
   52145 
   52146 #ifdef __LITTLE_ENDIAN__
   52147 __ai float64_t vminvq_f64(float64x2_t __p0) {
   52148   float64_t __ret;
   52149   __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__p0);
   52150   return __ret;
   52151 }
   52152 #else
   52153 __ai float64_t vminvq_f64(float64x2_t __p0) {
   52154   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52155   float64_t __ret;
   52156   __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__rev0);
   52157   return __ret;
   52158 }
   52159 #endif
   52160 
   52161 #ifdef __LITTLE_ENDIAN__
   52162 __ai float32_t vminvq_f32(float32x4_t __p0) {
   52163   float32_t __ret;
   52164   __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__p0);
   52165   return __ret;
   52166 }
   52167 #else
   52168 __ai float32_t vminvq_f32(float32x4_t __p0) {
   52169   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   52170   float32_t __ret;
   52171   __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__rev0);
   52172   return __ret;
   52173 }
   52174 #endif
   52175 
   52176 #ifdef __LITTLE_ENDIAN__
   52177 __ai int32_t vminvq_s32(int32x4_t __p0) {
   52178   int32_t __ret;
   52179   __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__p0);
   52180   return __ret;
   52181 }
   52182 #else
   52183 __ai int32_t vminvq_s32(int32x4_t __p0) {
   52184   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   52185   int32_t __ret;
   52186   __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__rev0);
   52187   return __ret;
   52188 }
   52189 #endif
   52190 
   52191 #ifdef __LITTLE_ENDIAN__
   52192 __ai int16_t vminvq_s16(int16x8_t __p0) {
   52193   int16_t __ret;
   52194   __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__p0);
   52195   return __ret;
   52196 }
   52197 #else
   52198 __ai int16_t vminvq_s16(int16x8_t __p0) {
   52199   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   52200   int16_t __ret;
   52201   __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__rev0);
   52202   return __ret;
   52203 }
   52204 #endif
   52205 
   52206 #ifdef __LITTLE_ENDIAN__
   52207 __ai uint8_t vminv_u8(uint8x8_t __p0) {
   52208   uint8_t __ret;
   52209   __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__p0);
   52210   return __ret;
   52211 }
   52212 #else
   52213 __ai uint8_t vminv_u8(uint8x8_t __p0) {
   52214   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   52215   uint8_t __ret;
   52216   __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__rev0);
   52217   return __ret;
   52218 }
   52219 #endif
   52220 
   52221 #ifdef __LITTLE_ENDIAN__
   52222 __ai uint32_t vminv_u32(uint32x2_t __p0) {
   52223   uint32_t __ret;
   52224   __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__p0);
   52225   return __ret;
   52226 }
   52227 #else
   52228 __ai uint32_t vminv_u32(uint32x2_t __p0) {
   52229   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52230   uint32_t __ret;
   52231   __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__rev0);
   52232   return __ret;
   52233 }
   52234 #endif
   52235 
   52236 #ifdef __LITTLE_ENDIAN__
   52237 __ai uint16_t vminv_u16(uint16x4_t __p0) {
   52238   uint16_t __ret;
   52239   __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__p0);
   52240   return __ret;
   52241 }
   52242 #else
   52243 __ai uint16_t vminv_u16(uint16x4_t __p0) {
   52244   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   52245   uint16_t __ret;
   52246   __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__rev0);
   52247   return __ret;
   52248 }
   52249 #endif
   52250 
   52251 #ifdef __LITTLE_ENDIAN__
   52252 __ai int8_t vminv_s8(int8x8_t __p0) {
   52253   int8_t __ret;
   52254   __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__p0);
   52255   return __ret;
   52256 }
   52257 #else
   52258 __ai int8_t vminv_s8(int8x8_t __p0) {
   52259   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   52260   int8_t __ret;
   52261   __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__rev0);
   52262   return __ret;
   52263 }
   52264 #endif
   52265 
   52266 #ifdef __LITTLE_ENDIAN__
   52267 __ai float32_t vminv_f32(float32x2_t __p0) {
   52268   float32_t __ret;
   52269   __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__p0);
   52270   return __ret;
   52271 }
   52272 #else
   52273 __ai float32_t vminv_f32(float32x2_t __p0) {
   52274   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52275   float32_t __ret;
   52276   __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__rev0);
   52277   return __ret;
   52278 }
   52279 #endif
   52280 
   52281 #ifdef __LITTLE_ENDIAN__
   52282 __ai int32_t vminv_s32(int32x2_t __p0) {
   52283   int32_t __ret;
   52284   __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__p0);
   52285   return __ret;
   52286 }
   52287 #else
   52288 __ai int32_t vminv_s32(int32x2_t __p0) {
   52289   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52290   int32_t __ret;
   52291   __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__rev0);
   52292   return __ret;
   52293 }
   52294 #endif
   52295 
   52296 #ifdef __LITTLE_ENDIAN__
   52297 __ai int16_t vminv_s16(int16x4_t __p0) {
   52298   int16_t __ret;
   52299   __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__p0);
   52300   return __ret;
   52301 }
   52302 #else
   52303 __ai int16_t vminv_s16(int16x4_t __p0) {
   52304   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   52305   int16_t __ret;
   52306   __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__rev0);
   52307   return __ret;
   52308 }
   52309 #endif
   52310 
   52311 #ifdef __LITTLE_ENDIAN__
   52312 __ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   52313   float64x2_t __ret;
   52314   __ret = __p0 + __p1 * __p2;
   52315   return __ret;
   52316 }
   52317 #else
   52318 __ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   52319   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52320   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   52321   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   52322   float64x2_t __ret;
   52323   __ret = __rev0 + __rev1 * __rev2;
   52324   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   52325   return __ret;
   52326 }
   52327 #endif
   52328 
   52329 #ifdef __LITTLE_ENDIAN__
   52330 __ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   52331   float64x1_t __ret;
   52332   __ret = __p0 + __p1 * __p2;
   52333   return __ret;
   52334 }
   52335 #else
   52336 __ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   52337   float64x1_t __ret;
   52338   __ret = __p0 + __p1 * __p2;
   52339   return __ret;
   52340 }
   52341 #endif
   52342 
   52343 #ifdef __LITTLE_ENDIAN__
   52344 #define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52345   uint32x4_t __s0 = __p0; \
   52346   uint32x4_t __s1 = __p1; \
   52347   uint32x4_t __s2 = __p2; \
   52348   uint32x4_t __ret; \
   52349   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   52350   __ret; \
   52351 })
   52352 #else
   52353 #define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52354   uint32x4_t __s0 = __p0; \
   52355   uint32x4_t __s1 = __p1; \
   52356   uint32x4_t __s2 = __p2; \
   52357   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52358   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52359   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52360   uint32x4_t __ret; \
   52361   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   52362   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52363   __ret; \
   52364 })
   52365 #endif
   52366 
   52367 #ifdef __LITTLE_ENDIAN__
   52368 #define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52369   uint16x8_t __s0 = __p0; \
   52370   uint16x8_t __s1 = __p1; \
   52371   uint16x8_t __s2 = __p2; \
   52372   uint16x8_t __ret; \
   52373   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   52374   __ret; \
   52375 })
   52376 #else
   52377 #define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52378   uint16x8_t __s0 = __p0; \
   52379   uint16x8_t __s1 = __p1; \
   52380   uint16x8_t __s2 = __p2; \
   52381   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   52382   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   52383   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   52384   uint16x8_t __ret; \
   52385   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   52386   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   52387   __ret; \
   52388 })
   52389 #endif
   52390 
   52391 #ifdef __LITTLE_ENDIAN__
   52392 #define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52393   float32x4_t __s0 = __p0; \
   52394   float32x4_t __s1 = __p1; \
   52395   float32x4_t __s2 = __p2; \
   52396   float32x4_t __ret; \
   52397   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   52398   __ret; \
   52399 })
   52400 #else
   52401 #define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52402   float32x4_t __s0 = __p0; \
   52403   float32x4_t __s1 = __p1; \
   52404   float32x4_t __s2 = __p2; \
   52405   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52406   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52407   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52408   float32x4_t __ret; \
   52409   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   52410   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52411   __ret; \
   52412 })
   52413 #endif
   52414 
   52415 #ifdef __LITTLE_ENDIAN__
   52416 #define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52417   int32x4_t __s0 = __p0; \
   52418   int32x4_t __s1 = __p1; \
   52419   int32x4_t __s2 = __p2; \
   52420   int32x4_t __ret; \
   52421   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   52422   __ret; \
   52423 })
   52424 #else
   52425 #define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52426   int32x4_t __s0 = __p0; \
   52427   int32x4_t __s1 = __p1; \
   52428   int32x4_t __s2 = __p2; \
   52429   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52430   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52431   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52432   int32x4_t __ret; \
   52433   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   52434   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52435   __ret; \
   52436 })
   52437 #endif
   52438 
   52439 #ifdef __LITTLE_ENDIAN__
   52440 #define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52441   int16x8_t __s0 = __p0; \
   52442   int16x8_t __s1 = __p1; \
   52443   int16x8_t __s2 = __p2; \
   52444   int16x8_t __ret; \
   52445   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   52446   __ret; \
   52447 })
   52448 #else
   52449 #define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52450   int16x8_t __s0 = __p0; \
   52451   int16x8_t __s1 = __p1; \
   52452   int16x8_t __s2 = __p2; \
   52453   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   52454   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   52455   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   52456   int16x8_t __ret; \
   52457   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   52458   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   52459   __ret; \
   52460 })
   52461 #endif
   52462 
   52463 #ifdef __LITTLE_ENDIAN__
   52464 #define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52465   uint32x2_t __s0 = __p0; \
   52466   uint32x2_t __s1 = __p1; \
   52467   uint32x4_t __s2 = __p2; \
   52468   uint32x2_t __ret; \
   52469   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   52470   __ret; \
   52471 })
   52472 #else
   52473 #define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52474   uint32x2_t __s0 = __p0; \
   52475   uint32x2_t __s1 = __p1; \
   52476   uint32x4_t __s2 = __p2; \
   52477   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   52478   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   52479   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52480   uint32x2_t __ret; \
   52481   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   52482   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   52483   __ret; \
   52484 })
   52485 #endif
   52486 
   52487 #ifdef __LITTLE_ENDIAN__
   52488 #define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52489   uint16x4_t __s0 = __p0; \
   52490   uint16x4_t __s1 = __p1; \
   52491   uint16x8_t __s2 = __p2; \
   52492   uint16x4_t __ret; \
   52493   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   52494   __ret; \
   52495 })
   52496 #else
   52497 #define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52498   uint16x4_t __s0 = __p0; \
   52499   uint16x4_t __s1 = __p1; \
   52500   uint16x8_t __s2 = __p2; \
   52501   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52502   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52503   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   52504   uint16x4_t __ret; \
   52505   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   52506   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52507   __ret; \
   52508 })
   52509 #endif
   52510 
   52511 #ifdef __LITTLE_ENDIAN__
   52512 #define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52513   float32x2_t __s0 = __p0; \
   52514   float32x2_t __s1 = __p1; \
   52515   float32x4_t __s2 = __p2; \
   52516   float32x2_t __ret; \
   52517   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   52518   __ret; \
   52519 })
   52520 #else
   52521 #define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52522   float32x2_t __s0 = __p0; \
   52523   float32x2_t __s1 = __p1; \
   52524   float32x4_t __s2 = __p2; \
   52525   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   52526   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   52527   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52528   float32x2_t __ret; \
   52529   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   52530   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   52531   __ret; \
   52532 })
   52533 #endif
   52534 
   52535 #ifdef __LITTLE_ENDIAN__
   52536 #define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52537   int32x2_t __s0 = __p0; \
   52538   int32x2_t __s1 = __p1; \
   52539   int32x4_t __s2 = __p2; \
   52540   int32x2_t __ret; \
   52541   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   52542   __ret; \
   52543 })
   52544 #else
   52545 #define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52546   int32x2_t __s0 = __p0; \
   52547   int32x2_t __s1 = __p1; \
   52548   int32x4_t __s2 = __p2; \
   52549   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   52550   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   52551   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52552   int32x2_t __ret; \
   52553   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   52554   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   52555   __ret; \
   52556 })
   52557 #endif
   52558 
   52559 #ifdef __LITTLE_ENDIAN__
   52560 #define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52561   int16x4_t __s0 = __p0; \
   52562   int16x4_t __s1 = __p1; \
   52563   int16x8_t __s2 = __p2; \
   52564   int16x4_t __ret; \
   52565   __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   52566   __ret; \
   52567 })
   52568 #else
   52569 #define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52570   int16x4_t __s0 = __p0; \
   52571   int16x4_t __s1 = __p1; \
   52572   int16x8_t __s2 = __p2; \
   52573   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52574   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52575   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   52576   int16x4_t __ret; \
   52577   __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   52578   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52579   __ret; \
   52580 })
   52581 #endif
   52582 
   52583 #ifdef __LITTLE_ENDIAN__
   52584 __ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
   52585   float64x2_t __ret;
   52586   __ret = __p0 + __p1 * (float64x2_t) {__p2, __p2};
   52587   return __ret;
   52588 }
   52589 #else
   52590 __ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
   52591   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52592   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   52593   float64x2_t __ret;
   52594   __ret = __rev0 + __rev1 * (float64x2_t) {__p2, __p2};
   52595   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   52596   return __ret;
   52597 }
   52598 #endif
   52599 
   52600 #ifdef __LITTLE_ENDIAN__
   52601 #define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52602   uint64x2_t __s0 = __p0; \
   52603   uint32x4_t __s1 = __p1; \
   52604   uint32x2_t __s2 = __p2; \
   52605   uint64x2_t __ret; \
   52606   __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   52607   __ret; \
   52608 })
   52609 #else
   52610 #define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52611   uint64x2_t __s0 = __p0; \
   52612   uint32x4_t __s1 = __p1; \
   52613   uint32x2_t __s2 = __p2; \
   52614   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   52615   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52616   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   52617   uint64x2_t __ret; \
   52618   __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   52619   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   52620   __ret; \
   52621 })
   52622 #endif
   52623 
   52624 #ifdef __LITTLE_ENDIAN__
   52625 #define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52626   uint32x4_t __s0 = __p0; \
   52627   uint16x8_t __s1 = __p1; \
   52628   uint16x4_t __s2 = __p2; \
   52629   uint32x4_t __ret; \
   52630   __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   52631   __ret; \
   52632 })
   52633 #else
   52634 #define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52635   uint32x4_t __s0 = __p0; \
   52636   uint16x8_t __s1 = __p1; \
   52637   uint16x4_t __s2 = __p2; \
   52638   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52639   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   52640   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52641   uint32x4_t __ret; \
   52642   __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   52643   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52644   __ret; \
   52645 })
   52646 #endif
   52647 
   52648 #ifdef __LITTLE_ENDIAN__
   52649 #define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52650   int64x2_t __s0 = __p0; \
   52651   int32x4_t __s1 = __p1; \
   52652   int32x2_t __s2 = __p2; \
   52653   int64x2_t __ret; \
   52654   __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   52655   __ret; \
   52656 })
   52657 #else
   52658 #define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52659   int64x2_t __s0 = __p0; \
   52660   int32x4_t __s1 = __p1; \
   52661   int32x2_t __s2 = __p2; \
   52662   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   52663   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52664   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   52665   int64x2_t __ret; \
   52666   __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   52667   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   52668   __ret; \
   52669 })
   52670 #endif
   52671 
   52672 #ifdef __LITTLE_ENDIAN__
   52673 #define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52674   int32x4_t __s0 = __p0; \
   52675   int16x8_t __s1 = __p1; \
   52676   int16x4_t __s2 = __p2; \
   52677   int32x4_t __ret; \
   52678   __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   52679   __ret; \
   52680 })
   52681 #else
   52682 #define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52683   int32x4_t __s0 = __p0; \
   52684   int16x8_t __s1 = __p1; \
   52685   int16x4_t __s2 = __p2; \
   52686   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52687   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   52688   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52689   int32x4_t __ret; \
   52690   __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   52691   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52692   __ret; \
   52693 })
   52694 #endif
   52695 
   52696 #ifdef __LITTLE_ENDIAN__
   52697 #define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52698   uint64x2_t __s0 = __p0; \
   52699   uint32x4_t __s1 = __p1; \
   52700   uint32x4_t __s2 = __p2; \
   52701   uint64x2_t __ret; \
   52702   __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   52703   __ret; \
   52704 })
   52705 #else
   52706 #define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52707   uint64x2_t __s0 = __p0; \
   52708   uint32x4_t __s1 = __p1; \
   52709   uint32x4_t __s2 = __p2; \
   52710   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   52711   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52712   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52713   uint64x2_t __ret; \
   52714   __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   52715   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   52716   __ret; \
   52717 })
   52718 #endif
   52719 
   52720 #ifdef __LITTLE_ENDIAN__
   52721 #define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52722   uint32x4_t __s0 = __p0; \
   52723   uint16x8_t __s1 = __p1; \
   52724   uint16x8_t __s2 = __p2; \
   52725   uint32x4_t __ret; \
   52726   __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   52727   __ret; \
   52728 })
   52729 #else
   52730 #define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52731   uint32x4_t __s0 = __p0; \
   52732   uint16x8_t __s1 = __p1; \
   52733   uint16x8_t __s2 = __p2; \
   52734   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52735   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   52736   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   52737   uint32x4_t __ret; \
   52738   __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   52739   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52740   __ret; \
   52741 })
   52742 #endif
   52743 
   52744 #ifdef __LITTLE_ENDIAN__
   52745 #define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52746   int64x2_t __s0 = __p0; \
   52747   int32x4_t __s1 = __p1; \
   52748   int32x4_t __s2 = __p2; \
   52749   int64x2_t __ret; \
   52750   __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   52751   __ret; \
   52752 })
   52753 #else
   52754 #define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52755   int64x2_t __s0 = __p0; \
   52756   int32x4_t __s1 = __p1; \
   52757   int32x4_t __s2 = __p2; \
   52758   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   52759   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52760   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52761   int64x2_t __ret; \
   52762   __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   52763   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   52764   __ret; \
   52765 })
   52766 #endif
   52767 
   52768 #ifdef __LITTLE_ENDIAN__
   52769 #define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52770   int32x4_t __s0 = __p0; \
   52771   int16x8_t __s1 = __p1; \
   52772   int16x8_t __s2 = __p2; \
   52773   int32x4_t __ret; \
   52774   __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   52775   __ret; \
   52776 })
   52777 #else
   52778 #define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52779   int32x4_t __s0 = __p0; \
   52780   int16x8_t __s1 = __p1; \
   52781   int16x8_t __s2 = __p2; \
   52782   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52783   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   52784   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   52785   int32x4_t __ret; \
   52786   __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   52787   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52788   __ret; \
   52789 })
   52790 #endif
   52791 
   52792 #ifdef __LITTLE_ENDIAN__
   52793 #define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52794   uint64x2_t __s0 = __p0; \
   52795   uint32x2_t __s1 = __p1; \
   52796   uint32x4_t __s2 = __p2; \
   52797   uint64x2_t __ret; \
   52798   __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   52799   __ret; \
   52800 })
   52801 #else
   52802 #define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52803   uint64x2_t __s0 = __p0; \
   52804   uint32x2_t __s1 = __p1; \
   52805   uint32x4_t __s2 = __p2; \
   52806   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   52807   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   52808   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52809   uint64x2_t __ret; \
   52810   __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   52811   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   52812   __ret; \
   52813 })
   52814 #endif
   52815 
   52816 #ifdef __LITTLE_ENDIAN__
   52817 #define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52818   uint32x4_t __s0 = __p0; \
   52819   uint16x4_t __s1 = __p1; \
   52820   uint16x8_t __s2 = __p2; \
   52821   uint32x4_t __ret; \
   52822   __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   52823   __ret; \
   52824 })
   52825 #else
   52826 #define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52827   uint32x4_t __s0 = __p0; \
   52828   uint16x4_t __s1 = __p1; \
   52829   uint16x8_t __s2 = __p2; \
   52830   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52831   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52832   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   52833   uint32x4_t __ret; \
   52834   __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   52835   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52836   __ret; \
   52837 })
   52838 #endif
   52839 
   52840 #ifdef __LITTLE_ENDIAN__
   52841 #define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52842   int64x2_t __s0 = __p0; \
   52843   int32x2_t __s1 = __p1; \
   52844   int32x4_t __s2 = __p2; \
   52845   int64x2_t __ret; \
   52846   __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   52847   __ret; \
   52848 })
   52849 #else
   52850 #define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52851   int64x2_t __s0 = __p0; \
   52852   int32x2_t __s1 = __p1; \
   52853   int32x4_t __s2 = __p2; \
   52854   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   52855   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   52856   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52857   int64x2_t __ret; \
   52858   __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   52859   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   52860   __ret; \
   52861 })
   52862 #endif
   52863 
   52864 #ifdef __LITTLE_ENDIAN__
   52865 #define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52866   int32x4_t __s0 = __p0; \
   52867   int16x4_t __s1 = __p1; \
   52868   int16x8_t __s2 = __p2; \
   52869   int32x4_t __ret; \
   52870   __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   52871   __ret; \
   52872 })
   52873 #else
   52874 #define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52875   int32x4_t __s0 = __p0; \
   52876   int16x4_t __s1 = __p1; \
   52877   int16x8_t __s2 = __p2; \
   52878   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52879   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52880   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   52881   int32x4_t __ret; \
   52882   __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   52883   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52884   __ret; \
   52885 })
   52886 #endif
   52887 
   52888 #ifdef __LITTLE_ENDIAN__
   52889 __ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   52890   float64x2_t __ret;
   52891   __ret = __p0 - __p1 * __p2;
   52892   return __ret;
   52893 }
   52894 #else
   52895 __ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
   52896   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   52897   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   52898   float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   52899   float64x2_t __ret;
   52900   __ret = __rev0 - __rev1 * __rev2;
   52901   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   52902   return __ret;
   52903 }
   52904 #endif
   52905 
   52906 #ifdef __LITTLE_ENDIAN__
   52907 __ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   52908   float64x1_t __ret;
   52909   __ret = __p0 - __p1 * __p2;
   52910   return __ret;
   52911 }
   52912 #else
   52913 __ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
   52914   float64x1_t __ret;
   52915   __ret = __p0 - __p1 * __p2;
   52916   return __ret;
   52917 }
   52918 #endif
   52919 
   52920 #ifdef __LITTLE_ENDIAN__
   52921 #define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52922   uint32x4_t __s0 = __p0; \
   52923   uint32x4_t __s1 = __p1; \
   52924   uint32x4_t __s2 = __p2; \
   52925   uint32x4_t __ret; \
   52926   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   52927   __ret; \
   52928 })
   52929 #else
   52930 #define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52931   uint32x4_t __s0 = __p0; \
   52932   uint32x4_t __s1 = __p1; \
   52933   uint32x4_t __s2 = __p2; \
   52934   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52935   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52936   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52937   uint32x4_t __ret; \
   52938   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   52939   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52940   __ret; \
   52941 })
   52942 #endif
   52943 
   52944 #ifdef __LITTLE_ENDIAN__
   52945 #define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52946   uint16x8_t __s0 = __p0; \
   52947   uint16x8_t __s1 = __p1; \
   52948   uint16x8_t __s2 = __p2; \
   52949   uint16x8_t __ret; \
   52950   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   52951   __ret; \
   52952 })
   52953 #else
   52954 #define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   52955   uint16x8_t __s0 = __p0; \
   52956   uint16x8_t __s1 = __p1; \
   52957   uint16x8_t __s2 = __p2; \
   52958   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   52959   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   52960   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   52961   uint16x8_t __ret; \
   52962   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   52963   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   52964   __ret; \
   52965 })
   52966 #endif
   52967 
   52968 #ifdef __LITTLE_ENDIAN__
   52969 #define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52970   float32x4_t __s0 = __p0; \
   52971   float32x4_t __s1 = __p1; \
   52972   float32x4_t __s2 = __p2; \
   52973   float32x4_t __ret; \
   52974   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   52975   __ret; \
   52976 })
   52977 #else
   52978 #define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52979   float32x4_t __s0 = __p0; \
   52980   float32x4_t __s1 = __p1; \
   52981   float32x4_t __s2 = __p2; \
   52982   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   52983   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   52984   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   52985   float32x4_t __ret; \
   52986   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   52987   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   52988   __ret; \
   52989 })
   52990 #endif
   52991 
   52992 #ifdef __LITTLE_ENDIAN__
   52993 #define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   52994   int32x4_t __s0 = __p0; \
   52995   int32x4_t __s1 = __p1; \
   52996   int32x4_t __s2 = __p2; \
   52997   int32x4_t __ret; \
   52998   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   52999   __ret; \
   53000 })
   53001 #else
   53002 #define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53003   int32x4_t __s0 = __p0; \
   53004   int32x4_t __s1 = __p1; \
   53005   int32x4_t __s2 = __p2; \
   53006   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53007   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53008   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53009   int32x4_t __ret; \
   53010   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   53011   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53012   __ret; \
   53013 })
   53014 #endif
   53015 
   53016 #ifdef __LITTLE_ENDIAN__
   53017 #define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53018   int16x8_t __s0 = __p0; \
   53019   int16x8_t __s1 = __p1; \
   53020   int16x8_t __s2 = __p2; \
   53021   int16x8_t __ret; \
   53022   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   53023   __ret; \
   53024 })
   53025 #else
   53026 #define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53027   int16x8_t __s0 = __p0; \
   53028   int16x8_t __s1 = __p1; \
   53029   int16x8_t __s2 = __p2; \
   53030   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   53031   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   53032   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   53033   int16x8_t __ret; \
   53034   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
   53035   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   53036   __ret; \
   53037 })
   53038 #endif
   53039 
   53040 #ifdef __LITTLE_ENDIAN__
   53041 #define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53042   uint32x2_t __s0 = __p0; \
   53043   uint32x2_t __s1 = __p1; \
   53044   uint32x4_t __s2 = __p2; \
   53045   uint32x2_t __ret; \
   53046   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   53047   __ret; \
   53048 })
   53049 #else
   53050 #define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53051   uint32x2_t __s0 = __p0; \
   53052   uint32x2_t __s1 = __p1; \
   53053   uint32x4_t __s2 = __p2; \
   53054   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53055   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   53056   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53057   uint32x2_t __ret; \
   53058   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   53059   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53060   __ret; \
   53061 })
   53062 #endif
   53063 
   53064 #ifdef __LITTLE_ENDIAN__
   53065 #define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53066   uint16x4_t __s0 = __p0; \
   53067   uint16x4_t __s1 = __p1; \
   53068   uint16x8_t __s2 = __p2; \
   53069   uint16x4_t __ret; \
   53070   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   53071   __ret; \
   53072 })
   53073 #else
   53074 #define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53075   uint16x4_t __s0 = __p0; \
   53076   uint16x4_t __s1 = __p1; \
   53077   uint16x8_t __s2 = __p2; \
   53078   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53079   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53080   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   53081   uint16x4_t __ret; \
   53082   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   53083   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53084   __ret; \
   53085 })
   53086 #endif
   53087 
   53088 #ifdef __LITTLE_ENDIAN__
   53089 #define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53090   float32x2_t __s0 = __p0; \
   53091   float32x2_t __s1 = __p1; \
   53092   float32x4_t __s2 = __p2; \
   53093   float32x2_t __ret; \
   53094   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   53095   __ret; \
   53096 })
   53097 #else
   53098 #define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53099   float32x2_t __s0 = __p0; \
   53100   float32x2_t __s1 = __p1; \
   53101   float32x4_t __s2 = __p2; \
   53102   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53103   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   53104   float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53105   float32x2_t __ret; \
   53106   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   53107   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53108   __ret; \
   53109 })
   53110 #endif
   53111 
   53112 #ifdef __LITTLE_ENDIAN__
   53113 #define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53114   int32x2_t __s0 = __p0; \
   53115   int32x2_t __s1 = __p1; \
   53116   int32x4_t __s2 = __p2; \
   53117   int32x2_t __ret; \
   53118   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
   53119   __ret; \
   53120 })
   53121 #else
   53122 #define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53123   int32x2_t __s0 = __p0; \
   53124   int32x2_t __s1 = __p1; \
   53125   int32x4_t __s2 = __p2; \
   53126   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53127   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   53128   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53129   int32x2_t __ret; \
   53130   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
   53131   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53132   __ret; \
   53133 })
   53134 #endif
   53135 
   53136 #ifdef __LITTLE_ENDIAN__
   53137 #define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53138   int16x4_t __s0 = __p0; \
   53139   int16x4_t __s1 = __p1; \
   53140   int16x8_t __s2 = __p2; \
   53141   int16x4_t __ret; \
   53142   __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
   53143   __ret; \
   53144 })
   53145 #else
   53146 #define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53147   int16x4_t __s0 = __p0; \
   53148   int16x4_t __s1 = __p1; \
   53149   int16x8_t __s2 = __p2; \
   53150   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53151   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53152   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   53153   int16x4_t __ret; \
   53154   __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
   53155   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53156   __ret; \
   53157 })
   53158 #endif
   53159 
   53160 #ifdef __LITTLE_ENDIAN__
   53161 __ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
   53162   float64x2_t __ret;
   53163   __ret = __p0 - __p1 * (float64x2_t) {__p2, __p2};
   53164   return __ret;
   53165 }
   53166 #else
   53167 __ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
   53168   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   53169   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   53170   float64x2_t __ret;
   53171   __ret = __rev0 - __rev1 * (float64x2_t) {__p2, __p2};
   53172   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   53173   return __ret;
   53174 }
   53175 #endif
   53176 
   53177 #ifdef __LITTLE_ENDIAN__
   53178 #define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53179   uint64x2_t __s0 = __p0; \
   53180   uint32x4_t __s1 = __p1; \
   53181   uint32x2_t __s2 = __p2; \
   53182   uint64x2_t __ret; \
   53183   __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   53184   __ret; \
   53185 })
   53186 #else
   53187 #define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53188   uint64x2_t __s0 = __p0; \
   53189   uint32x4_t __s1 = __p1; \
   53190   uint32x2_t __s2 = __p2; \
   53191   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53192   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53193   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   53194   uint64x2_t __ret; \
   53195   __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   53196   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53197   __ret; \
   53198 })
   53199 #endif
   53200 
   53201 #ifdef __LITTLE_ENDIAN__
   53202 #define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53203   uint32x4_t __s0 = __p0; \
   53204   uint16x8_t __s1 = __p1; \
   53205   uint16x4_t __s2 = __p2; \
   53206   uint32x4_t __ret; \
   53207   __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   53208   __ret; \
   53209 })
   53210 #else
   53211 #define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53212   uint32x4_t __s0 = __p0; \
   53213   uint16x8_t __s1 = __p1; \
   53214   uint16x4_t __s2 = __p2; \
   53215   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53216   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   53217   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53218   uint32x4_t __ret; \
   53219   __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   53220   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53221   __ret; \
   53222 })
   53223 #endif
   53224 
   53225 #ifdef __LITTLE_ENDIAN__
   53226 #define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53227   int64x2_t __s0 = __p0; \
   53228   int32x4_t __s1 = __p1; \
   53229   int32x2_t __s2 = __p2; \
   53230   int64x2_t __ret; \
   53231   __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   53232   __ret; \
   53233 })
   53234 #else
   53235 #define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53236   int64x2_t __s0 = __p0; \
   53237   int32x4_t __s1 = __p1; \
   53238   int32x2_t __s2 = __p2; \
   53239   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53240   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53241   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   53242   int64x2_t __ret; \
   53243   __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   53244   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53245   __ret; \
   53246 })
   53247 #endif
   53248 
   53249 #ifdef __LITTLE_ENDIAN__
   53250 #define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53251   int32x4_t __s0 = __p0; \
   53252   int16x8_t __s1 = __p1; \
   53253   int16x4_t __s2 = __p2; \
   53254   int32x4_t __ret; \
   53255   __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   53256   __ret; \
   53257 })
   53258 #else
   53259 #define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53260   int32x4_t __s0 = __p0; \
   53261   int16x8_t __s1 = __p1; \
   53262   int16x4_t __s2 = __p2; \
   53263   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53264   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   53265   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53266   int32x4_t __ret; \
   53267   __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   53268   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53269   __ret; \
   53270 })
   53271 #endif
   53272 
   53273 #ifdef __LITTLE_ENDIAN__
   53274 #define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53275   uint64x2_t __s0 = __p0; \
   53276   uint32x4_t __s1 = __p1; \
   53277   uint32x4_t __s2 = __p2; \
   53278   uint64x2_t __ret; \
   53279   __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   53280   __ret; \
   53281 })
   53282 #else
   53283 #define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53284   uint64x2_t __s0 = __p0; \
   53285   uint32x4_t __s1 = __p1; \
   53286   uint32x4_t __s2 = __p2; \
   53287   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53288   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53289   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53290   uint64x2_t __ret; \
   53291   __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   53292   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53293   __ret; \
   53294 })
   53295 #endif
   53296 
   53297 #ifdef __LITTLE_ENDIAN__
   53298 #define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53299   uint32x4_t __s0 = __p0; \
   53300   uint16x8_t __s1 = __p1; \
   53301   uint16x8_t __s2 = __p2; \
   53302   uint32x4_t __ret; \
   53303   __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   53304   __ret; \
   53305 })
   53306 #else
   53307 #define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53308   uint32x4_t __s0 = __p0; \
   53309   uint16x8_t __s1 = __p1; \
   53310   uint16x8_t __s2 = __p2; \
   53311   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53312   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   53313   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   53314   uint32x4_t __ret; \
   53315   __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   53316   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53317   __ret; \
   53318 })
   53319 #endif
   53320 
   53321 #ifdef __LITTLE_ENDIAN__
   53322 #define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53323   int64x2_t __s0 = __p0; \
   53324   int32x4_t __s1 = __p1; \
   53325   int32x4_t __s2 = __p2; \
   53326   int64x2_t __ret; \
   53327   __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   53328   __ret; \
   53329 })
   53330 #else
   53331 #define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53332   int64x2_t __s0 = __p0; \
   53333   int32x4_t __s1 = __p1; \
   53334   int32x4_t __s2 = __p2; \
   53335   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53336   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53337   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53338   int64x2_t __ret; \
   53339   __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   53340   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53341   __ret; \
   53342 })
   53343 #endif
   53344 
   53345 #ifdef __LITTLE_ENDIAN__
   53346 #define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53347   int32x4_t __s0 = __p0; \
   53348   int16x8_t __s1 = __p1; \
   53349   int16x8_t __s2 = __p2; \
   53350   int32x4_t __ret; \
   53351   __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   53352   __ret; \
   53353 })
   53354 #else
   53355 #define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53356   int32x4_t __s0 = __p0; \
   53357   int16x8_t __s1 = __p1; \
   53358   int16x8_t __s2 = __p2; \
   53359   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53360   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   53361   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   53362   int32x4_t __ret; \
   53363   __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   53364   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53365   __ret; \
   53366 })
   53367 #endif
   53368 
   53369 #ifdef __LITTLE_ENDIAN__
   53370 #define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53371   uint64x2_t __s0 = __p0; \
   53372   uint32x2_t __s1 = __p1; \
   53373   uint32x4_t __s2 = __p2; \
   53374   uint64x2_t __ret; \
   53375   __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   53376   __ret; \
   53377 })
   53378 #else
   53379 #define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53380   uint64x2_t __s0 = __p0; \
   53381   uint32x2_t __s1 = __p1; \
   53382   uint32x4_t __s2 = __p2; \
   53383   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53384   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   53385   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53386   uint64x2_t __ret; \
   53387   __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   53388   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53389   __ret; \
   53390 })
   53391 #endif
   53392 
   53393 #ifdef __LITTLE_ENDIAN__
   53394 #define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53395   uint32x4_t __s0 = __p0; \
   53396   uint16x4_t __s1 = __p1; \
   53397   uint16x8_t __s2 = __p2; \
   53398   uint32x4_t __ret; \
   53399   __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   53400   __ret; \
   53401 })
   53402 #else
   53403 #define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53404   uint32x4_t __s0 = __p0; \
   53405   uint16x4_t __s1 = __p1; \
   53406   uint16x8_t __s2 = __p2; \
   53407   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53408   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53409   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   53410   uint32x4_t __ret; \
   53411   __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   53412   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53413   __ret; \
   53414 })
   53415 #endif
   53416 
   53417 #ifdef __LITTLE_ENDIAN__
   53418 #define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53419   int64x2_t __s0 = __p0; \
   53420   int32x2_t __s1 = __p1; \
   53421   int32x4_t __s2 = __p2; \
   53422   int64x2_t __ret; \
   53423   __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   53424   __ret; \
   53425 })
   53426 #else
   53427 #define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   53428   int64x2_t __s0 = __p0; \
   53429   int32x2_t __s1 = __p1; \
   53430   int32x4_t __s2 = __p2; \
   53431   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53432   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   53433   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   53434   int64x2_t __ret; \
   53435   __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   53436   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53437   __ret; \
   53438 })
   53439 #endif
   53440 
   53441 #ifdef __LITTLE_ENDIAN__
   53442 #define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53443   int32x4_t __s0 = __p0; \
   53444   int16x4_t __s1 = __p1; \
   53445   int16x8_t __s2 = __p2; \
   53446   int32x4_t __ret; \
   53447   __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   53448   __ret; \
   53449 })
   53450 #else
   53451 #define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   53452   int32x4_t __s0 = __p0; \
   53453   int16x4_t __s1 = __p1; \
   53454   int16x8_t __s2 = __p2; \
   53455   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53456   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53457   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   53458   int32x4_t __ret; \
   53459   __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   53460   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53461   __ret; \
   53462 })
   53463 #endif
   53464 
   53465 #ifdef __LITTLE_ENDIAN__
   53466 __ai poly64x1_t vmov_n_p64(poly64_t __p0) {
   53467   poly64x1_t __ret;
   53468   __ret = (poly64x1_t) {__p0};
   53469   return __ret;
   53470 }
   53471 #else
   53472 __ai poly64x1_t vmov_n_p64(poly64_t __p0) {
   53473   poly64x1_t __ret;
   53474   __ret = (poly64x1_t) {__p0};
   53475   return __ret;
   53476 }
   53477 #endif
   53478 
   53479 #ifdef __LITTLE_ENDIAN__
   53480 __ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
   53481   poly64x2_t __ret;
   53482   __ret = (poly64x2_t) {__p0, __p0};
   53483   return __ret;
   53484 }
   53485 #else
   53486 __ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
   53487   poly64x2_t __ret;
   53488   __ret = (poly64x2_t) {__p0, __p0};
   53489   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   53490   return __ret;
   53491 }
   53492 #endif
   53493 
   53494 #ifdef __LITTLE_ENDIAN__
   53495 __ai float64x2_t vmovq_n_f64(float64_t __p0) {
   53496   float64x2_t __ret;
   53497   __ret = (float64x2_t) {__p0, __p0};
   53498   return __ret;
   53499 }
   53500 #else
   53501 __ai float64x2_t vmovq_n_f64(float64_t __p0) {
   53502   float64x2_t __ret;
   53503   __ret = (float64x2_t) {__p0, __p0};
   53504   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   53505   return __ret;
   53506 }
   53507 #endif
   53508 
   53509 #ifdef __LITTLE_ENDIAN__
   53510 __ai float64x1_t vmov_n_f64(float64_t __p0) {
   53511   float64x1_t __ret;
   53512   __ret = (float64x1_t) {__p0};
   53513   return __ret;
   53514 }
   53515 #else
   53516 __ai float64x1_t vmov_n_f64(float64_t __p0) {
   53517   float64x1_t __ret;
   53518   __ret = (float64x1_t) {__p0};
   53519   return __ret;
   53520 }
   53521 #endif
   53522 
   53523 #ifdef __LITTLE_ENDIAN__
   53524 __ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_112) {
   53525   uint16x8_t __ret_112;
   53526   uint8x8_t __a1_112 = vget_high_u8(__p0_112);
   53527   __ret_112 = (uint16x8_t)(vshll_n_u8(__a1_112, 0));
   53528   return __ret_112;
   53529 }
   53530 #else
   53531 __ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_113) {
   53532   uint8x16_t __rev0_113;  __rev0_113 = __builtin_shufflevector(__p0_113, __p0_113, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   53533   uint16x8_t __ret_113;
   53534   uint8x8_t __a1_113 = __noswap_vget_high_u8(__rev0_113);
   53535   __ret_113 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_113, 0));
   53536   __ret_113 = __builtin_shufflevector(__ret_113, __ret_113, 7, 6, 5, 4, 3, 2, 1, 0);
   53537   return __ret_113;
   53538 }
   53539 __ai uint16x8_t __noswap_vmovl_high_u8(uint8x16_t __p0_114) {
   53540   uint16x8_t __ret_114;
   53541   uint8x8_t __a1_114 = __noswap_vget_high_u8(__p0_114);
   53542   __ret_114 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_114, 0));
   53543   return __ret_114;
   53544 }
   53545 #endif
   53546 
   53547 #ifdef __LITTLE_ENDIAN__
   53548 __ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_115) {
   53549   uint64x2_t __ret_115;
   53550   uint32x2_t __a1_115 = vget_high_u32(__p0_115);
   53551   __ret_115 = (uint64x2_t)(vshll_n_u32(__a1_115, 0));
   53552   return __ret_115;
   53553 }
   53554 #else
   53555 __ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_116) {
   53556   uint32x4_t __rev0_116;  __rev0_116 = __builtin_shufflevector(__p0_116, __p0_116, 3, 2, 1, 0);
   53557   uint64x2_t __ret_116;
   53558   uint32x2_t __a1_116 = __noswap_vget_high_u32(__rev0_116);
   53559   __ret_116 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_116, 0));
   53560   __ret_116 = __builtin_shufflevector(__ret_116, __ret_116, 1, 0);
   53561   return __ret_116;
   53562 }
   53563 __ai uint64x2_t __noswap_vmovl_high_u32(uint32x4_t __p0_117) {
   53564   uint64x2_t __ret_117;
   53565   uint32x2_t __a1_117 = __noswap_vget_high_u32(__p0_117);
   53566   __ret_117 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_117, 0));
   53567   return __ret_117;
   53568 }
   53569 #endif
   53570 
   53571 #ifdef __LITTLE_ENDIAN__
   53572 __ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_118) {
   53573   uint32x4_t __ret_118;
   53574   uint16x4_t __a1_118 = vget_high_u16(__p0_118);
   53575   __ret_118 = (uint32x4_t)(vshll_n_u16(__a1_118, 0));
   53576   return __ret_118;
   53577 }
   53578 #else
   53579 __ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_119) {
   53580   uint16x8_t __rev0_119;  __rev0_119 = __builtin_shufflevector(__p0_119, __p0_119, 7, 6, 5, 4, 3, 2, 1, 0);
   53581   uint32x4_t __ret_119;
   53582   uint16x4_t __a1_119 = __noswap_vget_high_u16(__rev0_119);
   53583   __ret_119 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_119, 0));
   53584   __ret_119 = __builtin_shufflevector(__ret_119, __ret_119, 3, 2, 1, 0);
   53585   return __ret_119;
   53586 }
   53587 __ai uint32x4_t __noswap_vmovl_high_u16(uint16x8_t __p0_120) {
   53588   uint32x4_t __ret_120;
   53589   uint16x4_t __a1_120 = __noswap_vget_high_u16(__p0_120);
   53590   __ret_120 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_120, 0));
   53591   return __ret_120;
   53592 }
   53593 #endif
   53594 
   53595 #ifdef __LITTLE_ENDIAN__
   53596 __ai int16x8_t vmovl_high_s8(int8x16_t __p0_121) {
   53597   int16x8_t __ret_121;
   53598   int8x8_t __a1_121 = vget_high_s8(__p0_121);
   53599   __ret_121 = (int16x8_t)(vshll_n_s8(__a1_121, 0));
   53600   return __ret_121;
   53601 }
   53602 #else
   53603 __ai int16x8_t vmovl_high_s8(int8x16_t __p0_122) {
   53604   int8x16_t __rev0_122;  __rev0_122 = __builtin_shufflevector(__p0_122, __p0_122, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   53605   int16x8_t __ret_122;
   53606   int8x8_t __a1_122 = __noswap_vget_high_s8(__rev0_122);
   53607   __ret_122 = (int16x8_t)(__noswap_vshll_n_s8(__a1_122, 0));
   53608   __ret_122 = __builtin_shufflevector(__ret_122, __ret_122, 7, 6, 5, 4, 3, 2, 1, 0);
   53609   return __ret_122;
   53610 }
   53611 __ai int16x8_t __noswap_vmovl_high_s8(int8x16_t __p0_123) {
   53612   int16x8_t __ret_123;
   53613   int8x8_t __a1_123 = __noswap_vget_high_s8(__p0_123);
   53614   __ret_123 = (int16x8_t)(__noswap_vshll_n_s8(__a1_123, 0));
   53615   return __ret_123;
   53616 }
   53617 #endif
   53618 
   53619 #ifdef __LITTLE_ENDIAN__
   53620 __ai int64x2_t vmovl_high_s32(int32x4_t __p0_124) {
   53621   int64x2_t __ret_124;
   53622   int32x2_t __a1_124 = vget_high_s32(__p0_124);
   53623   __ret_124 = (int64x2_t)(vshll_n_s32(__a1_124, 0));
   53624   return __ret_124;
   53625 }
   53626 #else
   53627 __ai int64x2_t vmovl_high_s32(int32x4_t __p0_125) {
   53628   int32x4_t __rev0_125;  __rev0_125 = __builtin_shufflevector(__p0_125, __p0_125, 3, 2, 1, 0);
   53629   int64x2_t __ret_125;
   53630   int32x2_t __a1_125 = __noswap_vget_high_s32(__rev0_125);
   53631   __ret_125 = (int64x2_t)(__noswap_vshll_n_s32(__a1_125, 0));
   53632   __ret_125 = __builtin_shufflevector(__ret_125, __ret_125, 1, 0);
   53633   return __ret_125;
   53634 }
   53635 __ai int64x2_t __noswap_vmovl_high_s32(int32x4_t __p0_126) {
   53636   int64x2_t __ret_126;
   53637   int32x2_t __a1_126 = __noswap_vget_high_s32(__p0_126);
   53638   __ret_126 = (int64x2_t)(__noswap_vshll_n_s32(__a1_126, 0));
   53639   return __ret_126;
   53640 }
   53641 #endif
   53642 
   53643 #ifdef __LITTLE_ENDIAN__
   53644 __ai int32x4_t vmovl_high_s16(int16x8_t __p0_127) {
   53645   int32x4_t __ret_127;
   53646   int16x4_t __a1_127 = vget_high_s16(__p0_127);
   53647   __ret_127 = (int32x4_t)(vshll_n_s16(__a1_127, 0));
   53648   return __ret_127;
   53649 }
   53650 #else
   53651 __ai int32x4_t vmovl_high_s16(int16x8_t __p0_128) {
   53652   int16x8_t __rev0_128;  __rev0_128 = __builtin_shufflevector(__p0_128, __p0_128, 7, 6, 5, 4, 3, 2, 1, 0);
   53653   int32x4_t __ret_128;
   53654   int16x4_t __a1_128 = __noswap_vget_high_s16(__rev0_128);
   53655   __ret_128 = (int32x4_t)(__noswap_vshll_n_s16(__a1_128, 0));
   53656   __ret_128 = __builtin_shufflevector(__ret_128, __ret_128, 3, 2, 1, 0);
   53657   return __ret_128;
   53658 }
   53659 __ai int32x4_t __noswap_vmovl_high_s16(int16x8_t __p0_129) {
   53660   int32x4_t __ret_129;
   53661   int16x4_t __a1_129 = __noswap_vget_high_s16(__p0_129);
   53662   __ret_129 = (int32x4_t)(__noswap_vshll_n_s16(__a1_129, 0));
   53663   return __ret_129;
   53664 }
   53665 #endif
   53666 
   53667 #ifdef __LITTLE_ENDIAN__
   53668 __ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
   53669   uint16x8_t __ret;
   53670   __ret = vcombine_u16(__p0, vmovn_u32(__p1));
   53671   return __ret;
   53672 }
   53673 #else
   53674 __ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
   53675   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   53676   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   53677   uint16x8_t __ret;
   53678   __ret = __noswap_vcombine_u16(__rev0, __noswap_vmovn_u32(__rev1));
   53679   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   53680   return __ret;
   53681 }
   53682 #endif
   53683 
   53684 #ifdef __LITTLE_ENDIAN__
   53685 __ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
   53686   uint32x4_t __ret;
   53687   __ret = vcombine_u32(__p0, vmovn_u64(__p1));
   53688   return __ret;
   53689 }
   53690 #else
   53691 __ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
   53692   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   53693   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   53694   uint32x4_t __ret;
   53695   __ret = __noswap_vcombine_u32(__rev0, __noswap_vmovn_u64(__rev1));
   53696   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   53697   return __ret;
   53698 }
   53699 #endif
   53700 
   53701 #ifdef __LITTLE_ENDIAN__
   53702 __ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
   53703   uint8x16_t __ret;
   53704   __ret = vcombine_u8(__p0, vmovn_u16(__p1));
   53705   return __ret;
   53706 }
   53707 #else
   53708 __ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
   53709   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   53710   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   53711   uint8x16_t __ret;
   53712   __ret = __noswap_vcombine_u8(__rev0, __noswap_vmovn_u16(__rev1));
   53713   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   53714   return __ret;
   53715 }
   53716 #endif
   53717 
   53718 #ifdef __LITTLE_ENDIAN__
   53719 __ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
   53720   int16x8_t __ret;
   53721   __ret = vcombine_s16(__p0, vmovn_s32(__p1));
   53722   return __ret;
   53723 }
   53724 #else
   53725 __ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
   53726   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   53727   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   53728   int16x8_t __ret;
   53729   __ret = __noswap_vcombine_s16(__rev0, __noswap_vmovn_s32(__rev1));
   53730   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   53731   return __ret;
   53732 }
   53733 #endif
   53734 
   53735 #ifdef __LITTLE_ENDIAN__
   53736 __ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
   53737   int32x4_t __ret;
   53738   __ret = vcombine_s32(__p0, vmovn_s64(__p1));
   53739   return __ret;
   53740 }
   53741 #else
   53742 __ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
   53743   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   53744   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   53745   int32x4_t __ret;
   53746   __ret = __noswap_vcombine_s32(__rev0, __noswap_vmovn_s64(__rev1));
   53747   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   53748   return __ret;
   53749 }
   53750 #endif
   53751 
   53752 #ifdef __LITTLE_ENDIAN__
   53753 __ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
   53754   int8x16_t __ret;
   53755   __ret = vcombine_s8(__p0, vmovn_s16(__p1));
   53756   return __ret;
   53757 }
   53758 #else
   53759 __ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
   53760   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   53761   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   53762   int8x16_t __ret;
   53763   __ret = __noswap_vcombine_s8(__rev0, __noswap_vmovn_s16(__rev1));
   53764   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   53765   return __ret;
   53766 }
   53767 #endif
   53768 
   53769 #ifdef __LITTLE_ENDIAN__
   53770 __ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
   53771   float64x2_t __ret;
   53772   __ret = __p0 * __p1;
   53773   return __ret;
   53774 }
   53775 #else
   53776 __ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
   53777   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   53778   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   53779   float64x2_t __ret;
   53780   __ret = __rev0 * __rev1;
   53781   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   53782   return __ret;
   53783 }
   53784 #endif
   53785 
   53786 #ifdef __LITTLE_ENDIAN__
   53787 __ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
   53788   float64x1_t __ret;
   53789   __ret = __p0 * __p1;
   53790   return __ret;
   53791 }
   53792 #else
   53793 __ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
   53794   float64x1_t __ret;
   53795   __ret = __p0 * __p1;
   53796   return __ret;
   53797 }
   53798 #endif
   53799 
   53800 #ifdef __LITTLE_ENDIAN__
   53801 #define vmuld_lane_f64(__p0_130, __p1_130, __p2_130) __extension__ ({ \
   53802   float64_t __s0_130 = __p0_130; \
   53803   float64x1_t __s1_130 = __p1_130; \
   53804   float64_t __ret_130; \
   53805   __ret_130 = __s0_130 * vget_lane_f64(__s1_130, __p2_130); \
   53806   __ret_130; \
   53807 })
   53808 #else
   53809 #define vmuld_lane_f64(__p0_131, __p1_131, __p2_131) __extension__ ({ \
   53810   float64_t __s0_131 = __p0_131; \
   53811   float64x1_t __s1_131 = __p1_131; \
   53812   float64_t __ret_131; \
   53813   __ret_131 = __s0_131 * __noswap_vget_lane_f64(__s1_131, __p2_131); \
   53814   __ret_131; \
   53815 })
   53816 #endif
   53817 
   53818 #ifdef __LITTLE_ENDIAN__
   53819 #define vmuls_lane_f32(__p0_132, __p1_132, __p2_132) __extension__ ({ \
   53820   float32_t __s0_132 = __p0_132; \
   53821   float32x2_t __s1_132 = __p1_132; \
   53822   float32_t __ret_132; \
   53823   __ret_132 = __s0_132 * vget_lane_f32(__s1_132, __p2_132); \
   53824   __ret_132; \
   53825 })
   53826 #else
   53827 #define vmuls_lane_f32(__p0_133, __p1_133, __p2_133) __extension__ ({ \
   53828   float32_t __s0_133 = __p0_133; \
   53829   float32x2_t __s1_133 = __p1_133; \
   53830   float32x2_t __rev1_133;  __rev1_133 = __builtin_shufflevector(__s1_133, __s1_133, 1, 0); \
   53831   float32_t __ret_133; \
   53832   __ret_133 = __s0_133 * __noswap_vget_lane_f32(__rev1_133, __p2_133); \
   53833   __ret_133; \
   53834 })
   53835 #endif
   53836 
   53837 #ifdef __LITTLE_ENDIAN__
   53838 #define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   53839   float64x1_t __s0 = __p0; \
   53840   float64x1_t __s1 = __p1; \
   53841   float64x1_t __ret; \
   53842   __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
   53843   __ret; \
   53844 })
   53845 #else
   53846 #define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   53847   float64x1_t __s0 = __p0; \
   53848   float64x1_t __s1 = __p1; \
   53849   float64x1_t __ret; \
   53850   __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
   53851   __ret; \
   53852 })
   53853 #endif
   53854 
   53855 #ifdef __LITTLE_ENDIAN__
   53856 #define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   53857   float64x2_t __s0 = __p0; \
   53858   float64x1_t __s1 = __p1; \
   53859   float64x2_t __ret; \
   53860   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
   53861   __ret; \
   53862 })
   53863 #else
   53864 #define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   53865   float64x2_t __s0 = __p0; \
   53866   float64x1_t __s1 = __p1; \
   53867   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53868   float64x2_t __ret; \
   53869   __ret = __rev0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
   53870   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53871   __ret; \
   53872 })
   53873 #endif
   53874 
   53875 #ifdef __LITTLE_ENDIAN__
   53876 #define vmuld_laneq_f64(__p0_134, __p1_134, __p2_134) __extension__ ({ \
   53877   float64_t __s0_134 = __p0_134; \
   53878   float64x2_t __s1_134 = __p1_134; \
   53879   float64_t __ret_134; \
   53880   __ret_134 = __s0_134 * vgetq_lane_f64(__s1_134, __p2_134); \
   53881   __ret_134; \
   53882 })
   53883 #else
   53884 #define vmuld_laneq_f64(__p0_135, __p1_135, __p2_135) __extension__ ({ \
   53885   float64_t __s0_135 = __p0_135; \
   53886   float64x2_t __s1_135 = __p1_135; \
   53887   float64x2_t __rev1_135;  __rev1_135 = __builtin_shufflevector(__s1_135, __s1_135, 1, 0); \
   53888   float64_t __ret_135; \
   53889   __ret_135 = __s0_135 * __noswap_vgetq_lane_f64(__rev1_135, __p2_135); \
   53890   __ret_135; \
   53891 })
   53892 #endif
   53893 
   53894 #ifdef __LITTLE_ENDIAN__
   53895 #define vmuls_laneq_f32(__p0_136, __p1_136, __p2_136) __extension__ ({ \
   53896   float32_t __s0_136 = __p0_136; \
   53897   float32x4_t __s1_136 = __p1_136; \
   53898   float32_t __ret_136; \
   53899   __ret_136 = __s0_136 * vgetq_lane_f32(__s1_136, __p2_136); \
   53900   __ret_136; \
   53901 })
   53902 #else
   53903 #define vmuls_laneq_f32(__p0_137, __p1_137, __p2_137) __extension__ ({ \
   53904   float32_t __s0_137 = __p0_137; \
   53905   float32x4_t __s1_137 = __p1_137; \
   53906   float32x4_t __rev1_137;  __rev1_137 = __builtin_shufflevector(__s1_137, __s1_137, 3, 2, 1, 0); \
   53907   float32_t __ret_137; \
   53908   __ret_137 = __s0_137 * __noswap_vgetq_lane_f32(__rev1_137, __p2_137); \
   53909   __ret_137; \
   53910 })
   53911 #endif
   53912 
   53913 #ifdef __LITTLE_ENDIAN__
   53914 #define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
   53915   float64x1_t __s0 = __p0; \
   53916   float64x2_t __s1 = __p1; \
   53917   float64x1_t __ret; \
   53918   __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 10); \
   53919   __ret; \
   53920 })
   53921 #else
   53922 #define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
   53923   float64x1_t __s0 = __p0; \
   53924   float64x2_t __s1 = __p1; \
   53925   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   53926   float64x1_t __ret; \
   53927   __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__rev1, __p2, 10); \
   53928   __ret; \
   53929 })
   53930 #endif
   53931 
   53932 #ifdef __LITTLE_ENDIAN__
   53933 #define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
   53934   uint32x4_t __s0 = __p0; \
   53935   uint32x4_t __s1 = __p1; \
   53936   uint32x4_t __ret; \
   53937   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   53938   __ret; \
   53939 })
   53940 #else
   53941 #define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
   53942   uint32x4_t __s0 = __p0; \
   53943   uint32x4_t __s1 = __p1; \
   53944   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   53945   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   53946   uint32x4_t __ret; \
   53947   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   53948   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   53949   __ret; \
   53950 })
   53951 #endif
   53952 
   53953 #ifdef __LITTLE_ENDIAN__
   53954 #define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
   53955   uint16x8_t __s0 = __p0; \
   53956   uint16x8_t __s1 = __p1; \
   53957   uint16x8_t __ret; \
   53958   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
   53959   __ret; \
   53960 })
   53961 #else
   53962 #define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
   53963   uint16x8_t __s0 = __p0; \
   53964   uint16x8_t __s1 = __p1; \
   53965   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   53966   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   53967   uint16x8_t __ret; \
   53968   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
   53969   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   53970   __ret; \
   53971 })
   53972 #endif
   53973 
   53974 #ifdef __LITTLE_ENDIAN__
   53975 #define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
   53976   float64x2_t __s0 = __p0; \
   53977   float64x2_t __s1 = __p1; \
   53978   float64x2_t __ret; \
   53979   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
   53980   __ret; \
   53981 })
   53982 #else
   53983 #define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
   53984   float64x2_t __s0 = __p0; \
   53985   float64x2_t __s1 = __p1; \
   53986   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   53987   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   53988   float64x2_t __ret; \
   53989   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
   53990   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   53991   __ret; \
   53992 })
   53993 #endif
   53994 
   53995 #ifdef __LITTLE_ENDIAN__
   53996 #define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
   53997   float32x4_t __s0 = __p0; \
   53998   float32x4_t __s1 = __p1; \
   53999   float32x4_t __ret; \
   54000   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   54001   __ret; \
   54002 })
   54003 #else
   54004 #define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
   54005   float32x4_t __s0 = __p0; \
   54006   float32x4_t __s1 = __p1; \
   54007   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54008   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54009   float32x4_t __ret; \
   54010   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   54011   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54012   __ret; \
   54013 })
   54014 #endif
   54015 
   54016 #ifdef __LITTLE_ENDIAN__
   54017 #define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   54018   int32x4_t __s0 = __p0; \
   54019   int32x4_t __s1 = __p1; \
   54020   int32x4_t __ret; \
   54021   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   54022   __ret; \
   54023 })
   54024 #else
   54025 #define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   54026   int32x4_t __s0 = __p0; \
   54027   int32x4_t __s1 = __p1; \
   54028   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54029   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54030   int32x4_t __ret; \
   54031   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   54032   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54033   __ret; \
   54034 })
   54035 #endif
   54036 
   54037 #ifdef __LITTLE_ENDIAN__
   54038 #define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   54039   int16x8_t __s0 = __p0; \
   54040   int16x8_t __s1 = __p1; \
   54041   int16x8_t __ret; \
   54042   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
   54043   __ret; \
   54044 })
   54045 #else
   54046 #define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   54047   int16x8_t __s0 = __p0; \
   54048   int16x8_t __s1 = __p1; \
   54049   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   54050   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   54051   int16x8_t __ret; \
   54052   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
   54053   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   54054   __ret; \
   54055 })
   54056 #endif
   54057 
   54058 #ifdef __LITTLE_ENDIAN__
   54059 #define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
   54060   uint32x2_t __s0 = __p0; \
   54061   uint32x4_t __s1 = __p1; \
   54062   uint32x2_t __ret; \
   54063   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
   54064   __ret; \
   54065 })
   54066 #else
   54067 #define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
   54068   uint32x2_t __s0 = __p0; \
   54069   uint32x4_t __s1 = __p1; \
   54070   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   54071   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54072   uint32x2_t __ret; \
   54073   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
   54074   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54075   __ret; \
   54076 })
   54077 #endif
   54078 
   54079 #ifdef __LITTLE_ENDIAN__
   54080 #define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
   54081   uint16x4_t __s0 = __p0; \
   54082   uint16x8_t __s1 = __p1; \
   54083   uint16x4_t __ret; \
   54084   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   54085   __ret; \
   54086 })
   54087 #else
   54088 #define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
   54089   uint16x4_t __s0 = __p0; \
   54090   uint16x8_t __s1 = __p1; \
   54091   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54092   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   54093   uint16x4_t __ret; \
   54094   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   54095   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54096   __ret; \
   54097 })
   54098 #endif
   54099 
   54100 #ifdef __LITTLE_ENDIAN__
   54101 #define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
   54102   float32x2_t __s0 = __p0; \
   54103   float32x4_t __s1 = __p1; \
   54104   float32x2_t __ret; \
   54105   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
   54106   __ret; \
   54107 })
   54108 #else
   54109 #define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
   54110   float32x2_t __s0 = __p0; \
   54111   float32x4_t __s1 = __p1; \
   54112   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   54113   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54114   float32x2_t __ret; \
   54115   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
   54116   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54117   __ret; \
   54118 })
   54119 #endif
   54120 
   54121 #ifdef __LITTLE_ENDIAN__
   54122 #define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   54123   int32x2_t __s0 = __p0; \
   54124   int32x4_t __s1 = __p1; \
   54125   int32x2_t __ret; \
   54126   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
   54127   __ret; \
   54128 })
   54129 #else
   54130 #define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   54131   int32x2_t __s0 = __p0; \
   54132   int32x4_t __s1 = __p1; \
   54133   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   54134   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54135   int32x2_t __ret; \
   54136   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
   54137   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54138   __ret; \
   54139 })
   54140 #endif
   54141 
   54142 #ifdef __LITTLE_ENDIAN__
   54143 #define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   54144   int16x4_t __s0 = __p0; \
   54145   int16x8_t __s1 = __p1; \
   54146   int16x4_t __ret; \
   54147   __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
   54148   __ret; \
   54149 })
   54150 #else
   54151 #define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   54152   int16x4_t __s0 = __p0; \
   54153   int16x8_t __s1 = __p1; \
   54154   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54155   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   54156   int16x4_t __ret; \
   54157   __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
   54158   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54159   __ret; \
   54160 })
   54161 #endif
   54162 
   54163 #ifdef __LITTLE_ENDIAN__
   54164 __ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
   54165   float64x1_t __ret;
   54166   __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
   54167   return __ret;
   54168 }
   54169 #else
   54170 __ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
   54171   float64x1_t __ret;
   54172   __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
   54173   return __ret;
   54174 }
   54175 #endif
   54176 
   54177 #ifdef __LITTLE_ENDIAN__
   54178 __ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
   54179   float64x2_t __ret;
   54180   __ret = __p0 * (float64x2_t) {__p1, __p1};
   54181   return __ret;
   54182 }
   54183 #else
   54184 __ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
   54185   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   54186   float64x2_t __ret;
   54187   __ret = __rev0 * (float64x2_t) {__p1, __p1};
   54188   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   54189   return __ret;
   54190 }
   54191 #endif
   54192 
   54193 #ifdef __LITTLE_ENDIAN__
   54194 __ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
   54195   poly128_t __ret;
   54196   __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
   54197   return __ret;
   54198 }
   54199 #else
   54200 __ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
   54201   poly128_t __ret;
   54202   __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
   54203   return __ret;
   54204 }
   54205 __ai poly128_t __noswap_vmull_p64(poly64_t __p0, poly64_t __p1) {
   54206   poly128_t __ret;
   54207   __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
   54208   return __ret;
   54209 }
   54210 #endif
   54211 
   54212 #ifdef __LITTLE_ENDIAN__
   54213 __ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
   54214   poly16x8_t __ret;
   54215   __ret = vmull_p8(vget_high_p8(__p0), vget_high_p8(__p1));
   54216   return __ret;
   54217 }
   54218 #else
   54219 __ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
   54220   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   54221   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   54222   poly16x8_t __ret;
   54223   __ret = __noswap_vmull_p8(__noswap_vget_high_p8(__rev0), __noswap_vget_high_p8(__rev1));
   54224   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   54225   return __ret;
   54226 }
   54227 #endif
   54228 
   54229 #ifdef __LITTLE_ENDIAN__
   54230 __ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
   54231   uint16x8_t __ret;
   54232   __ret = vmull_u8(vget_high_u8(__p0), vget_high_u8(__p1));
   54233   return __ret;
   54234 }
   54235 #else
   54236 __ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
   54237   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   54238   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   54239   uint16x8_t __ret;
   54240   __ret = __noswap_vmull_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
   54241   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   54242   return __ret;
   54243 }
   54244 #endif
   54245 
   54246 #ifdef __LITTLE_ENDIAN__
   54247 __ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
   54248   uint64x2_t __ret;
   54249   __ret = vmull_u32(vget_high_u32(__p0), vget_high_u32(__p1));
   54250   return __ret;
   54251 }
   54252 #else
   54253 __ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
   54254   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   54255   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   54256   uint64x2_t __ret;
   54257   __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
   54258   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   54259   return __ret;
   54260 }
   54261 #endif
   54262 
   54263 #ifdef __LITTLE_ENDIAN__
   54264 __ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
   54265   uint32x4_t __ret;
   54266   __ret = vmull_u16(vget_high_u16(__p0), vget_high_u16(__p1));
   54267   return __ret;
   54268 }
   54269 #else
   54270 __ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
   54271   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   54272   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   54273   uint32x4_t __ret;
   54274   __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
   54275   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   54276   return __ret;
   54277 }
   54278 #endif
   54279 
   54280 #ifdef __LITTLE_ENDIAN__
   54281 __ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
   54282   int16x8_t __ret;
   54283   __ret = vmull_s8(vget_high_s8(__p0), vget_high_s8(__p1));
   54284   return __ret;
   54285 }
   54286 #else
   54287 __ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
   54288   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   54289   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   54290   int16x8_t __ret;
   54291   __ret = __noswap_vmull_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
   54292   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   54293   return __ret;
   54294 }
   54295 #endif
   54296 
   54297 #ifdef __LITTLE_ENDIAN__
   54298 __ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
   54299   int64x2_t __ret;
   54300   __ret = vmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
   54301   return __ret;
   54302 }
   54303 #else
   54304 __ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
   54305   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   54306   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   54307   int64x2_t __ret;
   54308   __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
   54309   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   54310   return __ret;
   54311 }
   54312 #endif
   54313 
   54314 #ifdef __LITTLE_ENDIAN__
   54315 __ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
   54316   int32x4_t __ret;
   54317   __ret = vmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
   54318   return __ret;
   54319 }
   54320 #else
   54321 __ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
   54322   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   54323   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   54324   int32x4_t __ret;
   54325   __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
   54326   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   54327   return __ret;
   54328 }
   54329 #endif
   54330 
   54331 #ifdef __LITTLE_ENDIAN__
   54332 __ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
   54333   poly128_t __ret;
   54334   __ret = vmull_p64((poly64_t)(vget_high_p64(__p0)), (poly64_t)(vget_high_p64(__p1)));
   54335   return __ret;
   54336 }
   54337 #else
   54338 __ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
   54339   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   54340   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   54341   poly128_t __ret;
   54342   __ret = __noswap_vmull_p64((poly64_t)(__noswap_vget_high_p64(__rev0)), (poly64_t)(__noswap_vget_high_p64(__rev1)));
   54343   return __ret;
   54344 }
   54345 #endif
   54346 
   54347 #ifdef __LITTLE_ENDIAN__
   54348 #define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   54349   uint32x4_t __s0 = __p0; \
   54350   uint32x2_t __s1 = __p1; \
   54351   uint64x2_t __ret; \
   54352   __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54353   __ret; \
   54354 })
   54355 #else
   54356 #define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
   54357   uint32x4_t __s0 = __p0; \
   54358   uint32x2_t __s1 = __p1; \
   54359   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54360   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   54361   uint64x2_t __ret; \
   54362   __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   54363   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54364   __ret; \
   54365 })
   54366 #endif
   54367 
   54368 #ifdef __LITTLE_ENDIAN__
   54369 #define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   54370   uint16x8_t __s0 = __p0; \
   54371   uint16x4_t __s1 = __p1; \
   54372   uint32x4_t __ret; \
   54373   __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   54374   __ret; \
   54375 })
   54376 #else
   54377 #define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
   54378   uint16x8_t __s0 = __p0; \
   54379   uint16x4_t __s1 = __p1; \
   54380   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   54381   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54382   uint32x4_t __ret; \
   54383   __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   54384   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54385   __ret; \
   54386 })
   54387 #endif
   54388 
   54389 #ifdef __LITTLE_ENDIAN__
   54390 #define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   54391   int32x4_t __s0 = __p0; \
   54392   int32x2_t __s1 = __p1; \
   54393   int64x2_t __ret; \
   54394   __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54395   __ret; \
   54396 })
   54397 #else
   54398 #define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   54399   int32x4_t __s0 = __p0; \
   54400   int32x2_t __s1 = __p1; \
   54401   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54402   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   54403   int64x2_t __ret; \
   54404   __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   54405   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54406   __ret; \
   54407 })
   54408 #endif
   54409 
   54410 #ifdef __LITTLE_ENDIAN__
   54411 #define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   54412   int16x8_t __s0 = __p0; \
   54413   int16x4_t __s1 = __p1; \
   54414   int32x4_t __ret; \
   54415   __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   54416   __ret; \
   54417 })
   54418 #else
   54419 #define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   54420   int16x8_t __s0 = __p0; \
   54421   int16x4_t __s1 = __p1; \
   54422   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   54423   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54424   int32x4_t __ret; \
   54425   __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   54426   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54427   __ret; \
   54428 })
   54429 #endif
   54430 
   54431 #ifdef __LITTLE_ENDIAN__
   54432 #define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
   54433   uint32x4_t __s0 = __p0; \
   54434   uint32x4_t __s1 = __p1; \
   54435   uint64x2_t __ret; \
   54436   __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54437   __ret; \
   54438 })
   54439 #else
   54440 #define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
   54441   uint32x4_t __s0 = __p0; \
   54442   uint32x4_t __s1 = __p1; \
   54443   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54444   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54445   uint64x2_t __ret; \
   54446   __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   54447   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54448   __ret; \
   54449 })
   54450 #endif
   54451 
   54452 #ifdef __LITTLE_ENDIAN__
   54453 #define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
   54454   uint16x8_t __s0 = __p0; \
   54455   uint16x8_t __s1 = __p1; \
   54456   uint32x4_t __ret; \
   54457   __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   54458   __ret; \
   54459 })
   54460 #else
   54461 #define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
   54462   uint16x8_t __s0 = __p0; \
   54463   uint16x8_t __s1 = __p1; \
   54464   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   54465   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   54466   uint32x4_t __ret; \
   54467   __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   54468   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54469   __ret; \
   54470 })
   54471 #endif
   54472 
   54473 #ifdef __LITTLE_ENDIAN__
   54474 #define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   54475   int32x4_t __s0 = __p0; \
   54476   int32x4_t __s1 = __p1; \
   54477   int64x2_t __ret; \
   54478   __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54479   __ret; \
   54480 })
   54481 #else
   54482 #define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   54483   int32x4_t __s0 = __p0; \
   54484   int32x4_t __s1 = __p1; \
   54485   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54486   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54487   int64x2_t __ret; \
   54488   __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   54489   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54490   __ret; \
   54491 })
   54492 #endif
   54493 
   54494 #ifdef __LITTLE_ENDIAN__
   54495 #define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   54496   int16x8_t __s0 = __p0; \
   54497   int16x8_t __s1 = __p1; \
   54498   int32x4_t __ret; \
   54499   __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   54500   __ret; \
   54501 })
   54502 #else
   54503 #define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   54504   int16x8_t __s0 = __p0; \
   54505   int16x8_t __s1 = __p1; \
   54506   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   54507   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   54508   int32x4_t __ret; \
   54509   __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   54510   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54511   __ret; \
   54512 })
   54513 #endif
   54514 
   54515 #ifdef __LITTLE_ENDIAN__
   54516 __ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
   54517   uint64x2_t __ret;
   54518   __ret = vmull_n_u32(vget_high_u32(__p0), __p1);
   54519   return __ret;
   54520 }
   54521 #else
   54522 __ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
   54523   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   54524   uint64x2_t __ret;
   54525   __ret = __noswap_vmull_n_u32(__noswap_vget_high_u32(__rev0), __p1);
   54526   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   54527   return __ret;
   54528 }
   54529 #endif
   54530 
   54531 #ifdef __LITTLE_ENDIAN__
   54532 __ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
   54533   uint32x4_t __ret;
   54534   __ret = vmull_n_u16(vget_high_u16(__p0), __p1);
   54535   return __ret;
   54536 }
   54537 #else
   54538 __ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
   54539   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   54540   uint32x4_t __ret;
   54541   __ret = __noswap_vmull_n_u16(__noswap_vget_high_u16(__rev0), __p1);
   54542   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   54543   return __ret;
   54544 }
   54545 #endif
   54546 
   54547 #ifdef __LITTLE_ENDIAN__
   54548 __ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
   54549   int64x2_t __ret;
   54550   __ret = vmull_n_s32(vget_high_s32(__p0), __p1);
   54551   return __ret;
   54552 }
   54553 #else
   54554 __ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
   54555   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   54556   int64x2_t __ret;
   54557   __ret = __noswap_vmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
   54558   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   54559   return __ret;
   54560 }
   54561 #endif
   54562 
   54563 #ifdef __LITTLE_ENDIAN__
   54564 __ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
   54565   int32x4_t __ret;
   54566   __ret = vmull_n_s16(vget_high_s16(__p0), __p1);
   54567   return __ret;
   54568 }
   54569 #else
   54570 __ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
   54571   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   54572   int32x4_t __ret;
   54573   __ret = __noswap_vmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
   54574   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   54575   return __ret;
   54576 }
   54577 #endif
   54578 
   54579 #ifdef __LITTLE_ENDIAN__
   54580 #define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
   54581   uint32x2_t __s0 = __p0; \
   54582   uint32x4_t __s1 = __p1; \
   54583   uint64x2_t __ret; \
   54584   __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54585   __ret; \
   54586 })
   54587 #else
   54588 #define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
   54589   uint32x2_t __s0 = __p0; \
   54590   uint32x4_t __s1 = __p1; \
   54591   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   54592   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54593   uint64x2_t __ret; \
   54594   __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   54595   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54596   __ret; \
   54597 })
   54598 #endif
   54599 
   54600 #ifdef __LITTLE_ENDIAN__
   54601 #define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
   54602   uint16x4_t __s0 = __p0; \
   54603   uint16x8_t __s1 = __p1; \
   54604   uint32x4_t __ret; \
   54605   __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   54606   __ret; \
   54607 })
   54608 #else
   54609 #define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
   54610   uint16x4_t __s0 = __p0; \
   54611   uint16x8_t __s1 = __p1; \
   54612   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54613   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   54614   uint32x4_t __ret; \
   54615   __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   54616   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54617   __ret; \
   54618 })
   54619 #endif
   54620 
   54621 #ifdef __LITTLE_ENDIAN__
   54622 #define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   54623   int32x2_t __s0 = __p0; \
   54624   int32x4_t __s1 = __p1; \
   54625   int64x2_t __ret; \
   54626   __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54627   __ret; \
   54628 })
   54629 #else
   54630 #define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   54631   int32x2_t __s0 = __p0; \
   54632   int32x4_t __s1 = __p1; \
   54633   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   54634   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54635   int64x2_t __ret; \
   54636   __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   54637   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54638   __ret; \
   54639 })
   54640 #endif
   54641 
   54642 #ifdef __LITTLE_ENDIAN__
   54643 #define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   54644   int16x4_t __s0 = __p0; \
   54645   int16x8_t __s1 = __p1; \
   54646   int32x4_t __ret; \
   54647   __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   54648   __ret; \
   54649 })
   54650 #else
   54651 #define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   54652   int16x4_t __s0 = __p0; \
   54653   int16x8_t __s1 = __p1; \
   54654   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54655   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   54656   int32x4_t __ret; \
   54657   __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   54658   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54659   __ret; \
   54660 })
   54661 #endif
   54662 
   54663 #ifdef __LITTLE_ENDIAN__
   54664 __ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
   54665   float64x2_t __ret;
   54666   __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   54667   return __ret;
   54668 }
   54669 #else
   54670 __ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
   54671   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   54672   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   54673   float64x2_t __ret;
   54674   __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   54675   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   54676   return __ret;
   54677 }
   54678 __ai float64x2_t __noswap_vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
   54679   float64x2_t __ret;
   54680   __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   54681   return __ret;
   54682 }
   54683 #endif
   54684 
   54685 #ifdef __LITTLE_ENDIAN__
   54686 __ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
   54687   float32x4_t __ret;
   54688   __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   54689   return __ret;
   54690 }
   54691 #else
   54692 __ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
   54693   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   54694   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   54695   float32x4_t __ret;
   54696   __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   54697   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   54698   return __ret;
   54699 }
   54700 __ai float32x4_t __noswap_vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
   54701   float32x4_t __ret;
   54702   __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   54703   return __ret;
   54704 }
   54705 #endif
   54706 
   54707 #ifdef __LITTLE_ENDIAN__
   54708 __ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
   54709   float64x1_t __ret;
   54710   __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   54711   return __ret;
   54712 }
   54713 #else
   54714 __ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
   54715   float64x1_t __ret;
   54716   __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   54717   return __ret;
   54718 }
   54719 #endif
   54720 
   54721 #ifdef __LITTLE_ENDIAN__
   54722 __ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
   54723   float32x2_t __ret;
   54724   __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   54725   return __ret;
   54726 }
   54727 #else
   54728 __ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
   54729   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   54730   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   54731   float32x2_t __ret;
   54732   __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   54733   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   54734   return __ret;
   54735 }
   54736 __ai float32x2_t __noswap_vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
   54737   float32x2_t __ret;
   54738   __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   54739   return __ret;
   54740 }
   54741 #endif
   54742 
   54743 #ifdef __LITTLE_ENDIAN__
   54744 __ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
   54745   float64_t __ret;
   54746   __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
   54747   return __ret;
   54748 }
   54749 #else
   54750 __ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
   54751   float64_t __ret;
   54752   __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
   54753   return __ret;
   54754 }
   54755 __ai float64_t __noswap_vmulxd_f64(float64_t __p0, float64_t __p1) {
   54756   float64_t __ret;
   54757   __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
   54758   return __ret;
   54759 }
   54760 #endif
   54761 
   54762 #ifdef __LITTLE_ENDIAN__
   54763 __ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
   54764   float32_t __ret;
   54765   __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
   54766   return __ret;
   54767 }
   54768 #else
   54769 __ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
   54770   float32_t __ret;
   54771   __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
   54772   return __ret;
   54773 }
   54774 __ai float32_t __noswap_vmulxs_f32(float32_t __p0, float32_t __p1) {
   54775   float32_t __ret;
   54776   __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
   54777   return __ret;
   54778 }
   54779 #endif
   54780 
   54781 #ifdef __LITTLE_ENDIAN__
   54782 #define vmulxd_lane_f64(__p0_138, __p1_138, __p2_138) __extension__ ({ \
   54783   float64_t __s0_138 = __p0_138; \
   54784   float64x1_t __s1_138 = __p1_138; \
   54785   float64_t __ret_138; \
   54786   __ret_138 = vmulxd_f64(__s0_138, vget_lane_f64(__s1_138, __p2_138)); \
   54787   __ret_138; \
   54788 })
   54789 #else
   54790 #define vmulxd_lane_f64(__p0_139, __p1_139, __p2_139) __extension__ ({ \
   54791   float64_t __s0_139 = __p0_139; \
   54792   float64x1_t __s1_139 = __p1_139; \
   54793   float64_t __ret_139; \
   54794   __ret_139 = __noswap_vmulxd_f64(__s0_139, __noswap_vget_lane_f64(__s1_139, __p2_139)); \
   54795   __ret_139; \
   54796 })
   54797 #endif
   54798 
   54799 #ifdef __LITTLE_ENDIAN__
   54800 #define vmulxs_lane_f32(__p0_140, __p1_140, __p2_140) __extension__ ({ \
   54801   float32_t __s0_140 = __p0_140; \
   54802   float32x2_t __s1_140 = __p1_140; \
   54803   float32_t __ret_140; \
   54804   __ret_140 = vmulxs_f32(__s0_140, vget_lane_f32(__s1_140, __p2_140)); \
   54805   __ret_140; \
   54806 })
   54807 #else
   54808 #define vmulxs_lane_f32(__p0_141, __p1_141, __p2_141) __extension__ ({ \
   54809   float32_t __s0_141 = __p0_141; \
   54810   float32x2_t __s1_141 = __p1_141; \
   54811   float32x2_t __rev1_141;  __rev1_141 = __builtin_shufflevector(__s1_141, __s1_141, 1, 0); \
   54812   float32_t __ret_141; \
   54813   __ret_141 = __noswap_vmulxs_f32(__s0_141, __noswap_vget_lane_f32(__rev1_141, __p2_141)); \
   54814   __ret_141; \
   54815 })
   54816 #endif
   54817 
   54818 #ifdef __LITTLE_ENDIAN__
   54819 #define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   54820   float64x2_t __s0 = __p0; \
   54821   float64x1_t __s1 = __p1; \
   54822   float64x2_t __ret; \
   54823   __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54824   __ret; \
   54825 })
   54826 #else
   54827 #define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   54828   float64x2_t __s0 = __p0; \
   54829   float64x1_t __s1 = __p1; \
   54830   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   54831   float64x2_t __ret; \
   54832   __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54833   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54834   __ret; \
   54835 })
   54836 #endif
   54837 
   54838 #ifdef __LITTLE_ENDIAN__
   54839 #define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   54840   float32x4_t __s0 = __p0; \
   54841   float32x2_t __s1 = __p1; \
   54842   float32x4_t __ret; \
   54843   __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   54844   __ret; \
   54845 })
   54846 #else
   54847 #define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   54848   float32x4_t __s0 = __p0; \
   54849   float32x2_t __s1 = __p1; \
   54850   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54851   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   54852   float32x4_t __ret; \
   54853   __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   54854   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54855   __ret; \
   54856 })
   54857 #endif
   54858 
   54859 #ifdef __LITTLE_ENDIAN__
   54860 #define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   54861   float32x2_t __s0 = __p0; \
   54862   float32x2_t __s1 = __p1; \
   54863   float32x2_t __ret; \
   54864   __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54865   __ret; \
   54866 })
   54867 #else
   54868 #define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
   54869   float32x2_t __s0 = __p0; \
   54870   float32x2_t __s1 = __p1; \
   54871   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   54872   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   54873   float32x2_t __ret; \
   54874   __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   54875   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54876   __ret; \
   54877 })
   54878 #endif
   54879 
   54880 #ifdef __LITTLE_ENDIAN__
   54881 #define vmulxd_laneq_f64(__p0_142, __p1_142, __p2_142) __extension__ ({ \
   54882   float64_t __s0_142 = __p0_142; \
   54883   float64x2_t __s1_142 = __p1_142; \
   54884   float64_t __ret_142; \
   54885   __ret_142 = vmulxd_f64(__s0_142, vgetq_lane_f64(__s1_142, __p2_142)); \
   54886   __ret_142; \
   54887 })
   54888 #else
   54889 #define vmulxd_laneq_f64(__p0_143, __p1_143, __p2_143) __extension__ ({ \
   54890   float64_t __s0_143 = __p0_143; \
   54891   float64x2_t __s1_143 = __p1_143; \
   54892   float64x2_t __rev1_143;  __rev1_143 = __builtin_shufflevector(__s1_143, __s1_143, 1, 0); \
   54893   float64_t __ret_143; \
   54894   __ret_143 = __noswap_vmulxd_f64(__s0_143, __noswap_vgetq_lane_f64(__rev1_143, __p2_143)); \
   54895   __ret_143; \
   54896 })
   54897 #endif
   54898 
   54899 #ifdef __LITTLE_ENDIAN__
   54900 #define vmulxs_laneq_f32(__p0_144, __p1_144, __p2_144) __extension__ ({ \
   54901   float32_t __s0_144 = __p0_144; \
   54902   float32x4_t __s1_144 = __p1_144; \
   54903   float32_t __ret_144; \
   54904   __ret_144 = vmulxs_f32(__s0_144, vgetq_lane_f32(__s1_144, __p2_144)); \
   54905   __ret_144; \
   54906 })
   54907 #else
   54908 #define vmulxs_laneq_f32(__p0_145, __p1_145, __p2_145) __extension__ ({ \
   54909   float32_t __s0_145 = __p0_145; \
   54910   float32x4_t __s1_145 = __p1_145; \
   54911   float32x4_t __rev1_145;  __rev1_145 = __builtin_shufflevector(__s1_145, __s1_145, 3, 2, 1, 0); \
   54912   float32_t __ret_145; \
   54913   __ret_145 = __noswap_vmulxs_f32(__s0_145, __noswap_vgetq_lane_f32(__rev1_145, __p2_145)); \
   54914   __ret_145; \
   54915 })
   54916 #endif
   54917 
   54918 #ifdef __LITTLE_ENDIAN__
   54919 #define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
   54920   float64x2_t __s0 = __p0; \
   54921   float64x2_t __s1 = __p1; \
   54922   float64x2_t __ret; \
   54923   __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54924   __ret; \
   54925 })
   54926 #else
   54927 #define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
   54928   float64x2_t __s0 = __p0; \
   54929   float64x2_t __s1 = __p1; \
   54930   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   54931   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   54932   float64x2_t __ret; \
   54933   __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   54934   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54935   __ret; \
   54936 })
   54937 #endif
   54938 
   54939 #ifdef __LITTLE_ENDIAN__
   54940 #define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
   54941   float32x4_t __s0 = __p0; \
   54942   float32x4_t __s1 = __p1; \
   54943   float32x4_t __ret; \
   54944   __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   54945   __ret; \
   54946 })
   54947 #else
   54948 #define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
   54949   float32x4_t __s0 = __p0; \
   54950   float32x4_t __s1 = __p1; \
   54951   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   54952   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54953   float32x4_t __ret; \
   54954   __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   54955   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   54956   __ret; \
   54957 })
   54958 #endif
   54959 
   54960 #ifdef __LITTLE_ENDIAN__
   54961 #define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
   54962   float32x2_t __s0 = __p0; \
   54963   float32x4_t __s1 = __p1; \
   54964   float32x2_t __ret; \
   54965   __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   54966   __ret; \
   54967 })
   54968 #else
   54969 #define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
   54970   float32x2_t __s0 = __p0; \
   54971   float32x4_t __s1 = __p1; \
   54972   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   54973   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   54974   float32x2_t __ret; \
   54975   __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   54976   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   54977   __ret; \
   54978 })
   54979 #endif
   54980 
   54981 #ifdef __LITTLE_ENDIAN__
   54982 __ai float64x2_t vnegq_f64(float64x2_t __p0) {
   54983   float64x2_t __ret;
   54984   __ret = -__p0;
   54985   return __ret;
   54986 }
   54987 #else
   54988 __ai float64x2_t vnegq_f64(float64x2_t __p0) {
   54989   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   54990   float64x2_t __ret;
   54991   __ret = -__rev0;
   54992   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   54993   return __ret;
   54994 }
   54995 #endif
   54996 
   54997 #ifdef __LITTLE_ENDIAN__
   54998 __ai int64x2_t vnegq_s64(int64x2_t __p0) {
   54999   int64x2_t __ret;
   55000   __ret = -__p0;
   55001   return __ret;
   55002 }
   55003 #else
   55004 __ai int64x2_t vnegq_s64(int64x2_t __p0) {
   55005   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55006   int64x2_t __ret;
   55007   __ret = -__rev0;
   55008   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55009   return __ret;
   55010 }
   55011 #endif
   55012 
   55013 #ifdef __LITTLE_ENDIAN__
   55014 __ai float64x1_t vneg_f64(float64x1_t __p0) {
   55015   float64x1_t __ret;
   55016   __ret = -__p0;
   55017   return __ret;
   55018 }
   55019 #else
   55020 __ai float64x1_t vneg_f64(float64x1_t __p0) {
   55021   float64x1_t __ret;
   55022   __ret = -__p0;
   55023   return __ret;
   55024 }
   55025 #endif
   55026 
   55027 #ifdef __LITTLE_ENDIAN__
   55028 __ai int64x1_t vneg_s64(int64x1_t __p0) {
   55029   int64x1_t __ret;
   55030   __ret = -__p0;
   55031   return __ret;
   55032 }
   55033 #else
   55034 __ai int64x1_t vneg_s64(int64x1_t __p0) {
   55035   int64x1_t __ret;
   55036   __ret = -__p0;
   55037   return __ret;
   55038 }
   55039 #endif
   55040 
   55041 #ifdef __LITTLE_ENDIAN__
   55042 __ai int64_t vnegd_s64(int64_t __p0) {
   55043   int64_t __ret;
   55044   __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
   55045   return __ret;
   55046 }
   55047 #else
   55048 __ai int64_t vnegd_s64(int64_t __p0) {
   55049   int64_t __ret;
   55050   __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
   55051   return __ret;
   55052 }
   55053 #endif
   55054 
   55055 #ifdef __LITTLE_ENDIAN__
   55056 __ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   55057   uint8x16_t __ret;
   55058   __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   55059   return __ret;
   55060 }
   55061 #else
   55062 __ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   55063   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55064   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55065   uint8x16_t __ret;
   55066   __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   55067   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55068   return __ret;
   55069 }
   55070 #endif
   55071 
   55072 #ifdef __LITTLE_ENDIAN__
   55073 __ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   55074   uint32x4_t __ret;
   55075   __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   55076   return __ret;
   55077 }
   55078 #else
   55079 __ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   55080   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55081   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55082   uint32x4_t __ret;
   55083   __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   55084   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55085   return __ret;
   55086 }
   55087 #endif
   55088 
   55089 #ifdef __LITTLE_ENDIAN__
   55090 __ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   55091   uint64x2_t __ret;
   55092   __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   55093   return __ret;
   55094 }
   55095 #else
   55096 __ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   55097   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55098   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   55099   uint64x2_t __ret;
   55100   __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   55101   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55102   return __ret;
   55103 }
   55104 #endif
   55105 
   55106 #ifdef __LITTLE_ENDIAN__
   55107 __ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   55108   uint16x8_t __ret;
   55109   __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   55110   return __ret;
   55111 }
   55112 #else
   55113 __ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   55114   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   55115   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   55116   uint16x8_t __ret;
   55117   __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   55118   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   55119   return __ret;
   55120 }
   55121 #endif
   55122 
   55123 #ifdef __LITTLE_ENDIAN__
   55124 __ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   55125   int8x16_t __ret;
   55126   __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   55127   return __ret;
   55128 }
   55129 #else
   55130 __ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   55131   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55132   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55133   int8x16_t __ret;
   55134   __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   55135   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55136   return __ret;
   55137 }
   55138 #endif
   55139 
   55140 #ifdef __LITTLE_ENDIAN__
   55141 __ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
   55142   float64x2_t __ret;
   55143   __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   55144   return __ret;
   55145 }
   55146 #else
   55147 __ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
   55148   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55149   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   55150   float64x2_t __ret;
   55151   __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   55152   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55153   return __ret;
   55154 }
   55155 #endif
   55156 
   55157 #ifdef __LITTLE_ENDIAN__
   55158 __ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
   55159   float32x4_t __ret;
   55160   __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   55161   return __ret;
   55162 }
   55163 #else
   55164 __ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
   55165   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55166   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55167   float32x4_t __ret;
   55168   __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   55169   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55170   return __ret;
   55171 }
   55172 #endif
   55173 
   55174 #ifdef __LITTLE_ENDIAN__
   55175 __ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   55176   int32x4_t __ret;
   55177   __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   55178   return __ret;
   55179 }
   55180 #else
   55181 __ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   55182   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55183   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55184   int32x4_t __ret;
   55185   __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   55186   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55187   return __ret;
   55188 }
   55189 #endif
   55190 
   55191 #ifdef __LITTLE_ENDIAN__
   55192 __ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
   55193   int64x2_t __ret;
   55194   __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
   55195   return __ret;
   55196 }
   55197 #else
   55198 __ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
   55199   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55200   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   55201   int64x2_t __ret;
   55202   __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
   55203   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55204   return __ret;
   55205 }
   55206 #endif
   55207 
   55208 #ifdef __LITTLE_ENDIAN__
   55209 __ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   55210   int16x8_t __ret;
   55211   __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   55212   return __ret;
   55213 }
   55214 #else
   55215 __ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   55216   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   55217   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   55218   int16x8_t __ret;
   55219   __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   55220   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   55221   return __ret;
   55222 }
   55223 #endif
   55224 
   55225 #ifdef __LITTLE_ENDIAN__
   55226 __ai uint64_t vpaddd_u64(uint64x2_t __p0) {
   55227   uint64_t __ret;
   55228   __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__p0);
   55229   return __ret;
   55230 }
   55231 #else
   55232 __ai uint64_t vpaddd_u64(uint64x2_t __p0) {
   55233   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55234   uint64_t __ret;
   55235   __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__rev0);
   55236   return __ret;
   55237 }
   55238 #endif
   55239 
   55240 #ifdef __LITTLE_ENDIAN__
   55241 __ai float64_t vpaddd_f64(float64x2_t __p0) {
   55242   float64_t __ret;
   55243   __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__p0);
   55244   return __ret;
   55245 }
   55246 #else
   55247 __ai float64_t vpaddd_f64(float64x2_t __p0) {
   55248   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55249   float64_t __ret;
   55250   __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__rev0);
   55251   return __ret;
   55252 }
   55253 #endif
   55254 
   55255 #ifdef __LITTLE_ENDIAN__
   55256 __ai int64_t vpaddd_s64(int64x2_t __p0) {
   55257   int64_t __ret;
   55258   __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__p0);
   55259   return __ret;
   55260 }
   55261 #else
   55262 __ai int64_t vpaddd_s64(int64x2_t __p0) {
   55263   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55264   int64_t __ret;
   55265   __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__rev0);
   55266   return __ret;
   55267 }
   55268 #endif
   55269 
   55270 #ifdef __LITTLE_ENDIAN__
   55271 __ai float32_t vpadds_f32(float32x2_t __p0) {
   55272   float32_t __ret;
   55273   __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__p0);
   55274   return __ret;
   55275 }
   55276 #else
   55277 __ai float32_t vpadds_f32(float32x2_t __p0) {
   55278   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55279   float32_t __ret;
   55280   __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__rev0);
   55281   return __ret;
   55282 }
   55283 #endif
   55284 
   55285 #ifdef __LITTLE_ENDIAN__
   55286 __ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   55287   uint8x16_t __ret;
   55288   __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   55289   return __ret;
   55290 }
   55291 #else
   55292 __ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   55293   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55294   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55295   uint8x16_t __ret;
   55296   __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   55297   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55298   return __ret;
   55299 }
   55300 #endif
   55301 
   55302 #ifdef __LITTLE_ENDIAN__
   55303 __ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   55304   uint32x4_t __ret;
   55305   __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   55306   return __ret;
   55307 }
   55308 #else
   55309 __ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   55310   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55311   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55312   uint32x4_t __ret;
   55313   __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   55314   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55315   return __ret;
   55316 }
   55317 #endif
   55318 
   55319 #ifdef __LITTLE_ENDIAN__
   55320 __ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   55321   uint16x8_t __ret;
   55322   __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   55323   return __ret;
   55324 }
   55325 #else
   55326 __ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   55327   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   55328   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   55329   uint16x8_t __ret;
   55330   __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   55331   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   55332   return __ret;
   55333 }
   55334 #endif
   55335 
   55336 #ifdef __LITTLE_ENDIAN__
   55337 __ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
   55338   int8x16_t __ret;
   55339   __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   55340   return __ret;
   55341 }
   55342 #else
   55343 __ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
   55344   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55345   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55346   int8x16_t __ret;
   55347   __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   55348   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55349   return __ret;
   55350 }
   55351 #endif
   55352 
   55353 #ifdef __LITTLE_ENDIAN__
   55354 __ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
   55355   float64x2_t __ret;
   55356   __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   55357   return __ret;
   55358 }
   55359 #else
   55360 __ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
   55361   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55362   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   55363   float64x2_t __ret;
   55364   __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   55365   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55366   return __ret;
   55367 }
   55368 #endif
   55369 
   55370 #ifdef __LITTLE_ENDIAN__
   55371 __ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
   55372   float32x4_t __ret;
   55373   __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   55374   return __ret;
   55375 }
   55376 #else
   55377 __ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
   55378   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55379   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55380   float32x4_t __ret;
   55381   __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   55382   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55383   return __ret;
   55384 }
   55385 #endif
   55386 
   55387 #ifdef __LITTLE_ENDIAN__
   55388 __ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
   55389   int32x4_t __ret;
   55390   __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   55391   return __ret;
   55392 }
   55393 #else
   55394 __ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
   55395   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55396   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55397   int32x4_t __ret;
   55398   __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   55399   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55400   return __ret;
   55401 }
   55402 #endif
   55403 
   55404 #ifdef __LITTLE_ENDIAN__
   55405 __ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
   55406   int16x8_t __ret;
   55407   __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   55408   return __ret;
   55409 }
   55410 #else
   55411 __ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
   55412   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   55413   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   55414   int16x8_t __ret;
   55415   __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   55416   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   55417   return __ret;
   55418 }
   55419 #endif
   55420 
   55421 #ifdef __LITTLE_ENDIAN__
   55422 __ai float64_t vpmaxqd_f64(float64x2_t __p0) {
   55423   float64_t __ret;
   55424   __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__p0);
   55425   return __ret;
   55426 }
   55427 #else
   55428 __ai float64_t vpmaxqd_f64(float64x2_t __p0) {
   55429   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55430   float64_t __ret;
   55431   __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__rev0);
   55432   return __ret;
   55433 }
   55434 #endif
   55435 
   55436 #ifdef __LITTLE_ENDIAN__
   55437 __ai float32_t vpmaxs_f32(float32x2_t __p0) {
   55438   float32_t __ret;
   55439   __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__p0);
   55440   return __ret;
   55441 }
   55442 #else
   55443 __ai float32_t vpmaxs_f32(float32x2_t __p0) {
   55444   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55445   float32_t __ret;
   55446   __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__rev0);
   55447   return __ret;
   55448 }
   55449 #endif
   55450 
   55451 #ifdef __LITTLE_ENDIAN__
   55452 __ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
   55453   float64x2_t __ret;
   55454   __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   55455   return __ret;
   55456 }
   55457 #else
   55458 __ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
   55459   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55460   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   55461   float64x2_t __ret;
   55462   __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   55463   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55464   return __ret;
   55465 }
   55466 #endif
   55467 
   55468 #ifdef __LITTLE_ENDIAN__
   55469 __ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
   55470   float32x4_t __ret;
   55471   __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   55472   return __ret;
   55473 }
   55474 #else
   55475 __ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
   55476   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55477   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55478   float32x4_t __ret;
   55479   __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   55480   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55481   return __ret;
   55482 }
   55483 #endif
   55484 
   55485 #ifdef __LITTLE_ENDIAN__
   55486 __ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
   55487   float32x2_t __ret;
   55488   __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   55489   return __ret;
   55490 }
   55491 #else
   55492 __ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
   55493   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55494   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   55495   float32x2_t __ret;
   55496   __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   55497   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55498   return __ret;
   55499 }
   55500 #endif
   55501 
   55502 #ifdef __LITTLE_ENDIAN__
   55503 __ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
   55504   float64_t __ret;
   55505   __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__p0);
   55506   return __ret;
   55507 }
   55508 #else
   55509 __ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
   55510   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55511   float64_t __ret;
   55512   __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__rev0);
   55513   return __ret;
   55514 }
   55515 #endif
   55516 
   55517 #ifdef __LITTLE_ENDIAN__
   55518 __ai float32_t vpmaxnms_f32(float32x2_t __p0) {
   55519   float32_t __ret;
   55520   __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__p0);
   55521   return __ret;
   55522 }
   55523 #else
   55524 __ai float32_t vpmaxnms_f32(float32x2_t __p0) {
   55525   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55526   float32_t __ret;
   55527   __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__rev0);
   55528   return __ret;
   55529 }
   55530 #endif
   55531 
   55532 #ifdef __LITTLE_ENDIAN__
   55533 __ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   55534   uint8x16_t __ret;
   55535   __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   55536   return __ret;
   55537 }
   55538 #else
   55539 __ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   55540   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55541   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55542   uint8x16_t __ret;
   55543   __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   55544   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55545   return __ret;
   55546 }
   55547 #endif
   55548 
   55549 #ifdef __LITTLE_ENDIAN__
   55550 __ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   55551   uint32x4_t __ret;
   55552   __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   55553   return __ret;
   55554 }
   55555 #else
   55556 __ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   55557   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55558   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55559   uint32x4_t __ret;
   55560   __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   55561   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55562   return __ret;
   55563 }
   55564 #endif
   55565 
   55566 #ifdef __LITTLE_ENDIAN__
   55567 __ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   55568   uint16x8_t __ret;
   55569   __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   55570   return __ret;
   55571 }
   55572 #else
   55573 __ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   55574   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   55575   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   55576   uint16x8_t __ret;
   55577   __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   55578   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   55579   return __ret;
   55580 }
   55581 #endif
   55582 
   55583 #ifdef __LITTLE_ENDIAN__
   55584 __ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
   55585   int8x16_t __ret;
   55586   __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   55587   return __ret;
   55588 }
   55589 #else
   55590 __ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
   55591   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55592   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55593   int8x16_t __ret;
   55594   __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   55595   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   55596   return __ret;
   55597 }
   55598 #endif
   55599 
   55600 #ifdef __LITTLE_ENDIAN__
   55601 __ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
   55602   float64x2_t __ret;
   55603   __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   55604   return __ret;
   55605 }
   55606 #else
   55607 __ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
   55608   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55609   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   55610   float64x2_t __ret;
   55611   __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   55612   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55613   return __ret;
   55614 }
   55615 #endif
   55616 
   55617 #ifdef __LITTLE_ENDIAN__
   55618 __ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
   55619   float32x4_t __ret;
   55620   __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   55621   return __ret;
   55622 }
   55623 #else
   55624 __ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
   55625   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55626   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55627   float32x4_t __ret;
   55628   __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   55629   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55630   return __ret;
   55631 }
   55632 #endif
   55633 
   55634 #ifdef __LITTLE_ENDIAN__
   55635 __ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
   55636   int32x4_t __ret;
   55637   __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   55638   return __ret;
   55639 }
   55640 #else
   55641 __ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
   55642   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55643   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55644   int32x4_t __ret;
   55645   __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   55646   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55647   return __ret;
   55648 }
   55649 #endif
   55650 
   55651 #ifdef __LITTLE_ENDIAN__
   55652 __ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
   55653   int16x8_t __ret;
   55654   __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   55655   return __ret;
   55656 }
   55657 #else
   55658 __ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
   55659   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   55660   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   55661   int16x8_t __ret;
   55662   __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   55663   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   55664   return __ret;
   55665 }
   55666 #endif
   55667 
   55668 #ifdef __LITTLE_ENDIAN__
   55669 __ai float64_t vpminqd_f64(float64x2_t __p0) {
   55670   float64_t __ret;
   55671   __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__p0);
   55672   return __ret;
   55673 }
   55674 #else
   55675 __ai float64_t vpminqd_f64(float64x2_t __p0) {
   55676   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55677   float64_t __ret;
   55678   __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__rev0);
   55679   return __ret;
   55680 }
   55681 #endif
   55682 
   55683 #ifdef __LITTLE_ENDIAN__
   55684 __ai float32_t vpmins_f32(float32x2_t __p0) {
   55685   float32_t __ret;
   55686   __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__p0);
   55687   return __ret;
   55688 }
   55689 #else
   55690 __ai float32_t vpmins_f32(float32x2_t __p0) {
   55691   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55692   float32_t __ret;
   55693   __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__rev0);
   55694   return __ret;
   55695 }
   55696 #endif
   55697 
   55698 #ifdef __LITTLE_ENDIAN__
   55699 __ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
   55700   float64x2_t __ret;
   55701   __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   55702   return __ret;
   55703 }
   55704 #else
   55705 __ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
   55706   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55707   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   55708   float64x2_t __ret;
   55709   __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   55710   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55711   return __ret;
   55712 }
   55713 #endif
   55714 
   55715 #ifdef __LITTLE_ENDIAN__
   55716 __ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
   55717   float32x4_t __ret;
   55718   __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
   55719   return __ret;
   55720 }
   55721 #else
   55722 __ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
   55723   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   55724   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   55725   float32x4_t __ret;
   55726   __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
   55727   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   55728   return __ret;
   55729 }
   55730 #endif
   55731 
   55732 #ifdef __LITTLE_ENDIAN__
   55733 __ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
   55734   float32x2_t __ret;
   55735   __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
   55736   return __ret;
   55737 }
   55738 #else
   55739 __ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
   55740   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55741   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   55742   float32x2_t __ret;
   55743   __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
   55744   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55745   return __ret;
   55746 }
   55747 #endif
   55748 
   55749 #ifdef __LITTLE_ENDIAN__
   55750 __ai float64_t vpminnmqd_f64(float64x2_t __p0) {
   55751   float64_t __ret;
   55752   __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__p0);
   55753   return __ret;
   55754 }
   55755 #else
   55756 __ai float64_t vpminnmqd_f64(float64x2_t __p0) {
   55757   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55758   float64_t __ret;
   55759   __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__rev0);
   55760   return __ret;
   55761 }
   55762 #endif
   55763 
   55764 #ifdef __LITTLE_ENDIAN__
   55765 __ai float32_t vpminnms_f32(float32x2_t __p0) {
   55766   float32_t __ret;
   55767   __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__p0);
   55768   return __ret;
   55769 }
   55770 #else
   55771 __ai float32_t vpminnms_f32(float32x2_t __p0) {
   55772   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55773   float32_t __ret;
   55774   __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__rev0);
   55775   return __ret;
   55776 }
   55777 #endif
   55778 
   55779 #ifdef __LITTLE_ENDIAN__
   55780 __ai int64x2_t vqabsq_s64(int64x2_t __p0) {
   55781   int64x2_t __ret;
   55782   __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 35);
   55783   return __ret;
   55784 }
   55785 #else
   55786 __ai int64x2_t vqabsq_s64(int64x2_t __p0) {
   55787   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   55788   int64x2_t __ret;
   55789   __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 35);
   55790   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   55791   return __ret;
   55792 }
   55793 #endif
   55794 
   55795 #ifdef __LITTLE_ENDIAN__
   55796 __ai int64x1_t vqabs_s64(int64x1_t __p0) {
   55797   int64x1_t __ret;
   55798   __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
   55799   return __ret;
   55800 }
   55801 #else
   55802 __ai int64x1_t vqabs_s64(int64x1_t __p0) {
   55803   int64x1_t __ret;
   55804   __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
   55805   return __ret;
   55806 }
   55807 #endif
   55808 
   55809 #ifdef __LITTLE_ENDIAN__
   55810 __ai int8_t vqabsb_s8(int8_t __p0) {
   55811   int8_t __ret;
   55812   __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
   55813   return __ret;
   55814 }
   55815 #else
   55816 __ai int8_t vqabsb_s8(int8_t __p0) {
   55817   int8_t __ret;
   55818   __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
   55819   return __ret;
   55820 }
   55821 #endif
   55822 
   55823 #ifdef __LITTLE_ENDIAN__
   55824 __ai int32_t vqabss_s32(int32_t __p0) {
   55825   int32_t __ret;
   55826   __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
   55827   return __ret;
   55828 }
   55829 #else
   55830 __ai int32_t vqabss_s32(int32_t __p0) {
   55831   int32_t __ret;
   55832   __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
   55833   return __ret;
   55834 }
   55835 #endif
   55836 
   55837 #ifdef __LITTLE_ENDIAN__
   55838 __ai int64_t vqabsd_s64(int64_t __p0) {
   55839   int64_t __ret;
   55840   __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
   55841   return __ret;
   55842 }
   55843 #else
   55844 __ai int64_t vqabsd_s64(int64_t __p0) {
   55845   int64_t __ret;
   55846   __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
   55847   return __ret;
   55848 }
   55849 #endif
   55850 
   55851 #ifdef __LITTLE_ENDIAN__
   55852 __ai int16_t vqabsh_s16(int16_t __p0) {
   55853   int16_t __ret;
   55854   __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
   55855   return __ret;
   55856 }
   55857 #else
   55858 __ai int16_t vqabsh_s16(int16_t __p0) {
   55859   int16_t __ret;
   55860   __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
   55861   return __ret;
   55862 }
   55863 #endif
   55864 
   55865 #ifdef __LITTLE_ENDIAN__
   55866 __ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
   55867   uint8_t __ret;
   55868   __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
   55869   return __ret;
   55870 }
   55871 #else
   55872 __ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
   55873   uint8_t __ret;
   55874   __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
   55875   return __ret;
   55876 }
   55877 #endif
   55878 
   55879 #ifdef __LITTLE_ENDIAN__
   55880 __ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
   55881   uint32_t __ret;
   55882   __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
   55883   return __ret;
   55884 }
   55885 #else
   55886 __ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
   55887   uint32_t __ret;
   55888   __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
   55889   return __ret;
   55890 }
   55891 #endif
   55892 
   55893 #ifdef __LITTLE_ENDIAN__
   55894 __ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
   55895   uint64_t __ret;
   55896   __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
   55897   return __ret;
   55898 }
   55899 #else
   55900 __ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
   55901   uint64_t __ret;
   55902   __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
   55903   return __ret;
   55904 }
   55905 #endif
   55906 
   55907 #ifdef __LITTLE_ENDIAN__
   55908 __ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
   55909   uint16_t __ret;
   55910   __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
   55911   return __ret;
   55912 }
   55913 #else
   55914 __ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
   55915   uint16_t __ret;
   55916   __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
   55917   return __ret;
   55918 }
   55919 #endif
   55920 
   55921 #ifdef __LITTLE_ENDIAN__
   55922 __ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
   55923   int8_t __ret;
   55924   __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
   55925   return __ret;
   55926 }
   55927 #else
   55928 __ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
   55929   int8_t __ret;
   55930   __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
   55931   return __ret;
   55932 }
   55933 #endif
   55934 
   55935 #ifdef __LITTLE_ENDIAN__
   55936 __ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
   55937   int32_t __ret;
   55938   __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
   55939   return __ret;
   55940 }
   55941 #else
   55942 __ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
   55943   int32_t __ret;
   55944   __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
   55945   return __ret;
   55946 }
   55947 __ai int32_t __noswap_vqadds_s32(int32_t __p0, int32_t __p1) {
   55948   int32_t __ret;
   55949   __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
   55950   return __ret;
   55951 }
   55952 #endif
   55953 
   55954 #ifdef __LITTLE_ENDIAN__
   55955 __ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
   55956   int64_t __ret;
   55957   __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
   55958   return __ret;
   55959 }
   55960 #else
   55961 __ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
   55962   int64_t __ret;
   55963   __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
   55964   return __ret;
   55965 }
   55966 #endif
   55967 
   55968 #ifdef __LITTLE_ENDIAN__
   55969 __ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
   55970   int16_t __ret;
   55971   __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
   55972   return __ret;
   55973 }
   55974 #else
   55975 __ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
   55976   int16_t __ret;
   55977   __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
   55978   return __ret;
   55979 }
   55980 __ai int16_t __noswap_vqaddh_s16(int16_t __p0, int16_t __p1) {
   55981   int16_t __ret;
   55982   __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
   55983   return __ret;
   55984 }
   55985 #endif
   55986 
   55987 #ifdef __LITTLE_ENDIAN__
   55988 __ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
   55989   int64_t __ret;
   55990   __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
   55991   return __ret;
   55992 }
   55993 #else
   55994 __ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
   55995   int64_t __ret;
   55996   __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
   55997   return __ret;
   55998 }
   55999 #endif
   56000 
   56001 #ifdef __LITTLE_ENDIAN__
   56002 __ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
   56003   int32_t __ret;
   56004   __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
   56005   return __ret;
   56006 }
   56007 #else
   56008 __ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
   56009   int32_t __ret;
   56010   __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
   56011   return __ret;
   56012 }
   56013 #endif
   56014 
   56015 #ifdef __LITTLE_ENDIAN__
   56016 __ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   56017   int64x2_t __ret;
   56018   __ret = vqdmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
   56019   return __ret;
   56020 }
   56021 #else
   56022 __ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   56023   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   56024   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   56025   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   56026   int64x2_t __ret;
   56027   __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
   56028   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   56029   return __ret;
   56030 }
   56031 #endif
   56032 
   56033 #ifdef __LITTLE_ENDIAN__
   56034 __ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   56035   int32x4_t __ret;
   56036   __ret = vqdmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
   56037   return __ret;
   56038 }
   56039 #else
   56040 __ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   56041   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   56042   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   56043   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   56044   int32x4_t __ret;
   56045   __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
   56046   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   56047   return __ret;
   56048 }
   56049 #endif
   56050 
   56051 #ifdef __LITTLE_ENDIAN__
   56052 #define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56053   int64x2_t __s0 = __p0; \
   56054   int32x4_t __s1 = __p1; \
   56055   int32x2_t __s2 = __p2; \
   56056   int64x2_t __ret; \
   56057   __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   56058   __ret; \
   56059 })
   56060 #else
   56061 #define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56062   int64x2_t __s0 = __p0; \
   56063   int32x4_t __s1 = __p1; \
   56064   int32x2_t __s2 = __p2; \
   56065   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   56066   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56067   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   56068   int64x2_t __ret; \
   56069   __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   56070   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   56071   __ret; \
   56072 })
   56073 #endif
   56074 
   56075 #ifdef __LITTLE_ENDIAN__
   56076 #define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56077   int32x4_t __s0 = __p0; \
   56078   int16x8_t __s1 = __p1; \
   56079   int16x4_t __s2 = __p2; \
   56080   int32x4_t __ret; \
   56081   __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   56082   __ret; \
   56083 })
   56084 #else
   56085 #define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56086   int32x4_t __s0 = __p0; \
   56087   int16x8_t __s1 = __p1; \
   56088   int16x4_t __s2 = __p2; \
   56089   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56090   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   56091   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56092   int32x4_t __ret; \
   56093   __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   56094   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56095   __ret; \
   56096 })
   56097 #endif
   56098 
   56099 #ifdef __LITTLE_ENDIAN__
   56100 #define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56101   int64x2_t __s0 = __p0; \
   56102   int32x4_t __s1 = __p1; \
   56103   int32x4_t __s2 = __p2; \
   56104   int64x2_t __ret; \
   56105   __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   56106   __ret; \
   56107 })
   56108 #else
   56109 #define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56110   int64x2_t __s0 = __p0; \
   56111   int32x4_t __s1 = __p1; \
   56112   int32x4_t __s2 = __p2; \
   56113   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   56114   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56115   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56116   int64x2_t __ret; \
   56117   __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   56118   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   56119   __ret; \
   56120 })
   56121 #endif
   56122 
   56123 #ifdef __LITTLE_ENDIAN__
   56124 #define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56125   int32x4_t __s0 = __p0; \
   56126   int16x8_t __s1 = __p1; \
   56127   int16x8_t __s2 = __p2; \
   56128   int32x4_t __ret; \
   56129   __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   56130   __ret; \
   56131 })
   56132 #else
   56133 #define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56134   int32x4_t __s0 = __p0; \
   56135   int16x8_t __s1 = __p1; \
   56136   int16x8_t __s2 = __p2; \
   56137   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56138   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   56139   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   56140   int32x4_t __ret; \
   56141   __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   56142   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56143   __ret; \
   56144 })
   56145 #endif
   56146 
   56147 #ifdef __LITTLE_ENDIAN__
   56148 __ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
   56149   int64x2_t __ret;
   56150   __ret = vqdmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
   56151   return __ret;
   56152 }
   56153 #else
   56154 __ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
   56155   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   56156   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   56157   int64x2_t __ret;
   56158   __ret = __noswap_vqdmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
   56159   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   56160   return __ret;
   56161 }
   56162 #endif
   56163 
   56164 #ifdef __LITTLE_ENDIAN__
   56165 __ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
   56166   int32x4_t __ret;
   56167   __ret = vqdmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
   56168   return __ret;
   56169 }
   56170 #else
   56171 __ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
   56172   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   56173   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   56174   int32x4_t __ret;
   56175   __ret = __noswap_vqdmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
   56176   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   56177   return __ret;
   56178 }
   56179 #endif
   56180 
   56181 #ifdef __LITTLE_ENDIAN__
   56182 #define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56183   int64_t __s0 = __p0; \
   56184   int32_t __s1 = __p1; \
   56185   int32x2_t __s2 = __p2; \
   56186   int64_t __ret; \
   56187   __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
   56188   __ret; \
   56189 })
   56190 #else
   56191 #define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56192   int64_t __s0 = __p0; \
   56193   int32_t __s1 = __p1; \
   56194   int32x2_t __s2 = __p2; \
   56195   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   56196   int64_t __ret; \
   56197   __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
   56198   __ret; \
   56199 })
   56200 #endif
   56201 
   56202 #ifdef __LITTLE_ENDIAN__
   56203 #define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56204   int32_t __s0 = __p0; \
   56205   int16_t __s1 = __p1; \
   56206   int16x4_t __s2 = __p2; \
   56207   int32_t __ret; \
   56208   __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
   56209   __ret; \
   56210 })
   56211 #else
   56212 #define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56213   int32_t __s0 = __p0; \
   56214   int16_t __s1 = __p1; \
   56215   int16x4_t __s2 = __p2; \
   56216   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56217   int32_t __ret; \
   56218   __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
   56219   __ret; \
   56220 })
   56221 #endif
   56222 
   56223 #ifdef __LITTLE_ENDIAN__
   56224 #define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56225   int64_t __s0 = __p0; \
   56226   int32_t __s1 = __p1; \
   56227   int32x4_t __s2 = __p2; \
   56228   int64_t __ret; \
   56229   __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
   56230   __ret; \
   56231 })
   56232 #else
   56233 #define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56234   int64_t __s0 = __p0; \
   56235   int32_t __s1 = __p1; \
   56236   int32x4_t __s2 = __p2; \
   56237   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56238   int64_t __ret; \
   56239   __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
   56240   __ret; \
   56241 })
   56242 #endif
   56243 
   56244 #ifdef __LITTLE_ENDIAN__
   56245 #define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56246   int32_t __s0 = __p0; \
   56247   int16_t __s1 = __p1; \
   56248   int16x8_t __s2 = __p2; \
   56249   int32_t __ret; \
   56250   __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
   56251   __ret; \
   56252 })
   56253 #else
   56254 #define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56255   int32_t __s0 = __p0; \
   56256   int16_t __s1 = __p1; \
   56257   int16x8_t __s2 = __p2; \
   56258   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   56259   int32_t __ret; \
   56260   __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
   56261   __ret; \
   56262 })
   56263 #endif
   56264 
   56265 #ifdef __LITTLE_ENDIAN__
   56266 #define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56267   int64x2_t __s0 = __p0; \
   56268   int32x2_t __s1 = __p1; \
   56269   int32x4_t __s2 = __p2; \
   56270   int64x2_t __ret; \
   56271   __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   56272   __ret; \
   56273 })
   56274 #else
   56275 #define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56276   int64x2_t __s0 = __p0; \
   56277   int32x2_t __s1 = __p1; \
   56278   int32x4_t __s2 = __p2; \
   56279   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   56280   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   56281   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56282   int64x2_t __ret; \
   56283   __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   56284   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   56285   __ret; \
   56286 })
   56287 #endif
   56288 
   56289 #ifdef __LITTLE_ENDIAN__
   56290 #define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56291   int32x4_t __s0 = __p0; \
   56292   int16x4_t __s1 = __p1; \
   56293   int16x8_t __s2 = __p2; \
   56294   int32x4_t __ret; \
   56295   __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   56296   __ret; \
   56297 })
   56298 #else
   56299 #define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56300   int32x4_t __s0 = __p0; \
   56301   int16x4_t __s1 = __p1; \
   56302   int16x8_t __s2 = __p2; \
   56303   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56304   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56305   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   56306   int32x4_t __ret; \
   56307   __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   56308   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56309   __ret; \
   56310 })
   56311 #endif
   56312 
   56313 #ifdef __LITTLE_ENDIAN__
   56314 __ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
   56315   int64_t __ret;
   56316   __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
   56317   return __ret;
   56318 }
   56319 #else
   56320 __ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
   56321   int64_t __ret;
   56322   __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
   56323   return __ret;
   56324 }
   56325 #endif
   56326 
   56327 #ifdef __LITTLE_ENDIAN__
   56328 __ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
   56329   int32_t __ret;
   56330   __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
   56331   return __ret;
   56332 }
   56333 #else
   56334 __ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
   56335   int32_t __ret;
   56336   __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
   56337   return __ret;
   56338 }
   56339 #endif
   56340 
   56341 #ifdef __LITTLE_ENDIAN__
   56342 __ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   56343   int64x2_t __ret;
   56344   __ret = vqdmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
   56345   return __ret;
   56346 }
   56347 #else
   56348 __ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   56349   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   56350   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   56351   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   56352   int64x2_t __ret;
   56353   __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
   56354   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   56355   return __ret;
   56356 }
   56357 #endif
   56358 
   56359 #ifdef __LITTLE_ENDIAN__
   56360 __ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   56361   int32x4_t __ret;
   56362   __ret = vqdmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
   56363   return __ret;
   56364 }
   56365 #else
   56366 __ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   56367   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   56368   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   56369   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   56370   int32x4_t __ret;
   56371   __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
   56372   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   56373   return __ret;
   56374 }
   56375 #endif
   56376 
   56377 #ifdef __LITTLE_ENDIAN__
   56378 #define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56379   int64x2_t __s0 = __p0; \
   56380   int32x4_t __s1 = __p1; \
   56381   int32x2_t __s2 = __p2; \
   56382   int64x2_t __ret; \
   56383   __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   56384   __ret; \
   56385 })
   56386 #else
   56387 #define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56388   int64x2_t __s0 = __p0; \
   56389   int32x4_t __s1 = __p1; \
   56390   int32x2_t __s2 = __p2; \
   56391   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   56392   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56393   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   56394   int64x2_t __ret; \
   56395   __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   56396   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   56397   __ret; \
   56398 })
   56399 #endif
   56400 
   56401 #ifdef __LITTLE_ENDIAN__
   56402 #define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56403   int32x4_t __s0 = __p0; \
   56404   int16x8_t __s1 = __p1; \
   56405   int16x4_t __s2 = __p2; \
   56406   int32x4_t __ret; \
   56407   __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   56408   __ret; \
   56409 })
   56410 #else
   56411 #define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56412   int32x4_t __s0 = __p0; \
   56413   int16x8_t __s1 = __p1; \
   56414   int16x4_t __s2 = __p2; \
   56415   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56416   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   56417   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56418   int32x4_t __ret; \
   56419   __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   56420   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56421   __ret; \
   56422 })
   56423 #endif
   56424 
   56425 #ifdef __LITTLE_ENDIAN__
   56426 #define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56427   int64x2_t __s0 = __p0; \
   56428   int32x4_t __s1 = __p1; \
   56429   int32x4_t __s2 = __p2; \
   56430   int64x2_t __ret; \
   56431   __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   56432   __ret; \
   56433 })
   56434 #else
   56435 #define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56436   int64x2_t __s0 = __p0; \
   56437   int32x4_t __s1 = __p1; \
   56438   int32x4_t __s2 = __p2; \
   56439   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   56440   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56441   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56442   int64x2_t __ret; \
   56443   __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   56444   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   56445   __ret; \
   56446 })
   56447 #endif
   56448 
   56449 #ifdef __LITTLE_ENDIAN__
   56450 #define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56451   int32x4_t __s0 = __p0; \
   56452   int16x8_t __s1 = __p1; \
   56453   int16x8_t __s2 = __p2; \
   56454   int32x4_t __ret; \
   56455   __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   56456   __ret; \
   56457 })
   56458 #else
   56459 #define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56460   int32x4_t __s0 = __p0; \
   56461   int16x8_t __s1 = __p1; \
   56462   int16x8_t __s2 = __p2; \
   56463   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56464   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   56465   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   56466   int32x4_t __ret; \
   56467   __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   56468   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56469   __ret; \
   56470 })
   56471 #endif
   56472 
   56473 #ifdef __LITTLE_ENDIAN__
   56474 __ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
   56475   int64x2_t __ret;
   56476   __ret = vqdmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
   56477   return __ret;
   56478 }
   56479 #else
   56480 __ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
   56481   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   56482   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   56483   int64x2_t __ret;
   56484   __ret = __noswap_vqdmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
   56485   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   56486   return __ret;
   56487 }
   56488 #endif
   56489 
   56490 #ifdef __LITTLE_ENDIAN__
   56491 __ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
   56492   int32x4_t __ret;
   56493   __ret = vqdmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
   56494   return __ret;
   56495 }
   56496 #else
   56497 __ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
   56498   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   56499   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   56500   int32x4_t __ret;
   56501   __ret = __noswap_vqdmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
   56502   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   56503   return __ret;
   56504 }
   56505 #endif
   56506 
   56507 #ifdef __LITTLE_ENDIAN__
   56508 #define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56509   int64_t __s0 = __p0; \
   56510   int32_t __s1 = __p1; \
   56511   int32x2_t __s2 = __p2; \
   56512   int64_t __ret; \
   56513   __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
   56514   __ret; \
   56515 })
   56516 #else
   56517 #define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56518   int64_t __s0 = __p0; \
   56519   int32_t __s1 = __p1; \
   56520   int32x2_t __s2 = __p2; \
   56521   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   56522   int64_t __ret; \
   56523   __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
   56524   __ret; \
   56525 })
   56526 #endif
   56527 
   56528 #ifdef __LITTLE_ENDIAN__
   56529 #define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56530   int32_t __s0 = __p0; \
   56531   int16_t __s1 = __p1; \
   56532   int16x4_t __s2 = __p2; \
   56533   int32_t __ret; \
   56534   __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
   56535   __ret; \
   56536 })
   56537 #else
   56538 #define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56539   int32_t __s0 = __p0; \
   56540   int16_t __s1 = __p1; \
   56541   int16x4_t __s2 = __p2; \
   56542   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56543   int32_t __ret; \
   56544   __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
   56545   __ret; \
   56546 })
   56547 #endif
   56548 
   56549 #ifdef __LITTLE_ENDIAN__
   56550 #define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56551   int64_t __s0 = __p0; \
   56552   int32_t __s1 = __p1; \
   56553   int32x4_t __s2 = __p2; \
   56554   int64_t __ret; \
   56555   __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
   56556   __ret; \
   56557 })
   56558 #else
   56559 #define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56560   int64_t __s0 = __p0; \
   56561   int32_t __s1 = __p1; \
   56562   int32x4_t __s2 = __p2; \
   56563   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56564   int64_t __ret; \
   56565   __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
   56566   __ret; \
   56567 })
   56568 #endif
   56569 
   56570 #ifdef __LITTLE_ENDIAN__
   56571 #define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56572   int32_t __s0 = __p0; \
   56573   int16_t __s1 = __p1; \
   56574   int16x8_t __s2 = __p2; \
   56575   int32_t __ret; \
   56576   __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
   56577   __ret; \
   56578 })
   56579 #else
   56580 #define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56581   int32_t __s0 = __p0; \
   56582   int16_t __s1 = __p1; \
   56583   int16x8_t __s2 = __p2; \
   56584   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   56585   int32_t __ret; \
   56586   __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
   56587   __ret; \
   56588 })
   56589 #endif
   56590 
   56591 #ifdef __LITTLE_ENDIAN__
   56592 #define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56593   int64x2_t __s0 = __p0; \
   56594   int32x2_t __s1 = __p1; \
   56595   int32x4_t __s2 = __p2; \
   56596   int64x2_t __ret; \
   56597   __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   56598   __ret; \
   56599 })
   56600 #else
   56601 #define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   56602   int64x2_t __s0 = __p0; \
   56603   int32x2_t __s1 = __p1; \
   56604   int32x4_t __s2 = __p2; \
   56605   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   56606   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   56607   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   56608   int64x2_t __ret; \
   56609   __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   56610   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   56611   __ret; \
   56612 })
   56613 #endif
   56614 
   56615 #ifdef __LITTLE_ENDIAN__
   56616 #define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56617   int32x4_t __s0 = __p0; \
   56618   int16x4_t __s1 = __p1; \
   56619   int16x8_t __s2 = __p2; \
   56620   int32x4_t __ret; \
   56621   __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   56622   __ret; \
   56623 })
   56624 #else
   56625 #define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   56626   int32x4_t __s0 = __p0; \
   56627   int16x4_t __s1 = __p1; \
   56628   int16x8_t __s2 = __p2; \
   56629   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56630   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56631   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
   56632   int32x4_t __ret; \
   56633   __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   56634   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56635   __ret; \
   56636 })
   56637 #endif
   56638 
   56639 #ifdef __LITTLE_ENDIAN__
   56640 __ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
   56641   int32_t __ret;
   56642   __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
   56643   return __ret;
   56644 }
   56645 #else
   56646 __ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
   56647   int32_t __ret;
   56648   __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
   56649   return __ret;
   56650 }
   56651 __ai int32_t __noswap_vqdmulhs_s32(int32_t __p0, int32_t __p1) {
   56652   int32_t __ret;
   56653   __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
   56654   return __ret;
   56655 }
   56656 #endif
   56657 
   56658 #ifdef __LITTLE_ENDIAN__
   56659 __ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
   56660   int16_t __ret;
   56661   __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
   56662   return __ret;
   56663 }
   56664 #else
   56665 __ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
   56666   int16_t __ret;
   56667   __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
   56668   return __ret;
   56669 }
   56670 __ai int16_t __noswap_vqdmulhh_s16(int16_t __p0, int16_t __p1) {
   56671   int16_t __ret;
   56672   __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
   56673   return __ret;
   56674 }
   56675 #endif
   56676 
   56677 #ifdef __LITTLE_ENDIAN__
   56678 #define vqdmulhs_lane_s32(__p0_146, __p1_146, __p2_146) __extension__ ({ \
   56679   int32_t __s0_146 = __p0_146; \
   56680   int32x2_t __s1_146 = __p1_146; \
   56681   int32_t __ret_146; \
   56682   __ret_146 = vqdmulhs_s32(__s0_146, vget_lane_s32(__s1_146, __p2_146)); \
   56683   __ret_146; \
   56684 })
   56685 #else
   56686 #define vqdmulhs_lane_s32(__p0_147, __p1_147, __p2_147) __extension__ ({ \
   56687   int32_t __s0_147 = __p0_147; \
   56688   int32x2_t __s1_147 = __p1_147; \
   56689   int32x2_t __rev1_147;  __rev1_147 = __builtin_shufflevector(__s1_147, __s1_147, 1, 0); \
   56690   int32_t __ret_147; \
   56691   __ret_147 = __noswap_vqdmulhs_s32(__s0_147, __noswap_vget_lane_s32(__rev1_147, __p2_147)); \
   56692   __ret_147; \
   56693 })
   56694 #endif
   56695 
   56696 #ifdef __LITTLE_ENDIAN__
   56697 #define vqdmulhh_lane_s16(__p0_148, __p1_148, __p2_148) __extension__ ({ \
   56698   int16_t __s0_148 = __p0_148; \
   56699   int16x4_t __s1_148 = __p1_148; \
   56700   int16_t __ret_148; \
   56701   __ret_148 = vqdmulhh_s16(__s0_148, vget_lane_s16(__s1_148, __p2_148)); \
   56702   __ret_148; \
   56703 })
   56704 #else
   56705 #define vqdmulhh_lane_s16(__p0_149, __p1_149, __p2_149) __extension__ ({ \
   56706   int16_t __s0_149 = __p0_149; \
   56707   int16x4_t __s1_149 = __p1_149; \
   56708   int16x4_t __rev1_149;  __rev1_149 = __builtin_shufflevector(__s1_149, __s1_149, 3, 2, 1, 0); \
   56709   int16_t __ret_149; \
   56710   __ret_149 = __noswap_vqdmulhh_s16(__s0_149, __noswap_vget_lane_s16(__rev1_149, __p2_149)); \
   56711   __ret_149; \
   56712 })
   56713 #endif
   56714 
   56715 #ifdef __LITTLE_ENDIAN__
   56716 #define vqdmulhs_laneq_s32(__p0_150, __p1_150, __p2_150) __extension__ ({ \
   56717   int32_t __s0_150 = __p0_150; \
   56718   int32x4_t __s1_150 = __p1_150; \
   56719   int32_t __ret_150; \
   56720   __ret_150 = vqdmulhs_s32(__s0_150, vgetq_lane_s32(__s1_150, __p2_150)); \
   56721   __ret_150; \
   56722 })
   56723 #else
   56724 #define vqdmulhs_laneq_s32(__p0_151, __p1_151, __p2_151) __extension__ ({ \
   56725   int32_t __s0_151 = __p0_151; \
   56726   int32x4_t __s1_151 = __p1_151; \
   56727   int32x4_t __rev1_151;  __rev1_151 = __builtin_shufflevector(__s1_151, __s1_151, 3, 2, 1, 0); \
   56728   int32_t __ret_151; \
   56729   __ret_151 = __noswap_vqdmulhs_s32(__s0_151, __noswap_vgetq_lane_s32(__rev1_151, __p2_151)); \
   56730   __ret_151; \
   56731 })
   56732 #endif
   56733 
   56734 #ifdef __LITTLE_ENDIAN__
   56735 #define vqdmulhh_laneq_s16(__p0_152, __p1_152, __p2_152) __extension__ ({ \
   56736   int16_t __s0_152 = __p0_152; \
   56737   int16x8_t __s1_152 = __p1_152; \
   56738   int16_t __ret_152; \
   56739   __ret_152 = vqdmulhh_s16(__s0_152, vgetq_lane_s16(__s1_152, __p2_152)); \
   56740   __ret_152; \
   56741 })
   56742 #else
   56743 #define vqdmulhh_laneq_s16(__p0_153, __p1_153, __p2_153) __extension__ ({ \
   56744   int16_t __s0_153 = __p0_153; \
   56745   int16x8_t __s1_153 = __p1_153; \
   56746   int16x8_t __rev1_153;  __rev1_153 = __builtin_shufflevector(__s1_153, __s1_153, 7, 6, 5, 4, 3, 2, 1, 0); \
   56747   int16_t __ret_153; \
   56748   __ret_153 = __noswap_vqdmulhh_s16(__s0_153, __noswap_vgetq_lane_s16(__rev1_153, __p2_153)); \
   56749   __ret_153; \
   56750 })
   56751 #endif
   56752 
   56753 #ifdef __LITTLE_ENDIAN__
   56754 #define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   56755   int32x4_t __s0 = __p0; \
   56756   int32x4_t __s1 = __p1; \
   56757   int32x4_t __ret; \
   56758   __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   56759   __ret; \
   56760 })
   56761 #else
   56762 #define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   56763   int32x4_t __s0 = __p0; \
   56764   int32x4_t __s1 = __p1; \
   56765   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56766   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56767   int32x4_t __ret; \
   56768   __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   56769   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56770   __ret; \
   56771 })
   56772 #endif
   56773 
   56774 #ifdef __LITTLE_ENDIAN__
   56775 #define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   56776   int16x8_t __s0 = __p0; \
   56777   int16x8_t __s1 = __p1; \
   56778   int16x8_t __ret; \
   56779   __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
   56780   __ret; \
   56781 })
   56782 #else
   56783 #define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   56784   int16x8_t __s0 = __p0; \
   56785   int16x8_t __s1 = __p1; \
   56786   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   56787   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   56788   int16x8_t __ret; \
   56789   __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
   56790   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   56791   __ret; \
   56792 })
   56793 #endif
   56794 
   56795 #ifdef __LITTLE_ENDIAN__
   56796 #define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   56797   int32x2_t __s0 = __p0; \
   56798   int32x4_t __s1 = __p1; \
   56799   int32x2_t __ret; \
   56800   __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   56801   __ret; \
   56802 })
   56803 #else
   56804 #define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   56805   int32x2_t __s0 = __p0; \
   56806   int32x4_t __s1 = __p1; \
   56807   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   56808   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56809   int32x2_t __ret; \
   56810   __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   56811   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   56812   __ret; \
   56813 })
   56814 #endif
   56815 
   56816 #ifdef __LITTLE_ENDIAN__
   56817 #define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   56818   int16x4_t __s0 = __p0; \
   56819   int16x8_t __s1 = __p1; \
   56820   int16x4_t __ret; \
   56821   __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   56822   __ret; \
   56823 })
   56824 #else
   56825 #define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   56826   int16x4_t __s0 = __p0; \
   56827   int16x8_t __s1 = __p1; \
   56828   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56829   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   56830   int16x4_t __ret; \
   56831   __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   56832   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56833   __ret; \
   56834 })
   56835 #endif
   56836 
   56837 #ifdef __LITTLE_ENDIAN__
   56838 __ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
   56839   int64_t __ret;
   56840   __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
   56841   return __ret;
   56842 }
   56843 #else
   56844 __ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
   56845   int64_t __ret;
   56846   __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
   56847   return __ret;
   56848 }
   56849 __ai int64_t __noswap_vqdmulls_s32(int32_t __p0, int32_t __p1) {
   56850   int64_t __ret;
   56851   __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
   56852   return __ret;
   56853 }
   56854 #endif
   56855 
   56856 #ifdef __LITTLE_ENDIAN__
   56857 __ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
   56858   int32_t __ret;
   56859   __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
   56860   return __ret;
   56861 }
   56862 #else
   56863 __ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
   56864   int32_t __ret;
   56865   __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
   56866   return __ret;
   56867 }
   56868 __ai int32_t __noswap_vqdmullh_s16(int16_t __p0, int16_t __p1) {
   56869   int32_t __ret;
   56870   __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
   56871   return __ret;
   56872 }
   56873 #endif
   56874 
   56875 #ifdef __LITTLE_ENDIAN__
   56876 __ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
   56877   int64x2_t __ret;
   56878   __ret = vqdmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
   56879   return __ret;
   56880 }
   56881 #else
   56882 __ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
   56883   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   56884   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   56885   int64x2_t __ret;
   56886   __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
   56887   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   56888   return __ret;
   56889 }
   56890 #endif
   56891 
   56892 #ifdef __LITTLE_ENDIAN__
   56893 __ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
   56894   int32x4_t __ret;
   56895   __ret = vqdmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
   56896   return __ret;
   56897 }
   56898 #else
   56899 __ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
   56900   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   56901   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   56902   int32x4_t __ret;
   56903   __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
   56904   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   56905   return __ret;
   56906 }
   56907 #endif
   56908 
   56909 #ifdef __LITTLE_ENDIAN__
   56910 #define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   56911   int32x4_t __s0 = __p0; \
   56912   int32x2_t __s1 = __p1; \
   56913   int64x2_t __ret; \
   56914   __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   56915   __ret; \
   56916 })
   56917 #else
   56918 #define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
   56919   int32x4_t __s0 = __p0; \
   56920   int32x2_t __s1 = __p1; \
   56921   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56922   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   56923   int64x2_t __ret; \
   56924   __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   56925   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   56926   __ret; \
   56927 })
   56928 #endif
   56929 
   56930 #ifdef __LITTLE_ENDIAN__
   56931 #define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   56932   int16x8_t __s0 = __p0; \
   56933   int16x4_t __s1 = __p1; \
   56934   int32x4_t __ret; \
   56935   __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   56936   __ret; \
   56937 })
   56938 #else
   56939 #define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
   56940   int16x8_t __s0 = __p0; \
   56941   int16x4_t __s1 = __p1; \
   56942   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   56943   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56944   int32x4_t __ret; \
   56945   __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   56946   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56947   __ret; \
   56948 })
   56949 #endif
   56950 
   56951 #ifdef __LITTLE_ENDIAN__
   56952 #define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   56953   int32x4_t __s0 = __p0; \
   56954   int32x4_t __s1 = __p1; \
   56955   int64x2_t __ret; \
   56956   __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   56957   __ret; \
   56958 })
   56959 #else
   56960 #define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   56961   int32x4_t __s0 = __p0; \
   56962   int32x4_t __s1 = __p1; \
   56963   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   56964   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   56965   int64x2_t __ret; \
   56966   __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   56967   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   56968   __ret; \
   56969 })
   56970 #endif
   56971 
   56972 #ifdef __LITTLE_ENDIAN__
   56973 #define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   56974   int16x8_t __s0 = __p0; \
   56975   int16x8_t __s1 = __p1; \
   56976   int32x4_t __ret; \
   56977   __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   56978   __ret; \
   56979 })
   56980 #else
   56981 #define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   56982   int16x8_t __s0 = __p0; \
   56983   int16x8_t __s1 = __p1; \
   56984   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   56985   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   56986   int32x4_t __ret; \
   56987   __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   56988   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   56989   __ret; \
   56990 })
   56991 #endif
   56992 
   56993 #ifdef __LITTLE_ENDIAN__
   56994 __ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
   56995   int64x2_t __ret;
   56996   __ret = vqdmull_n_s32(vget_high_s32(__p0), __p1);
   56997   return __ret;
   56998 }
   56999 #else
   57000 __ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
   57001   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   57002   int64x2_t __ret;
   57003   __ret = __noswap_vqdmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
   57004   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   57005   return __ret;
   57006 }
   57007 #endif
   57008 
   57009 #ifdef __LITTLE_ENDIAN__
   57010 __ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
   57011   int32x4_t __ret;
   57012   __ret = vqdmull_n_s16(vget_high_s16(__p0), __p1);
   57013   return __ret;
   57014 }
   57015 #else
   57016 __ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
   57017   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   57018   int32x4_t __ret;
   57019   __ret = __noswap_vqdmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
   57020   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   57021   return __ret;
   57022 }
   57023 #endif
   57024 
   57025 #ifdef __LITTLE_ENDIAN__
   57026 #define vqdmulls_lane_s32(__p0_154, __p1_154, __p2_154) __extension__ ({ \
   57027   int32_t __s0_154 = __p0_154; \
   57028   int32x2_t __s1_154 = __p1_154; \
   57029   int64_t __ret_154; \
   57030   __ret_154 = vqdmulls_s32(__s0_154, vget_lane_s32(__s1_154, __p2_154)); \
   57031   __ret_154; \
   57032 })
   57033 #else
   57034 #define vqdmulls_lane_s32(__p0_155, __p1_155, __p2_155) __extension__ ({ \
   57035   int32_t __s0_155 = __p0_155; \
   57036   int32x2_t __s1_155 = __p1_155; \
   57037   int32x2_t __rev1_155;  __rev1_155 = __builtin_shufflevector(__s1_155, __s1_155, 1, 0); \
   57038   int64_t __ret_155; \
   57039   __ret_155 = __noswap_vqdmulls_s32(__s0_155, __noswap_vget_lane_s32(__rev1_155, __p2_155)); \
   57040   __ret_155; \
   57041 })
   57042 #endif
   57043 
   57044 #ifdef __LITTLE_ENDIAN__
   57045 #define vqdmullh_lane_s16(__p0_156, __p1_156, __p2_156) __extension__ ({ \
   57046   int16_t __s0_156 = __p0_156; \
   57047   int16x4_t __s1_156 = __p1_156; \
   57048   int32_t __ret_156; \
   57049   __ret_156 = vqdmullh_s16(__s0_156, vget_lane_s16(__s1_156, __p2_156)); \
   57050   __ret_156; \
   57051 })
   57052 #else
   57053 #define vqdmullh_lane_s16(__p0_157, __p1_157, __p2_157) __extension__ ({ \
   57054   int16_t __s0_157 = __p0_157; \
   57055   int16x4_t __s1_157 = __p1_157; \
   57056   int16x4_t __rev1_157;  __rev1_157 = __builtin_shufflevector(__s1_157, __s1_157, 3, 2, 1, 0); \
   57057   int32_t __ret_157; \
   57058   __ret_157 = __noswap_vqdmullh_s16(__s0_157, __noswap_vget_lane_s16(__rev1_157, __p2_157)); \
   57059   __ret_157; \
   57060 })
   57061 #endif
   57062 
   57063 #ifdef __LITTLE_ENDIAN__
   57064 #define vqdmulls_laneq_s32(__p0_158, __p1_158, __p2_158) __extension__ ({ \
   57065   int32_t __s0_158 = __p0_158; \
   57066   int32x4_t __s1_158 = __p1_158; \
   57067   int64_t __ret_158; \
   57068   __ret_158 = vqdmulls_s32(__s0_158, vgetq_lane_s32(__s1_158, __p2_158)); \
   57069   __ret_158; \
   57070 })
   57071 #else
   57072 #define vqdmulls_laneq_s32(__p0_159, __p1_159, __p2_159) __extension__ ({ \
   57073   int32_t __s0_159 = __p0_159; \
   57074   int32x4_t __s1_159 = __p1_159; \
   57075   int32x4_t __rev1_159;  __rev1_159 = __builtin_shufflevector(__s1_159, __s1_159, 3, 2, 1, 0); \
   57076   int64_t __ret_159; \
   57077   __ret_159 = __noswap_vqdmulls_s32(__s0_159, __noswap_vgetq_lane_s32(__rev1_159, __p2_159)); \
   57078   __ret_159; \
   57079 })
   57080 #endif
   57081 
   57082 #ifdef __LITTLE_ENDIAN__
   57083 #define vqdmullh_laneq_s16(__p0_160, __p1_160, __p2_160) __extension__ ({ \
   57084   int16_t __s0_160 = __p0_160; \
   57085   int16x8_t __s1_160 = __p1_160; \
   57086   int32_t __ret_160; \
   57087   __ret_160 = vqdmullh_s16(__s0_160, vgetq_lane_s16(__s1_160, __p2_160)); \
   57088   __ret_160; \
   57089 })
   57090 #else
   57091 #define vqdmullh_laneq_s16(__p0_161, __p1_161, __p2_161) __extension__ ({ \
   57092   int16_t __s0_161 = __p0_161; \
   57093   int16x8_t __s1_161 = __p1_161; \
   57094   int16x8_t __rev1_161;  __rev1_161 = __builtin_shufflevector(__s1_161, __s1_161, 7, 6, 5, 4, 3, 2, 1, 0); \
   57095   int32_t __ret_161; \
   57096   __ret_161 = __noswap_vqdmullh_s16(__s0_161, __noswap_vgetq_lane_s16(__rev1_161, __p2_161)); \
   57097   __ret_161; \
   57098 })
   57099 #endif
   57100 
   57101 #ifdef __LITTLE_ENDIAN__
   57102 #define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   57103   int32x2_t __s0 = __p0; \
   57104   int32x4_t __s1 = __p1; \
   57105   int64x2_t __ret; \
   57106   __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   57107   __ret; \
   57108 })
   57109 #else
   57110 #define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   57111   int32x2_t __s0 = __p0; \
   57112   int32x4_t __s1 = __p1; \
   57113   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   57114   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   57115   int64x2_t __ret; \
   57116   __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   57117   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   57118   __ret; \
   57119 })
   57120 #endif
   57121 
   57122 #ifdef __LITTLE_ENDIAN__
   57123 #define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   57124   int16x4_t __s0 = __p0; \
   57125   int16x8_t __s1 = __p1; \
   57126   int32x4_t __ret; \
   57127   __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   57128   __ret; \
   57129 })
   57130 #else
   57131 #define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   57132   int16x4_t __s0 = __p0; \
   57133   int16x8_t __s1 = __p1; \
   57134   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   57135   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   57136   int32x4_t __ret; \
   57137   __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   57138   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   57139   __ret; \
   57140 })
   57141 #endif
   57142 
   57143 #ifdef __LITTLE_ENDIAN__
   57144 __ai int16_t vqmovns_s32(int32_t __p0) {
   57145   int16_t __ret;
   57146   __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
   57147   return __ret;
   57148 }
   57149 #else
   57150 __ai int16_t vqmovns_s32(int32_t __p0) {
   57151   int16_t __ret;
   57152   __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
   57153   return __ret;
   57154 }
   57155 #endif
   57156 
   57157 #ifdef __LITTLE_ENDIAN__
   57158 __ai int32_t vqmovnd_s64(int64_t __p0) {
   57159   int32_t __ret;
   57160   __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
   57161   return __ret;
   57162 }
   57163 #else
   57164 __ai int32_t vqmovnd_s64(int64_t __p0) {
   57165   int32_t __ret;
   57166   __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
   57167   return __ret;
   57168 }
   57169 #endif
   57170 
   57171 #ifdef __LITTLE_ENDIAN__
   57172 __ai int8_t vqmovnh_s16(int16_t __p0) {
   57173   int8_t __ret;
   57174   __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
   57175   return __ret;
   57176 }
   57177 #else
   57178 __ai int8_t vqmovnh_s16(int16_t __p0) {
   57179   int8_t __ret;
   57180   __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
   57181   return __ret;
   57182 }
   57183 #endif
   57184 
   57185 #ifdef __LITTLE_ENDIAN__
   57186 __ai uint16_t vqmovns_u32(uint32_t __p0) {
   57187   uint16_t __ret;
   57188   __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
   57189   return __ret;
   57190 }
   57191 #else
   57192 __ai uint16_t vqmovns_u32(uint32_t __p0) {
   57193   uint16_t __ret;
   57194   __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
   57195   return __ret;
   57196 }
   57197 #endif
   57198 
   57199 #ifdef __LITTLE_ENDIAN__
   57200 __ai uint32_t vqmovnd_u64(uint64_t __p0) {
   57201   uint32_t __ret;
   57202   __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
   57203   return __ret;
   57204 }
   57205 #else
   57206 __ai uint32_t vqmovnd_u64(uint64_t __p0) {
   57207   uint32_t __ret;
   57208   __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
   57209   return __ret;
   57210 }
   57211 #endif
   57212 
   57213 #ifdef __LITTLE_ENDIAN__
   57214 __ai uint8_t vqmovnh_u16(uint16_t __p0) {
   57215   uint8_t __ret;
   57216   __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
   57217   return __ret;
   57218 }
   57219 #else
   57220 __ai uint8_t vqmovnh_u16(uint16_t __p0) {
   57221   uint8_t __ret;
   57222   __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
   57223   return __ret;
   57224 }
   57225 #endif
   57226 
   57227 #ifdef __LITTLE_ENDIAN__
   57228 __ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
   57229   uint16x8_t __ret;
   57230   __ret = vcombine_u16(__p0, vqmovn_u32(__p1));
   57231   return __ret;
   57232 }
   57233 #else
   57234 __ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
   57235   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   57236   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   57237   uint16x8_t __ret;
   57238   __ret = __noswap_vcombine_u16(__rev0, __noswap_vqmovn_u32(__rev1));
   57239   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   57240   return __ret;
   57241 }
   57242 #endif
   57243 
   57244 #ifdef __LITTLE_ENDIAN__
   57245 __ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
   57246   uint32x4_t __ret;
   57247   __ret = vcombine_u32(__p0, vqmovn_u64(__p1));
   57248   return __ret;
   57249 }
   57250 #else
   57251 __ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
   57252   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   57253   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   57254   uint32x4_t __ret;
   57255   __ret = __noswap_vcombine_u32(__rev0, __noswap_vqmovn_u64(__rev1));
   57256   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   57257   return __ret;
   57258 }
   57259 #endif
   57260 
   57261 #ifdef __LITTLE_ENDIAN__
   57262 __ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
   57263   uint8x16_t __ret;
   57264   __ret = vcombine_u8(__p0, vqmovn_u16(__p1));
   57265   return __ret;
   57266 }
   57267 #else
   57268 __ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
   57269   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   57270   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   57271   uint8x16_t __ret;
   57272   __ret = __noswap_vcombine_u8(__rev0, __noswap_vqmovn_u16(__rev1));
   57273   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   57274   return __ret;
   57275 }
   57276 #endif
   57277 
   57278 #ifdef __LITTLE_ENDIAN__
   57279 __ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
   57280   int16x8_t __ret;
   57281   __ret = vcombine_s16(__p0, vqmovn_s32(__p1));
   57282   return __ret;
   57283 }
   57284 #else
   57285 __ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
   57286   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   57287   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   57288   int16x8_t __ret;
   57289   __ret = __noswap_vcombine_s16(__rev0, __noswap_vqmovn_s32(__rev1));
   57290   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   57291   return __ret;
   57292 }
   57293 #endif
   57294 
   57295 #ifdef __LITTLE_ENDIAN__
   57296 __ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
   57297   int32x4_t __ret;
   57298   __ret = vcombine_s32(__p0, vqmovn_s64(__p1));
   57299   return __ret;
   57300 }
   57301 #else
   57302 __ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
   57303   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   57304   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   57305   int32x4_t __ret;
   57306   __ret = __noswap_vcombine_s32(__rev0, __noswap_vqmovn_s64(__rev1));
   57307   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   57308   return __ret;
   57309 }
   57310 #endif
   57311 
   57312 #ifdef __LITTLE_ENDIAN__
   57313 __ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
   57314   int8x16_t __ret;
   57315   __ret = vcombine_s8(__p0, vqmovn_s16(__p1));
   57316   return __ret;
   57317 }
   57318 #else
   57319 __ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
   57320   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   57321   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   57322   int8x16_t __ret;
   57323   __ret = __noswap_vcombine_s8(__rev0, __noswap_vqmovn_s16(__rev1));
   57324   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   57325   return __ret;
   57326 }
   57327 #endif
   57328 
   57329 #ifdef __LITTLE_ENDIAN__
   57330 __ai int16_t vqmovuns_s32(int32_t __p0) {
   57331   int16_t __ret;
   57332   __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
   57333   return __ret;
   57334 }
   57335 #else
   57336 __ai int16_t vqmovuns_s32(int32_t __p0) {
   57337   int16_t __ret;
   57338   __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
   57339   return __ret;
   57340 }
   57341 #endif
   57342 
   57343 #ifdef __LITTLE_ENDIAN__
   57344 __ai int32_t vqmovund_s64(int64_t __p0) {
   57345   int32_t __ret;
   57346   __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
   57347   return __ret;
   57348 }
   57349 #else
   57350 __ai int32_t vqmovund_s64(int64_t __p0) {
   57351   int32_t __ret;
   57352   __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
   57353   return __ret;
   57354 }
   57355 #endif
   57356 
   57357 #ifdef __LITTLE_ENDIAN__
   57358 __ai int8_t vqmovunh_s16(int16_t __p0) {
   57359   int8_t __ret;
   57360   __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
   57361   return __ret;
   57362 }
   57363 #else
   57364 __ai int8_t vqmovunh_s16(int16_t __p0) {
   57365   int8_t __ret;
   57366   __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
   57367   return __ret;
   57368 }
   57369 #endif
   57370 
   57371 #ifdef __LITTLE_ENDIAN__
   57372 __ai uint16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
   57373   uint16x8_t __ret;
   57374   __ret = vcombine_u16((uint16x4_t)(__p0), vqmovun_s32(__p1));
   57375   return __ret;
   57376 }
   57377 #else
   57378 __ai uint16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
   57379   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   57380   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   57381   uint16x8_t __ret;
   57382   __ret = __noswap_vcombine_u16((uint16x4_t)(__rev0), __noswap_vqmovun_s32(__rev1));
   57383   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   57384   return __ret;
   57385 }
   57386 #endif
   57387 
   57388 #ifdef __LITTLE_ENDIAN__
   57389 __ai uint32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
   57390   uint32x4_t __ret;
   57391   __ret = vcombine_u32((uint32x2_t)(__p0), vqmovun_s64(__p1));
   57392   return __ret;
   57393 }
   57394 #else
   57395 __ai uint32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
   57396   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   57397   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   57398   uint32x4_t __ret;
   57399   __ret = __noswap_vcombine_u32((uint32x2_t)(__rev0), __noswap_vqmovun_s64(__rev1));
   57400   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   57401   return __ret;
   57402 }
   57403 #endif
   57404 
   57405 #ifdef __LITTLE_ENDIAN__
   57406 __ai uint8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
   57407   uint8x16_t __ret;
   57408   __ret = vcombine_u8((uint8x8_t)(__p0), vqmovun_s16(__p1));
   57409   return __ret;
   57410 }
   57411 #else
   57412 __ai uint8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
   57413   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   57414   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   57415   uint8x16_t __ret;
   57416   __ret = __noswap_vcombine_u8((uint8x8_t)(__rev0), __noswap_vqmovun_s16(__rev1));
   57417   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   57418   return __ret;
   57419 }
   57420 #endif
   57421 
   57422 #ifdef __LITTLE_ENDIAN__
   57423 __ai int64x2_t vqnegq_s64(int64x2_t __p0) {
   57424   int64x2_t __ret;
   57425   __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 35);
   57426   return __ret;
   57427 }
   57428 #else
   57429 __ai int64x2_t vqnegq_s64(int64x2_t __p0) {
   57430   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   57431   int64x2_t __ret;
   57432   __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 35);
   57433   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   57434   return __ret;
   57435 }
   57436 #endif
   57437 
   57438 #ifdef __LITTLE_ENDIAN__
   57439 __ai int64x1_t vqneg_s64(int64x1_t __p0) {
   57440   int64x1_t __ret;
   57441   __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
   57442   return __ret;
   57443 }
   57444 #else
   57445 __ai int64x1_t vqneg_s64(int64x1_t __p0) {
   57446   int64x1_t __ret;
   57447   __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
   57448   return __ret;
   57449 }
   57450 #endif
   57451 
   57452 #ifdef __LITTLE_ENDIAN__
   57453 __ai int8_t vqnegb_s8(int8_t __p0) {
   57454   int8_t __ret;
   57455   __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
   57456   return __ret;
   57457 }
   57458 #else
   57459 __ai int8_t vqnegb_s8(int8_t __p0) {
   57460   int8_t __ret;
   57461   __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
   57462   return __ret;
   57463 }
   57464 #endif
   57465 
   57466 #ifdef __LITTLE_ENDIAN__
   57467 __ai int32_t vqnegs_s32(int32_t __p0) {
   57468   int32_t __ret;
   57469   __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
   57470   return __ret;
   57471 }
   57472 #else
   57473 __ai int32_t vqnegs_s32(int32_t __p0) {
   57474   int32_t __ret;
   57475   __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
   57476   return __ret;
   57477 }
   57478 #endif
   57479 
   57480 #ifdef __LITTLE_ENDIAN__
   57481 __ai int64_t vqnegd_s64(int64_t __p0) {
   57482   int64_t __ret;
   57483   __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
   57484   return __ret;
   57485 }
   57486 #else
   57487 __ai int64_t vqnegd_s64(int64_t __p0) {
   57488   int64_t __ret;
   57489   __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
   57490   return __ret;
   57491 }
   57492 #endif
   57493 
   57494 #ifdef __LITTLE_ENDIAN__
   57495 __ai int16_t vqnegh_s16(int16_t __p0) {
   57496   int16_t __ret;
   57497   __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
   57498   return __ret;
   57499 }
   57500 #else
   57501 __ai int16_t vqnegh_s16(int16_t __p0) {
   57502   int16_t __ret;
   57503   __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
   57504   return __ret;
   57505 }
   57506 #endif
   57507 
   57508 #ifdef __LITTLE_ENDIAN__
   57509 __ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
   57510   int32_t __ret;
   57511   __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
   57512   return __ret;
   57513 }
   57514 #else
   57515 __ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
   57516   int32_t __ret;
   57517   __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
   57518   return __ret;
   57519 }
   57520 __ai int32_t __noswap_vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
   57521   int32_t __ret;
   57522   __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
   57523   return __ret;
   57524 }
   57525 #endif
   57526 
   57527 #ifdef __LITTLE_ENDIAN__
   57528 __ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
   57529   int16_t __ret;
   57530   __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
   57531   return __ret;
   57532 }
   57533 #else
   57534 __ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
   57535   int16_t __ret;
   57536   __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
   57537   return __ret;
   57538 }
   57539 __ai int16_t __noswap_vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
   57540   int16_t __ret;
   57541   __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
   57542   return __ret;
   57543 }
   57544 #endif
   57545 
   57546 #ifdef __LITTLE_ENDIAN__
   57547 #define vqrdmulhs_lane_s32(__p0_162, __p1_162, __p2_162) __extension__ ({ \
   57548   int32_t __s0_162 = __p0_162; \
   57549   int32x2_t __s1_162 = __p1_162; \
   57550   int32_t __ret_162; \
   57551   __ret_162 = vqrdmulhs_s32(__s0_162, vget_lane_s32(__s1_162, __p2_162)); \
   57552   __ret_162; \
   57553 })
   57554 #else
   57555 #define vqrdmulhs_lane_s32(__p0_163, __p1_163, __p2_163) __extension__ ({ \
   57556   int32_t __s0_163 = __p0_163; \
   57557   int32x2_t __s1_163 = __p1_163; \
   57558   int32x2_t __rev1_163;  __rev1_163 = __builtin_shufflevector(__s1_163, __s1_163, 1, 0); \
   57559   int32_t __ret_163; \
   57560   __ret_163 = __noswap_vqrdmulhs_s32(__s0_163, __noswap_vget_lane_s32(__rev1_163, __p2_163)); \
   57561   __ret_163; \
   57562 })
   57563 #endif
   57564 
   57565 #ifdef __LITTLE_ENDIAN__
   57566 #define vqrdmulhh_lane_s16(__p0_164, __p1_164, __p2_164) __extension__ ({ \
   57567   int16_t __s0_164 = __p0_164; \
   57568   int16x4_t __s1_164 = __p1_164; \
   57569   int16_t __ret_164; \
   57570   __ret_164 = vqrdmulhh_s16(__s0_164, vget_lane_s16(__s1_164, __p2_164)); \
   57571   __ret_164; \
   57572 })
   57573 #else
   57574 #define vqrdmulhh_lane_s16(__p0_165, __p1_165, __p2_165) __extension__ ({ \
   57575   int16_t __s0_165 = __p0_165; \
   57576   int16x4_t __s1_165 = __p1_165; \
   57577   int16x4_t __rev1_165;  __rev1_165 = __builtin_shufflevector(__s1_165, __s1_165, 3, 2, 1, 0); \
   57578   int16_t __ret_165; \
   57579   __ret_165 = __noswap_vqrdmulhh_s16(__s0_165, __noswap_vget_lane_s16(__rev1_165, __p2_165)); \
   57580   __ret_165; \
   57581 })
   57582 #endif
   57583 
   57584 #ifdef __LITTLE_ENDIAN__
   57585 #define vqrdmulhs_laneq_s32(__p0_166, __p1_166, __p2_166) __extension__ ({ \
   57586   int32_t __s0_166 = __p0_166; \
   57587   int32x4_t __s1_166 = __p1_166; \
   57588   int32_t __ret_166; \
   57589   __ret_166 = vqrdmulhs_s32(__s0_166, vgetq_lane_s32(__s1_166, __p2_166)); \
   57590   __ret_166; \
   57591 })
   57592 #else
   57593 #define vqrdmulhs_laneq_s32(__p0_167, __p1_167, __p2_167) __extension__ ({ \
   57594   int32_t __s0_167 = __p0_167; \
   57595   int32x4_t __s1_167 = __p1_167; \
   57596   int32x4_t __rev1_167;  __rev1_167 = __builtin_shufflevector(__s1_167, __s1_167, 3, 2, 1, 0); \
   57597   int32_t __ret_167; \
   57598   __ret_167 = __noswap_vqrdmulhs_s32(__s0_167, __noswap_vgetq_lane_s32(__rev1_167, __p2_167)); \
   57599   __ret_167; \
   57600 })
   57601 #endif
   57602 
   57603 #ifdef __LITTLE_ENDIAN__
   57604 #define vqrdmulhh_laneq_s16(__p0_168, __p1_168, __p2_168) __extension__ ({ \
   57605   int16_t __s0_168 = __p0_168; \
   57606   int16x8_t __s1_168 = __p1_168; \
   57607   int16_t __ret_168; \
   57608   __ret_168 = vqrdmulhh_s16(__s0_168, vgetq_lane_s16(__s1_168, __p2_168)); \
   57609   __ret_168; \
   57610 })
   57611 #else
   57612 #define vqrdmulhh_laneq_s16(__p0_169, __p1_169, __p2_169) __extension__ ({ \
   57613   int16_t __s0_169 = __p0_169; \
   57614   int16x8_t __s1_169 = __p1_169; \
   57615   int16x8_t __rev1_169;  __rev1_169 = __builtin_shufflevector(__s1_169, __s1_169, 7, 6, 5, 4, 3, 2, 1, 0); \
   57616   int16_t __ret_169; \
   57617   __ret_169 = __noswap_vqrdmulhh_s16(__s0_169, __noswap_vgetq_lane_s16(__rev1_169, __p2_169)); \
   57618   __ret_169; \
   57619 })
   57620 #endif
   57621 
   57622 #ifdef __LITTLE_ENDIAN__
   57623 #define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   57624   int32x4_t __s0 = __p0; \
   57625   int32x4_t __s1 = __p1; \
   57626   int32x4_t __ret; \
   57627   __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   57628   __ret; \
   57629 })
   57630 #else
   57631 #define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   57632   int32x4_t __s0 = __p0; \
   57633   int32x4_t __s1 = __p1; \
   57634   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   57635   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   57636   int32x4_t __ret; \
   57637   __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   57638   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   57639   __ret; \
   57640 })
   57641 #endif
   57642 
   57643 #ifdef __LITTLE_ENDIAN__
   57644 #define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   57645   int16x8_t __s0 = __p0; \
   57646   int16x8_t __s1 = __p1; \
   57647   int16x8_t __ret; \
   57648   __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
   57649   __ret; \
   57650 })
   57651 #else
   57652 #define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   57653   int16x8_t __s0 = __p0; \
   57654   int16x8_t __s1 = __p1; \
   57655   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
   57656   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   57657   int16x8_t __ret; \
   57658   __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
   57659   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
   57660   __ret; \
   57661 })
   57662 #endif
   57663 
   57664 #ifdef __LITTLE_ENDIAN__
   57665 #define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   57666   int32x2_t __s0 = __p0; \
   57667   int32x4_t __s1 = __p1; \
   57668   int32x2_t __ret; \
   57669   __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
   57670   __ret; \
   57671 })
   57672 #else
   57673 #define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
   57674   int32x2_t __s0 = __p0; \
   57675   int32x4_t __s1 = __p1; \
   57676   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   57677   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   57678   int32x2_t __ret; \
   57679   __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
   57680   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   57681   __ret; \
   57682 })
   57683 #endif
   57684 
   57685 #ifdef __LITTLE_ENDIAN__
   57686 #define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   57687   int16x4_t __s0 = __p0; \
   57688   int16x8_t __s1 = __p1; \
   57689   int16x4_t __ret; \
   57690   __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
   57691   __ret; \
   57692 })
   57693 #else
   57694 #define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
   57695   int16x4_t __s0 = __p0; \
   57696   int16x8_t __s1 = __p1; \
   57697   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   57698   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
   57699   int16x4_t __ret; \
   57700   __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
   57701   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   57702   __ret; \
   57703 })
   57704 #endif
   57705 
   57706 #ifdef __LITTLE_ENDIAN__
   57707 __ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
   57708   uint8_t __ret;
   57709   __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
   57710   return __ret;
   57711 }
   57712 #else
   57713 __ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
   57714   uint8_t __ret;
   57715   __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
   57716   return __ret;
   57717 }
   57718 #endif
   57719 
   57720 #ifdef __LITTLE_ENDIAN__
   57721 __ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
   57722   uint32_t __ret;
   57723   __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
   57724   return __ret;
   57725 }
   57726 #else
   57727 __ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
   57728   uint32_t __ret;
   57729   __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
   57730   return __ret;
   57731 }
   57732 #endif
   57733 
   57734 #ifdef __LITTLE_ENDIAN__
   57735 __ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
   57736   uint64_t __ret;
   57737   __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
   57738   return __ret;
   57739 }
   57740 #else
   57741 __ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
   57742   uint64_t __ret;
   57743   __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
   57744   return __ret;
   57745 }
   57746 #endif
   57747 
   57748 #ifdef __LITTLE_ENDIAN__
   57749 __ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
   57750   uint16_t __ret;
   57751   __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
   57752   return __ret;
   57753 }
   57754 #else
   57755 __ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
   57756   uint16_t __ret;
   57757   __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
   57758   return __ret;
   57759 }
   57760 #endif
   57761 
   57762 #ifdef __LITTLE_ENDIAN__
   57763 __ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
   57764   int8_t __ret;
   57765   __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
   57766   return __ret;
   57767 }
   57768 #else
   57769 __ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
   57770   int8_t __ret;
   57771   __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
   57772   return __ret;
   57773 }
   57774 #endif
   57775 
   57776 #ifdef __LITTLE_ENDIAN__
   57777 __ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
   57778   int32_t __ret;
   57779   __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
   57780   return __ret;
   57781 }
   57782 #else
   57783 __ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
   57784   int32_t __ret;
   57785   __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
   57786   return __ret;
   57787 }
   57788 #endif
   57789 
   57790 #ifdef __LITTLE_ENDIAN__
   57791 __ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
   57792   int64_t __ret;
   57793   __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
   57794   return __ret;
   57795 }
   57796 #else
   57797 __ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
   57798   int64_t __ret;
   57799   __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
   57800   return __ret;
   57801 }
   57802 #endif
   57803 
   57804 #ifdef __LITTLE_ENDIAN__
   57805 __ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
   57806   int16_t __ret;
   57807   __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
   57808   return __ret;
   57809 }
   57810 #else
   57811 __ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
   57812   int16_t __ret;
   57813   __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
   57814   return __ret;
   57815 }
   57816 #endif
   57817 
   57818 #ifdef __LITTLE_ENDIAN__
   57819 #define vqrshrn_high_n_u32(__p0_170, __p1_170, __p2_170) __extension__ ({ \
   57820   uint16x4_t __s0_170 = __p0_170; \
   57821   uint32x4_t __s1_170 = __p1_170; \
   57822   uint16x8_t __ret_170; \
   57823   __ret_170 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_170), (uint16x4_t)(vqrshrn_n_u32(__s1_170, __p2_170)))); \
   57824   __ret_170; \
   57825 })
   57826 #else
   57827 #define vqrshrn_high_n_u32(__p0_171, __p1_171, __p2_171) __extension__ ({ \
   57828   uint16x4_t __s0_171 = __p0_171; \
   57829   uint32x4_t __s1_171 = __p1_171; \
   57830   uint16x4_t __rev0_171;  __rev0_171 = __builtin_shufflevector(__s0_171, __s0_171, 3, 2, 1, 0); \
   57831   uint32x4_t __rev1_171;  __rev1_171 = __builtin_shufflevector(__s1_171, __s1_171, 3, 2, 1, 0); \
   57832   uint16x8_t __ret_171; \
   57833   __ret_171 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_171), (uint16x4_t)(__noswap_vqrshrn_n_u32(__rev1_171, __p2_171)))); \
   57834   __ret_171 = __builtin_shufflevector(__ret_171, __ret_171, 7, 6, 5, 4, 3, 2, 1, 0); \
   57835   __ret_171; \
   57836 })
   57837 #endif
   57838 
   57839 #ifdef __LITTLE_ENDIAN__
   57840 #define vqrshrn_high_n_u64(__p0_172, __p1_172, __p2_172) __extension__ ({ \
   57841   uint32x2_t __s0_172 = __p0_172; \
   57842   uint64x2_t __s1_172 = __p1_172; \
   57843   uint32x4_t __ret_172; \
   57844   __ret_172 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_172), (uint32x2_t)(vqrshrn_n_u64(__s1_172, __p2_172)))); \
   57845   __ret_172; \
   57846 })
   57847 #else
   57848 #define vqrshrn_high_n_u64(__p0_173, __p1_173, __p2_173) __extension__ ({ \
   57849   uint32x2_t __s0_173 = __p0_173; \
   57850   uint64x2_t __s1_173 = __p1_173; \
   57851   uint32x2_t __rev0_173;  __rev0_173 = __builtin_shufflevector(__s0_173, __s0_173, 1, 0); \
   57852   uint64x2_t __rev1_173;  __rev1_173 = __builtin_shufflevector(__s1_173, __s1_173, 1, 0); \
   57853   uint32x4_t __ret_173; \
   57854   __ret_173 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_173), (uint32x2_t)(__noswap_vqrshrn_n_u64(__rev1_173, __p2_173)))); \
   57855   __ret_173 = __builtin_shufflevector(__ret_173, __ret_173, 3, 2, 1, 0); \
   57856   __ret_173; \
   57857 })
   57858 #endif
   57859 
   57860 #ifdef __LITTLE_ENDIAN__
   57861 #define vqrshrn_high_n_u16(__p0_174, __p1_174, __p2_174) __extension__ ({ \
   57862   uint8x8_t __s0_174 = __p0_174; \
   57863   uint16x8_t __s1_174 = __p1_174; \
   57864   uint8x16_t __ret_174; \
   57865   __ret_174 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_174), (uint8x8_t)(vqrshrn_n_u16(__s1_174, __p2_174)))); \
   57866   __ret_174; \
   57867 })
   57868 #else
   57869 #define vqrshrn_high_n_u16(__p0_175, __p1_175, __p2_175) __extension__ ({ \
   57870   uint8x8_t __s0_175 = __p0_175; \
   57871   uint16x8_t __s1_175 = __p1_175; \
   57872   uint8x8_t __rev0_175;  __rev0_175 = __builtin_shufflevector(__s0_175, __s0_175, 7, 6, 5, 4, 3, 2, 1, 0); \
   57873   uint16x8_t __rev1_175;  __rev1_175 = __builtin_shufflevector(__s1_175, __s1_175, 7, 6, 5, 4, 3, 2, 1, 0); \
   57874   uint8x16_t __ret_175; \
   57875   __ret_175 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_175), (uint8x8_t)(__noswap_vqrshrn_n_u16(__rev1_175, __p2_175)))); \
   57876   __ret_175 = __builtin_shufflevector(__ret_175, __ret_175, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   57877   __ret_175; \
   57878 })
   57879 #endif
   57880 
   57881 #ifdef __LITTLE_ENDIAN__
   57882 #define vqrshrn_high_n_s32(__p0_176, __p1_176, __p2_176) __extension__ ({ \
   57883   int16x4_t __s0_176 = __p0_176; \
   57884   int32x4_t __s1_176 = __p1_176; \
   57885   int16x8_t __ret_176; \
   57886   __ret_176 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_176), (int16x4_t)(vqrshrn_n_s32(__s1_176, __p2_176)))); \
   57887   __ret_176; \
   57888 })
   57889 #else
   57890 #define vqrshrn_high_n_s32(__p0_177, __p1_177, __p2_177) __extension__ ({ \
   57891   int16x4_t __s0_177 = __p0_177; \
   57892   int32x4_t __s1_177 = __p1_177; \
   57893   int16x4_t __rev0_177;  __rev0_177 = __builtin_shufflevector(__s0_177, __s0_177, 3, 2, 1, 0); \
   57894   int32x4_t __rev1_177;  __rev1_177 = __builtin_shufflevector(__s1_177, __s1_177, 3, 2, 1, 0); \
   57895   int16x8_t __ret_177; \
   57896   __ret_177 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_177), (int16x4_t)(__noswap_vqrshrn_n_s32(__rev1_177, __p2_177)))); \
   57897   __ret_177 = __builtin_shufflevector(__ret_177, __ret_177, 7, 6, 5, 4, 3, 2, 1, 0); \
   57898   __ret_177; \
   57899 })
   57900 #endif
   57901 
   57902 #ifdef __LITTLE_ENDIAN__
   57903 #define vqrshrn_high_n_s64(__p0_178, __p1_178, __p2_178) __extension__ ({ \
   57904   int32x2_t __s0_178 = __p0_178; \
   57905   int64x2_t __s1_178 = __p1_178; \
   57906   int32x4_t __ret_178; \
   57907   __ret_178 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_178), (int32x2_t)(vqrshrn_n_s64(__s1_178, __p2_178)))); \
   57908   __ret_178; \
   57909 })
   57910 #else
   57911 #define vqrshrn_high_n_s64(__p0_179, __p1_179, __p2_179) __extension__ ({ \
   57912   int32x2_t __s0_179 = __p0_179; \
   57913   int64x2_t __s1_179 = __p1_179; \
   57914   int32x2_t __rev0_179;  __rev0_179 = __builtin_shufflevector(__s0_179, __s0_179, 1, 0); \
   57915   int64x2_t __rev1_179;  __rev1_179 = __builtin_shufflevector(__s1_179, __s1_179, 1, 0); \
   57916   int32x4_t __ret_179; \
   57917   __ret_179 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_179), (int32x2_t)(__noswap_vqrshrn_n_s64(__rev1_179, __p2_179)))); \
   57918   __ret_179 = __builtin_shufflevector(__ret_179, __ret_179, 3, 2, 1, 0); \
   57919   __ret_179; \
   57920 })
   57921 #endif
   57922 
   57923 #ifdef __LITTLE_ENDIAN__
   57924 #define vqrshrn_high_n_s16(__p0_180, __p1_180, __p2_180) __extension__ ({ \
   57925   int8x8_t __s0_180 = __p0_180; \
   57926   int16x8_t __s1_180 = __p1_180; \
   57927   int8x16_t __ret_180; \
   57928   __ret_180 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_180), (int8x8_t)(vqrshrn_n_s16(__s1_180, __p2_180)))); \
   57929   __ret_180; \
   57930 })
   57931 #else
   57932 #define vqrshrn_high_n_s16(__p0_181, __p1_181, __p2_181) __extension__ ({ \
   57933   int8x8_t __s0_181 = __p0_181; \
   57934   int16x8_t __s1_181 = __p1_181; \
   57935   int8x8_t __rev0_181;  __rev0_181 = __builtin_shufflevector(__s0_181, __s0_181, 7, 6, 5, 4, 3, 2, 1, 0); \
   57936   int16x8_t __rev1_181;  __rev1_181 = __builtin_shufflevector(__s1_181, __s1_181, 7, 6, 5, 4, 3, 2, 1, 0); \
   57937   int8x16_t __ret_181; \
   57938   __ret_181 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_181), (int8x8_t)(__noswap_vqrshrn_n_s16(__rev1_181, __p2_181)))); \
   57939   __ret_181 = __builtin_shufflevector(__ret_181, __ret_181, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   57940   __ret_181; \
   57941 })
   57942 #endif
   57943 
   57944 #ifdef __LITTLE_ENDIAN__
   57945 #define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
   57946   uint32_t __s0 = __p0; \
   57947   uint16_t __ret; \
   57948   __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
   57949   __ret; \
   57950 })
   57951 #else
   57952 #define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
   57953   uint32_t __s0 = __p0; \
   57954   uint16_t __ret; \
   57955   __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
   57956   __ret; \
   57957 })
   57958 #endif
   57959 
   57960 #ifdef __LITTLE_ENDIAN__
   57961 #define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
   57962   uint64_t __s0 = __p0; \
   57963   uint32_t __ret; \
   57964   __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
   57965   __ret; \
   57966 })
   57967 #else
   57968 #define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
   57969   uint64_t __s0 = __p0; \
   57970   uint32_t __ret; \
   57971   __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
   57972   __ret; \
   57973 })
   57974 #endif
   57975 
   57976 #ifdef __LITTLE_ENDIAN__
   57977 #define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
   57978   uint16_t __s0 = __p0; \
   57979   uint8_t __ret; \
   57980   __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
   57981   __ret; \
   57982 })
   57983 #else
   57984 #define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
   57985   uint16_t __s0 = __p0; \
   57986   uint8_t __ret; \
   57987   __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
   57988   __ret; \
   57989 })
   57990 #endif
   57991 
   57992 #ifdef __LITTLE_ENDIAN__
   57993 #define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
   57994   int32_t __s0 = __p0; \
   57995   int16_t __ret; \
   57996   __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
   57997   __ret; \
   57998 })
   57999 #else
   58000 #define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
   58001   int32_t __s0 = __p0; \
   58002   int16_t __ret; \
   58003   __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
   58004   __ret; \
   58005 })
   58006 #endif
   58007 
   58008 #ifdef __LITTLE_ENDIAN__
   58009 #define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
   58010   int64_t __s0 = __p0; \
   58011   int32_t __ret; \
   58012   __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
   58013   __ret; \
   58014 })
   58015 #else
   58016 #define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
   58017   int64_t __s0 = __p0; \
   58018   int32_t __ret; \
   58019   __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
   58020   __ret; \
   58021 })
   58022 #endif
   58023 
   58024 #ifdef __LITTLE_ENDIAN__
   58025 #define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
   58026   int16_t __s0 = __p0; \
   58027   int8_t __ret; \
   58028   __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
   58029   __ret; \
   58030 })
   58031 #else
   58032 #define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
   58033   int16_t __s0 = __p0; \
   58034   int8_t __ret; \
   58035   __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
   58036   __ret; \
   58037 })
   58038 #endif
   58039 
   58040 #ifdef __LITTLE_ENDIAN__
   58041 #define vqrshrun_high_n_s32(__p0_182, __p1_182, __p2_182) __extension__ ({ \
   58042   int16x4_t __s0_182 = __p0_182; \
   58043   int32x4_t __s1_182 = __p1_182; \
   58044   int16x8_t __ret_182; \
   58045   __ret_182 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_182), (int16x4_t)(vqrshrun_n_s32(__s1_182, __p2_182)))); \
   58046   __ret_182; \
   58047 })
   58048 #else
   58049 #define vqrshrun_high_n_s32(__p0_183, __p1_183, __p2_183) __extension__ ({ \
   58050   int16x4_t __s0_183 = __p0_183; \
   58051   int32x4_t __s1_183 = __p1_183; \
   58052   int16x4_t __rev0_183;  __rev0_183 = __builtin_shufflevector(__s0_183, __s0_183, 3, 2, 1, 0); \
   58053   int32x4_t __rev1_183;  __rev1_183 = __builtin_shufflevector(__s1_183, __s1_183, 3, 2, 1, 0); \
   58054   int16x8_t __ret_183; \
   58055   __ret_183 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_183), (int16x4_t)(__noswap_vqrshrun_n_s32(__rev1_183, __p2_183)))); \
   58056   __ret_183 = __builtin_shufflevector(__ret_183, __ret_183, 7, 6, 5, 4, 3, 2, 1, 0); \
   58057   __ret_183; \
   58058 })
   58059 #endif
   58060 
   58061 #ifdef __LITTLE_ENDIAN__
   58062 #define vqrshrun_high_n_s64(__p0_184, __p1_184, __p2_184) __extension__ ({ \
   58063   int32x2_t __s0_184 = __p0_184; \
   58064   int64x2_t __s1_184 = __p1_184; \
   58065   int32x4_t __ret_184; \
   58066   __ret_184 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_184), (int32x2_t)(vqrshrun_n_s64(__s1_184, __p2_184)))); \
   58067   __ret_184; \
   58068 })
   58069 #else
   58070 #define vqrshrun_high_n_s64(__p0_185, __p1_185, __p2_185) __extension__ ({ \
   58071   int32x2_t __s0_185 = __p0_185; \
   58072   int64x2_t __s1_185 = __p1_185; \
   58073   int32x2_t __rev0_185;  __rev0_185 = __builtin_shufflevector(__s0_185, __s0_185, 1, 0); \
   58074   int64x2_t __rev1_185;  __rev1_185 = __builtin_shufflevector(__s1_185, __s1_185, 1, 0); \
   58075   int32x4_t __ret_185; \
   58076   __ret_185 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_185), (int32x2_t)(__noswap_vqrshrun_n_s64(__rev1_185, __p2_185)))); \
   58077   __ret_185 = __builtin_shufflevector(__ret_185, __ret_185, 3, 2, 1, 0); \
   58078   __ret_185; \
   58079 })
   58080 #endif
   58081 
   58082 #ifdef __LITTLE_ENDIAN__
   58083 #define vqrshrun_high_n_s16(__p0_186, __p1_186, __p2_186) __extension__ ({ \
   58084   int8x8_t __s0_186 = __p0_186; \
   58085   int16x8_t __s1_186 = __p1_186; \
   58086   int8x16_t __ret_186; \
   58087   __ret_186 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_186), (int8x8_t)(vqrshrun_n_s16(__s1_186, __p2_186)))); \
   58088   __ret_186; \
   58089 })
   58090 #else
   58091 #define vqrshrun_high_n_s16(__p0_187, __p1_187, __p2_187) __extension__ ({ \
   58092   int8x8_t __s0_187 = __p0_187; \
   58093   int16x8_t __s1_187 = __p1_187; \
   58094   int8x8_t __rev0_187;  __rev0_187 = __builtin_shufflevector(__s0_187, __s0_187, 7, 6, 5, 4, 3, 2, 1, 0); \
   58095   int16x8_t __rev1_187;  __rev1_187 = __builtin_shufflevector(__s1_187, __s1_187, 7, 6, 5, 4, 3, 2, 1, 0); \
   58096   int8x16_t __ret_187; \
   58097   __ret_187 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_187), (int8x8_t)(__noswap_vqrshrun_n_s16(__rev1_187, __p2_187)))); \
   58098   __ret_187 = __builtin_shufflevector(__ret_187, __ret_187, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   58099   __ret_187; \
   58100 })
   58101 #endif
   58102 
   58103 #ifdef __LITTLE_ENDIAN__
   58104 #define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
   58105   int32_t __s0 = __p0; \
   58106   int16_t __ret; \
   58107   __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
   58108   __ret; \
   58109 })
   58110 #else
   58111 #define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
   58112   int32_t __s0 = __p0; \
   58113   int16_t __ret; \
   58114   __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
   58115   __ret; \
   58116 })
   58117 #endif
   58118 
   58119 #ifdef __LITTLE_ENDIAN__
   58120 #define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
   58121   int64_t __s0 = __p0; \
   58122   int32_t __ret; \
   58123   __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
   58124   __ret; \
   58125 })
   58126 #else
   58127 #define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
   58128   int64_t __s0 = __p0; \
   58129   int32_t __ret; \
   58130   __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
   58131   __ret; \
   58132 })
   58133 #endif
   58134 
   58135 #ifdef __LITTLE_ENDIAN__
   58136 #define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
   58137   int16_t __s0 = __p0; \
   58138   int8_t __ret; \
   58139   __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
   58140   __ret; \
   58141 })
   58142 #else
   58143 #define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
   58144   int16_t __s0 = __p0; \
   58145   int8_t __ret; \
   58146   __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
   58147   __ret; \
   58148 })
   58149 #endif
   58150 
   58151 #ifdef __LITTLE_ENDIAN__
   58152 __ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
   58153   uint8_t __ret;
   58154   __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
   58155   return __ret;
   58156 }
   58157 #else
   58158 __ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
   58159   uint8_t __ret;
   58160   __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
   58161   return __ret;
   58162 }
   58163 #endif
   58164 
   58165 #ifdef __LITTLE_ENDIAN__
   58166 __ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
   58167   uint32_t __ret;
   58168   __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
   58169   return __ret;
   58170 }
   58171 #else
   58172 __ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
   58173   uint32_t __ret;
   58174   __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
   58175   return __ret;
   58176 }
   58177 #endif
   58178 
   58179 #ifdef __LITTLE_ENDIAN__
   58180 __ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
   58181   uint64_t __ret;
   58182   __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
   58183   return __ret;
   58184 }
   58185 #else
   58186 __ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
   58187   uint64_t __ret;
   58188   __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
   58189   return __ret;
   58190 }
   58191 #endif
   58192 
   58193 #ifdef __LITTLE_ENDIAN__
   58194 __ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
   58195   uint16_t __ret;
   58196   __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
   58197   return __ret;
   58198 }
   58199 #else
   58200 __ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
   58201   uint16_t __ret;
   58202   __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
   58203   return __ret;
   58204 }
   58205 #endif
   58206 
   58207 #ifdef __LITTLE_ENDIAN__
   58208 __ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
   58209   int8_t __ret;
   58210   __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
   58211   return __ret;
   58212 }
   58213 #else
   58214 __ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
   58215   int8_t __ret;
   58216   __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
   58217   return __ret;
   58218 }
   58219 #endif
   58220 
   58221 #ifdef __LITTLE_ENDIAN__
   58222 __ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
   58223   int32_t __ret;
   58224   __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
   58225   return __ret;
   58226 }
   58227 #else
   58228 __ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
   58229   int32_t __ret;
   58230   __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
   58231   return __ret;
   58232 }
   58233 #endif
   58234 
   58235 #ifdef __LITTLE_ENDIAN__
   58236 __ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
   58237   int64_t __ret;
   58238   __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
   58239   return __ret;
   58240 }
   58241 #else
   58242 __ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
   58243   int64_t __ret;
   58244   __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
   58245   return __ret;
   58246 }
   58247 #endif
   58248 
   58249 #ifdef __LITTLE_ENDIAN__
   58250 __ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
   58251   int16_t __ret;
   58252   __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
   58253   return __ret;
   58254 }
   58255 #else
   58256 __ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
   58257   int16_t __ret;
   58258   __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
   58259   return __ret;
   58260 }
   58261 #endif
   58262 
   58263 #ifdef __LITTLE_ENDIAN__
   58264 #define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
   58265   uint8_t __s0 = __p0; \
   58266   uint8_t __ret; \
   58267   __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
   58268   __ret; \
   58269 })
   58270 #else
   58271 #define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
   58272   uint8_t __s0 = __p0; \
   58273   uint8_t __ret; \
   58274   __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
   58275   __ret; \
   58276 })
   58277 #endif
   58278 
   58279 #ifdef __LITTLE_ENDIAN__
   58280 #define vqshls_n_u32(__p0, __p1) __extension__ ({ \
   58281   uint32_t __s0 = __p0; \
   58282   uint32_t __ret; \
   58283   __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
   58284   __ret; \
   58285 })
   58286 #else
   58287 #define vqshls_n_u32(__p0, __p1) __extension__ ({ \
   58288   uint32_t __s0 = __p0; \
   58289   uint32_t __ret; \
   58290   __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
   58291   __ret; \
   58292 })
   58293 #endif
   58294 
   58295 #ifdef __LITTLE_ENDIAN__
   58296 #define vqshld_n_u64(__p0, __p1) __extension__ ({ \
   58297   uint64_t __s0 = __p0; \
   58298   uint64_t __ret; \
   58299   __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
   58300   __ret; \
   58301 })
   58302 #else
   58303 #define vqshld_n_u64(__p0, __p1) __extension__ ({ \
   58304   uint64_t __s0 = __p0; \
   58305   uint64_t __ret; \
   58306   __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
   58307   __ret; \
   58308 })
   58309 #endif
   58310 
   58311 #ifdef __LITTLE_ENDIAN__
   58312 #define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
   58313   uint16_t __s0 = __p0; \
   58314   uint16_t __ret; \
   58315   __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
   58316   __ret; \
   58317 })
   58318 #else
   58319 #define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
   58320   uint16_t __s0 = __p0; \
   58321   uint16_t __ret; \
   58322   __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
   58323   __ret; \
   58324 })
   58325 #endif
   58326 
   58327 #ifdef __LITTLE_ENDIAN__
   58328 #define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
   58329   int8_t __s0 = __p0; \
   58330   int8_t __ret; \
   58331   __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
   58332   __ret; \
   58333 })
   58334 #else
   58335 #define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
   58336   int8_t __s0 = __p0; \
   58337   int8_t __ret; \
   58338   __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
   58339   __ret; \
   58340 })
   58341 #endif
   58342 
   58343 #ifdef __LITTLE_ENDIAN__
   58344 #define vqshls_n_s32(__p0, __p1) __extension__ ({ \
   58345   int32_t __s0 = __p0; \
   58346   int32_t __ret; \
   58347   __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
   58348   __ret; \
   58349 })
   58350 #else
   58351 #define vqshls_n_s32(__p0, __p1) __extension__ ({ \
   58352   int32_t __s0 = __p0; \
   58353   int32_t __ret; \
   58354   __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
   58355   __ret; \
   58356 })
   58357 #endif
   58358 
   58359 #ifdef __LITTLE_ENDIAN__
   58360 #define vqshld_n_s64(__p0, __p1) __extension__ ({ \
   58361   int64_t __s0 = __p0; \
   58362   int64_t __ret; \
   58363   __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
   58364   __ret; \
   58365 })
   58366 #else
   58367 #define vqshld_n_s64(__p0, __p1) __extension__ ({ \
   58368   int64_t __s0 = __p0; \
   58369   int64_t __ret; \
   58370   __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
   58371   __ret; \
   58372 })
   58373 #endif
   58374 
   58375 #ifdef __LITTLE_ENDIAN__
   58376 #define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
   58377   int16_t __s0 = __p0; \
   58378   int16_t __ret; \
   58379   __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
   58380   __ret; \
   58381 })
   58382 #else
   58383 #define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
   58384   int16_t __s0 = __p0; \
   58385   int16_t __ret; \
   58386   __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
   58387   __ret; \
   58388 })
   58389 #endif
   58390 
   58391 #ifdef __LITTLE_ENDIAN__
   58392 #define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
   58393   int8_t __s0 = __p0; \
   58394   int8_t __ret; \
   58395   __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
   58396   __ret; \
   58397 })
   58398 #else
   58399 #define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
   58400   int8_t __s0 = __p0; \
   58401   int8_t __ret; \
   58402   __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
   58403   __ret; \
   58404 })
   58405 #endif
   58406 
   58407 #ifdef __LITTLE_ENDIAN__
   58408 #define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
   58409   int32_t __s0 = __p0; \
   58410   int32_t __ret; \
   58411   __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
   58412   __ret; \
   58413 })
   58414 #else
   58415 #define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
   58416   int32_t __s0 = __p0; \
   58417   int32_t __ret; \
   58418   __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
   58419   __ret; \
   58420 })
   58421 #endif
   58422 
   58423 #ifdef __LITTLE_ENDIAN__
   58424 #define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
   58425   int64_t __s0 = __p0; \
   58426   int64_t __ret; \
   58427   __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
   58428   __ret; \
   58429 })
   58430 #else
   58431 #define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
   58432   int64_t __s0 = __p0; \
   58433   int64_t __ret; \
   58434   __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
   58435   __ret; \
   58436 })
   58437 #endif
   58438 
   58439 #ifdef __LITTLE_ENDIAN__
   58440 #define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
   58441   int16_t __s0 = __p0; \
   58442   int16_t __ret; \
   58443   __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
   58444   __ret; \
   58445 })
   58446 #else
   58447 #define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
   58448   int16_t __s0 = __p0; \
   58449   int16_t __ret; \
   58450   __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
   58451   __ret; \
   58452 })
   58453 #endif
   58454 
   58455 #ifdef __LITTLE_ENDIAN__
   58456 #define vqshrn_high_n_u32(__p0_188, __p1_188, __p2_188) __extension__ ({ \
   58457   uint16x4_t __s0_188 = __p0_188; \
   58458   uint32x4_t __s1_188 = __p1_188; \
   58459   uint16x8_t __ret_188; \
   58460   __ret_188 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_188), (uint16x4_t)(vqshrn_n_u32(__s1_188, __p2_188)))); \
   58461   __ret_188; \
   58462 })
   58463 #else
   58464 #define vqshrn_high_n_u32(__p0_189, __p1_189, __p2_189) __extension__ ({ \
   58465   uint16x4_t __s0_189 = __p0_189; \
   58466   uint32x4_t __s1_189 = __p1_189; \
   58467   uint16x4_t __rev0_189;  __rev0_189 = __builtin_shufflevector(__s0_189, __s0_189, 3, 2, 1, 0); \
   58468   uint32x4_t __rev1_189;  __rev1_189 = __builtin_shufflevector(__s1_189, __s1_189, 3, 2, 1, 0); \
   58469   uint16x8_t __ret_189; \
   58470   __ret_189 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_189), (uint16x4_t)(__noswap_vqshrn_n_u32(__rev1_189, __p2_189)))); \
   58471   __ret_189 = __builtin_shufflevector(__ret_189, __ret_189, 7, 6, 5, 4, 3, 2, 1, 0); \
   58472   __ret_189; \
   58473 })
   58474 #endif
   58475 
   58476 #ifdef __LITTLE_ENDIAN__
   58477 #define vqshrn_high_n_u64(__p0_190, __p1_190, __p2_190) __extension__ ({ \
   58478   uint32x2_t __s0_190 = __p0_190; \
   58479   uint64x2_t __s1_190 = __p1_190; \
   58480   uint32x4_t __ret_190; \
   58481   __ret_190 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_190), (uint32x2_t)(vqshrn_n_u64(__s1_190, __p2_190)))); \
   58482   __ret_190; \
   58483 })
   58484 #else
   58485 #define vqshrn_high_n_u64(__p0_191, __p1_191, __p2_191) __extension__ ({ \
   58486   uint32x2_t __s0_191 = __p0_191; \
   58487   uint64x2_t __s1_191 = __p1_191; \
   58488   uint32x2_t __rev0_191;  __rev0_191 = __builtin_shufflevector(__s0_191, __s0_191, 1, 0); \
   58489   uint64x2_t __rev1_191;  __rev1_191 = __builtin_shufflevector(__s1_191, __s1_191, 1, 0); \
   58490   uint32x4_t __ret_191; \
   58491   __ret_191 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_191), (uint32x2_t)(__noswap_vqshrn_n_u64(__rev1_191, __p2_191)))); \
   58492   __ret_191 = __builtin_shufflevector(__ret_191, __ret_191, 3, 2, 1, 0); \
   58493   __ret_191; \
   58494 })
   58495 #endif
   58496 
   58497 #ifdef __LITTLE_ENDIAN__
   58498 #define vqshrn_high_n_u16(__p0_192, __p1_192, __p2_192) __extension__ ({ \
   58499   uint8x8_t __s0_192 = __p0_192; \
   58500   uint16x8_t __s1_192 = __p1_192; \
   58501   uint8x16_t __ret_192; \
   58502   __ret_192 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_192), (uint8x8_t)(vqshrn_n_u16(__s1_192, __p2_192)))); \
   58503   __ret_192; \
   58504 })
   58505 #else
   58506 #define vqshrn_high_n_u16(__p0_193, __p1_193, __p2_193) __extension__ ({ \
   58507   uint8x8_t __s0_193 = __p0_193; \
   58508   uint16x8_t __s1_193 = __p1_193; \
   58509   uint8x8_t __rev0_193;  __rev0_193 = __builtin_shufflevector(__s0_193, __s0_193, 7, 6, 5, 4, 3, 2, 1, 0); \
   58510   uint16x8_t __rev1_193;  __rev1_193 = __builtin_shufflevector(__s1_193, __s1_193, 7, 6, 5, 4, 3, 2, 1, 0); \
   58511   uint8x16_t __ret_193; \
   58512   __ret_193 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_193), (uint8x8_t)(__noswap_vqshrn_n_u16(__rev1_193, __p2_193)))); \
   58513   __ret_193 = __builtin_shufflevector(__ret_193, __ret_193, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   58514   __ret_193; \
   58515 })
   58516 #endif
   58517 
   58518 #ifdef __LITTLE_ENDIAN__
   58519 #define vqshrn_high_n_s32(__p0_194, __p1_194, __p2_194) __extension__ ({ \
   58520   int16x4_t __s0_194 = __p0_194; \
   58521   int32x4_t __s1_194 = __p1_194; \
   58522   int16x8_t __ret_194; \
   58523   __ret_194 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_194), (int16x4_t)(vqshrn_n_s32(__s1_194, __p2_194)))); \
   58524   __ret_194; \
   58525 })
   58526 #else
   58527 #define vqshrn_high_n_s32(__p0_195, __p1_195, __p2_195) __extension__ ({ \
   58528   int16x4_t __s0_195 = __p0_195; \
   58529   int32x4_t __s1_195 = __p1_195; \
   58530   int16x4_t __rev0_195;  __rev0_195 = __builtin_shufflevector(__s0_195, __s0_195, 3, 2, 1, 0); \
   58531   int32x4_t __rev1_195;  __rev1_195 = __builtin_shufflevector(__s1_195, __s1_195, 3, 2, 1, 0); \
   58532   int16x8_t __ret_195; \
   58533   __ret_195 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_195), (int16x4_t)(__noswap_vqshrn_n_s32(__rev1_195, __p2_195)))); \
   58534   __ret_195 = __builtin_shufflevector(__ret_195, __ret_195, 7, 6, 5, 4, 3, 2, 1, 0); \
   58535   __ret_195; \
   58536 })
   58537 #endif
   58538 
   58539 #ifdef __LITTLE_ENDIAN__
   58540 #define vqshrn_high_n_s64(__p0_196, __p1_196, __p2_196) __extension__ ({ \
   58541   int32x2_t __s0_196 = __p0_196; \
   58542   int64x2_t __s1_196 = __p1_196; \
   58543   int32x4_t __ret_196; \
   58544   __ret_196 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_196), (int32x2_t)(vqshrn_n_s64(__s1_196, __p2_196)))); \
   58545   __ret_196; \
   58546 })
   58547 #else
   58548 #define vqshrn_high_n_s64(__p0_197, __p1_197, __p2_197) __extension__ ({ \
   58549   int32x2_t __s0_197 = __p0_197; \
   58550   int64x2_t __s1_197 = __p1_197; \
   58551   int32x2_t __rev0_197;  __rev0_197 = __builtin_shufflevector(__s0_197, __s0_197, 1, 0); \
   58552   int64x2_t __rev1_197;  __rev1_197 = __builtin_shufflevector(__s1_197, __s1_197, 1, 0); \
   58553   int32x4_t __ret_197; \
   58554   __ret_197 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_197), (int32x2_t)(__noswap_vqshrn_n_s64(__rev1_197, __p2_197)))); \
   58555   __ret_197 = __builtin_shufflevector(__ret_197, __ret_197, 3, 2, 1, 0); \
   58556   __ret_197; \
   58557 })
   58558 #endif
   58559 
   58560 #ifdef __LITTLE_ENDIAN__
   58561 #define vqshrn_high_n_s16(__p0_198, __p1_198, __p2_198) __extension__ ({ \
   58562   int8x8_t __s0_198 = __p0_198; \
   58563   int16x8_t __s1_198 = __p1_198; \
   58564   int8x16_t __ret_198; \
   58565   __ret_198 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_198), (int8x8_t)(vqshrn_n_s16(__s1_198, __p2_198)))); \
   58566   __ret_198; \
   58567 })
   58568 #else
   58569 #define vqshrn_high_n_s16(__p0_199, __p1_199, __p2_199) __extension__ ({ \
   58570   int8x8_t __s0_199 = __p0_199; \
   58571   int16x8_t __s1_199 = __p1_199; \
   58572   int8x8_t __rev0_199;  __rev0_199 = __builtin_shufflevector(__s0_199, __s0_199, 7, 6, 5, 4, 3, 2, 1, 0); \
   58573   int16x8_t __rev1_199;  __rev1_199 = __builtin_shufflevector(__s1_199, __s1_199, 7, 6, 5, 4, 3, 2, 1, 0); \
   58574   int8x16_t __ret_199; \
   58575   __ret_199 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_199), (int8x8_t)(__noswap_vqshrn_n_s16(__rev1_199, __p2_199)))); \
   58576   __ret_199 = __builtin_shufflevector(__ret_199, __ret_199, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   58577   __ret_199; \
   58578 })
   58579 #endif
   58580 
   58581 #ifdef __LITTLE_ENDIAN__
   58582 #define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
   58583   uint32_t __s0 = __p0; \
   58584   uint16_t __ret; \
   58585   __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
   58586   __ret; \
   58587 })
   58588 #else
   58589 #define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
   58590   uint32_t __s0 = __p0; \
   58591   uint16_t __ret; \
   58592   __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
   58593   __ret; \
   58594 })
   58595 #endif
   58596 
   58597 #ifdef __LITTLE_ENDIAN__
   58598 #define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
   58599   uint64_t __s0 = __p0; \
   58600   uint32_t __ret; \
   58601   __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
   58602   __ret; \
   58603 })
   58604 #else
   58605 #define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
   58606   uint64_t __s0 = __p0; \
   58607   uint32_t __ret; \
   58608   __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
   58609   __ret; \
   58610 })
   58611 #endif
   58612 
   58613 #ifdef __LITTLE_ENDIAN__
   58614 #define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
   58615   uint16_t __s0 = __p0; \
   58616   uint8_t __ret; \
   58617   __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
   58618   __ret; \
   58619 })
   58620 #else
   58621 #define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
   58622   uint16_t __s0 = __p0; \
   58623   uint8_t __ret; \
   58624   __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
   58625   __ret; \
   58626 })
   58627 #endif
   58628 
   58629 #ifdef __LITTLE_ENDIAN__
   58630 #define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
   58631   int32_t __s0 = __p0; \
   58632   int16_t __ret; \
   58633   __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
   58634   __ret; \
   58635 })
   58636 #else
   58637 #define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
   58638   int32_t __s0 = __p0; \
   58639   int16_t __ret; \
   58640   __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
   58641   __ret; \
   58642 })
   58643 #endif
   58644 
   58645 #ifdef __LITTLE_ENDIAN__
   58646 #define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
   58647   int64_t __s0 = __p0; \
   58648   int32_t __ret; \
   58649   __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
   58650   __ret; \
   58651 })
   58652 #else
   58653 #define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
   58654   int64_t __s0 = __p0; \
   58655   int32_t __ret; \
   58656   __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
   58657   __ret; \
   58658 })
   58659 #endif
   58660 
   58661 #ifdef __LITTLE_ENDIAN__
   58662 #define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
   58663   int16_t __s0 = __p0; \
   58664   int8_t __ret; \
   58665   __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
   58666   __ret; \
   58667 })
   58668 #else
   58669 #define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
   58670   int16_t __s0 = __p0; \
   58671   int8_t __ret; \
   58672   __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
   58673   __ret; \
   58674 })
   58675 #endif
   58676 
   58677 #ifdef __LITTLE_ENDIAN__
   58678 #define vqshrun_high_n_s32(__p0_200, __p1_200, __p2_200) __extension__ ({ \
   58679   int16x4_t __s0_200 = __p0_200; \
   58680   int32x4_t __s1_200 = __p1_200; \
   58681   int16x8_t __ret_200; \
   58682   __ret_200 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_200), (int16x4_t)(vqshrun_n_s32(__s1_200, __p2_200)))); \
   58683   __ret_200; \
   58684 })
   58685 #else
   58686 #define vqshrun_high_n_s32(__p0_201, __p1_201, __p2_201) __extension__ ({ \
   58687   int16x4_t __s0_201 = __p0_201; \
   58688   int32x4_t __s1_201 = __p1_201; \
   58689   int16x4_t __rev0_201;  __rev0_201 = __builtin_shufflevector(__s0_201, __s0_201, 3, 2, 1, 0); \
   58690   int32x4_t __rev1_201;  __rev1_201 = __builtin_shufflevector(__s1_201, __s1_201, 3, 2, 1, 0); \
   58691   int16x8_t __ret_201; \
   58692   __ret_201 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_201), (int16x4_t)(__noswap_vqshrun_n_s32(__rev1_201, __p2_201)))); \
   58693   __ret_201 = __builtin_shufflevector(__ret_201, __ret_201, 7, 6, 5, 4, 3, 2, 1, 0); \
   58694   __ret_201; \
   58695 })
   58696 #endif
   58697 
   58698 #ifdef __LITTLE_ENDIAN__
   58699 #define vqshrun_high_n_s64(__p0_202, __p1_202, __p2_202) __extension__ ({ \
   58700   int32x2_t __s0_202 = __p0_202; \
   58701   int64x2_t __s1_202 = __p1_202; \
   58702   int32x4_t __ret_202; \
   58703   __ret_202 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_202), (int32x2_t)(vqshrun_n_s64(__s1_202, __p2_202)))); \
   58704   __ret_202; \
   58705 })
   58706 #else
   58707 #define vqshrun_high_n_s64(__p0_203, __p1_203, __p2_203) __extension__ ({ \
   58708   int32x2_t __s0_203 = __p0_203; \
   58709   int64x2_t __s1_203 = __p1_203; \
   58710   int32x2_t __rev0_203;  __rev0_203 = __builtin_shufflevector(__s0_203, __s0_203, 1, 0); \
   58711   int64x2_t __rev1_203;  __rev1_203 = __builtin_shufflevector(__s1_203, __s1_203, 1, 0); \
   58712   int32x4_t __ret_203; \
   58713   __ret_203 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_203), (int32x2_t)(__noswap_vqshrun_n_s64(__rev1_203, __p2_203)))); \
   58714   __ret_203 = __builtin_shufflevector(__ret_203, __ret_203, 3, 2, 1, 0); \
   58715   __ret_203; \
   58716 })
   58717 #endif
   58718 
   58719 #ifdef __LITTLE_ENDIAN__
   58720 #define vqshrun_high_n_s16(__p0_204, __p1_204, __p2_204) __extension__ ({ \
   58721   int8x8_t __s0_204 = __p0_204; \
   58722   int16x8_t __s1_204 = __p1_204; \
   58723   int8x16_t __ret_204; \
   58724   __ret_204 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_204), (int8x8_t)(vqshrun_n_s16(__s1_204, __p2_204)))); \
   58725   __ret_204; \
   58726 })
   58727 #else
   58728 #define vqshrun_high_n_s16(__p0_205, __p1_205, __p2_205) __extension__ ({ \
   58729   int8x8_t __s0_205 = __p0_205; \
   58730   int16x8_t __s1_205 = __p1_205; \
   58731   int8x8_t __rev0_205;  __rev0_205 = __builtin_shufflevector(__s0_205, __s0_205, 7, 6, 5, 4, 3, 2, 1, 0); \
   58732   int16x8_t __rev1_205;  __rev1_205 = __builtin_shufflevector(__s1_205, __s1_205, 7, 6, 5, 4, 3, 2, 1, 0); \
   58733   int8x16_t __ret_205; \
   58734   __ret_205 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_205), (int8x8_t)(__noswap_vqshrun_n_s16(__rev1_205, __p2_205)))); \
   58735   __ret_205 = __builtin_shufflevector(__ret_205, __ret_205, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   58736   __ret_205; \
   58737 })
   58738 #endif
   58739 
   58740 #ifdef __LITTLE_ENDIAN__
   58741 #define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
   58742   int32_t __s0 = __p0; \
   58743   int16_t __ret; \
   58744   __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
   58745   __ret; \
   58746 })
   58747 #else
   58748 #define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
   58749   int32_t __s0 = __p0; \
   58750   int16_t __ret; \
   58751   __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
   58752   __ret; \
   58753 })
   58754 #endif
   58755 
   58756 #ifdef __LITTLE_ENDIAN__
   58757 #define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
   58758   int64_t __s0 = __p0; \
   58759   int32_t __ret; \
   58760   __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
   58761   __ret; \
   58762 })
   58763 #else
   58764 #define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
   58765   int64_t __s0 = __p0; \
   58766   int32_t __ret; \
   58767   __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
   58768   __ret; \
   58769 })
   58770 #endif
   58771 
   58772 #ifdef __LITTLE_ENDIAN__
   58773 #define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
   58774   int16_t __s0 = __p0; \
   58775   int8_t __ret; \
   58776   __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
   58777   __ret; \
   58778 })
   58779 #else
   58780 #define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
   58781   int16_t __s0 = __p0; \
   58782   int8_t __ret; \
   58783   __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
   58784   __ret; \
   58785 })
   58786 #endif
   58787 
   58788 #ifdef __LITTLE_ENDIAN__
   58789 __ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
   58790   uint8_t __ret;
   58791   __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
   58792   return __ret;
   58793 }
   58794 #else
   58795 __ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
   58796   uint8_t __ret;
   58797   __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
   58798   return __ret;
   58799 }
   58800 #endif
   58801 
   58802 #ifdef __LITTLE_ENDIAN__
   58803 __ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
   58804   uint32_t __ret;
   58805   __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
   58806   return __ret;
   58807 }
   58808 #else
   58809 __ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
   58810   uint32_t __ret;
   58811   __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
   58812   return __ret;
   58813 }
   58814 #endif
   58815 
   58816 #ifdef __LITTLE_ENDIAN__
   58817 __ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
   58818   uint64_t __ret;
   58819   __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
   58820   return __ret;
   58821 }
   58822 #else
   58823 __ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
   58824   uint64_t __ret;
   58825   __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
   58826   return __ret;
   58827 }
   58828 #endif
   58829 
   58830 #ifdef __LITTLE_ENDIAN__
   58831 __ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
   58832   uint16_t __ret;
   58833   __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
   58834   return __ret;
   58835 }
   58836 #else
   58837 __ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
   58838   uint16_t __ret;
   58839   __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
   58840   return __ret;
   58841 }
   58842 #endif
   58843 
   58844 #ifdef __LITTLE_ENDIAN__
   58845 __ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
   58846   int8_t __ret;
   58847   __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
   58848   return __ret;
   58849 }
   58850 #else
   58851 __ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
   58852   int8_t __ret;
   58853   __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
   58854   return __ret;
   58855 }
   58856 #endif
   58857 
   58858 #ifdef __LITTLE_ENDIAN__
   58859 __ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
   58860   int32_t __ret;
   58861   __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
   58862   return __ret;
   58863 }
   58864 #else
   58865 __ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
   58866   int32_t __ret;
   58867   __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
   58868   return __ret;
   58869 }
   58870 __ai int32_t __noswap_vqsubs_s32(int32_t __p0, int32_t __p1) {
   58871   int32_t __ret;
   58872   __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
   58873   return __ret;
   58874 }
   58875 #endif
   58876 
   58877 #ifdef __LITTLE_ENDIAN__
   58878 __ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
   58879   int64_t __ret;
   58880   __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
   58881   return __ret;
   58882 }
   58883 #else
   58884 __ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
   58885   int64_t __ret;
   58886   __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
   58887   return __ret;
   58888 }
   58889 #endif
   58890 
   58891 #ifdef __LITTLE_ENDIAN__
   58892 __ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
   58893   int16_t __ret;
   58894   __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
   58895   return __ret;
   58896 }
   58897 #else
   58898 __ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
   58899   int16_t __ret;
   58900   __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
   58901   return __ret;
   58902 }
   58903 __ai int16_t __noswap_vqsubh_s16(int16_t __p0, int16_t __p1) {
   58904   int16_t __ret;
   58905   __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
   58906   return __ret;
   58907 }
   58908 #endif
   58909 
   58910 #ifdef __LITTLE_ENDIAN__
   58911 __ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
   58912   poly8x8_t __ret;
   58913   __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 4);
   58914   return __ret;
   58915 }
   58916 #else
   58917 __ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
   58918   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58919   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   58920   poly8x8_t __ret;
   58921   __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 4);
   58922   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   58923   return __ret;
   58924 }
   58925 #endif
   58926 
   58927 #ifdef __LITTLE_ENDIAN__
   58928 __ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
   58929   poly8x16_t __ret;
   58930   __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
   58931   return __ret;
   58932 }
   58933 #else
   58934 __ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
   58935   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58936   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58937   poly8x16_t __ret;
   58938   __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
   58939   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58940   return __ret;
   58941 }
   58942 #endif
   58943 
   58944 #ifdef __LITTLE_ENDIAN__
   58945 __ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   58946   uint8x16_t __ret;
   58947   __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   58948   return __ret;
   58949 }
   58950 #else
   58951 __ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   58952   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58953   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58954   uint8x16_t __ret;
   58955   __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   58956   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58957   return __ret;
   58958 }
   58959 #endif
   58960 
   58961 #ifdef __LITTLE_ENDIAN__
   58962 __ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
   58963   int8x16_t __ret;
   58964   __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   58965   return __ret;
   58966 }
   58967 #else
   58968 __ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
   58969   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58970   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58971   int8x16_t __ret;
   58972   __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   58973   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58974   return __ret;
   58975 }
   58976 #endif
   58977 
   58978 #ifdef __LITTLE_ENDIAN__
   58979 __ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
   58980   uint8x8_t __ret;
   58981   __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 16);
   58982   return __ret;
   58983 }
   58984 #else
   58985 __ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
   58986   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   58987   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   58988   uint8x8_t __ret;
   58989   __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 16);
   58990   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   58991   return __ret;
   58992 }
   58993 #endif
   58994 
   58995 #ifdef __LITTLE_ENDIAN__
   58996 __ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
   58997   int8x8_t __ret;
   58998   __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 0);
   58999   return __ret;
   59000 }
   59001 #else
   59002 __ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
   59003   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59004   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59005   int8x8_t __ret;
   59006   __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 0);
   59007   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59008   return __ret;
   59009 }
   59010 #endif
   59011 
   59012 #ifdef __LITTLE_ENDIAN__
   59013 __ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
   59014   poly8x8_t __ret;
   59015   __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 4);
   59016   return __ret;
   59017 }
   59018 #else
   59019 __ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
   59020   poly8x16x2_t __rev0;
   59021   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59022   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59023   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59024   poly8x8_t __ret;
   59025   __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 4);
   59026   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59027   return __ret;
   59028 }
   59029 #endif
   59030 
   59031 #ifdef __LITTLE_ENDIAN__
   59032 __ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
   59033   poly8x16_t __ret;
   59034   __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 36);
   59035   return __ret;
   59036 }
   59037 #else
   59038 __ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
   59039   poly8x16x2_t __rev0;
   59040   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59041   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59042   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59043   poly8x16_t __ret;
   59044   __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 36);
   59045   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59046   return __ret;
   59047 }
   59048 #endif
   59049 
   59050 #ifdef __LITTLE_ENDIAN__
   59051 __ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
   59052   uint8x16_t __ret;
   59053   __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 48);
   59054   return __ret;
   59055 }
   59056 #else
   59057 __ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
   59058   uint8x16x2_t __rev0;
   59059   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59060   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59061   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59062   uint8x16_t __ret;
   59063   __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 48);
   59064   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59065   return __ret;
   59066 }
   59067 #endif
   59068 
   59069 #ifdef __LITTLE_ENDIAN__
   59070 __ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
   59071   int8x16_t __ret;
   59072   __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 32);
   59073   return __ret;
   59074 }
   59075 #else
   59076 __ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
   59077   int8x16x2_t __rev0;
   59078   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59079   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59080   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59081   int8x16_t __ret;
   59082   __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 32);
   59083   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59084   return __ret;
   59085 }
   59086 #endif
   59087 
   59088 #ifdef __LITTLE_ENDIAN__
   59089 __ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
   59090   uint8x8_t __ret;
   59091   __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 16);
   59092   return __ret;
   59093 }
   59094 #else
   59095 __ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
   59096   uint8x16x2_t __rev0;
   59097   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59098   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59099   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59100   uint8x8_t __ret;
   59101   __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 16);
   59102   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59103   return __ret;
   59104 }
   59105 #endif
   59106 
   59107 #ifdef __LITTLE_ENDIAN__
   59108 __ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
   59109   int8x8_t __ret;
   59110   __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 0);
   59111   return __ret;
   59112 }
   59113 #else
   59114 __ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
   59115   int8x16x2_t __rev0;
   59116   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59117   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59118   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59119   int8x8_t __ret;
   59120   __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 0);
   59121   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59122   return __ret;
   59123 }
   59124 #endif
   59125 
   59126 #ifdef __LITTLE_ENDIAN__
   59127 __ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
   59128   poly8x8_t __ret;
   59129   __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 4);
   59130   return __ret;
   59131 }
   59132 #else
   59133 __ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
   59134   poly8x16x3_t __rev0;
   59135   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59136   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59137   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59138   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59139   poly8x8_t __ret;
   59140   __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 4);
   59141   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59142   return __ret;
   59143 }
   59144 #endif
   59145 
   59146 #ifdef __LITTLE_ENDIAN__
   59147 __ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
   59148   poly8x16_t __ret;
   59149   __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 36);
   59150   return __ret;
   59151 }
   59152 #else
   59153 __ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
   59154   poly8x16x3_t __rev0;
   59155   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59156   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59157   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59158   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59159   poly8x16_t __ret;
   59160   __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 36);
   59161   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59162   return __ret;
   59163 }
   59164 #endif
   59165 
   59166 #ifdef __LITTLE_ENDIAN__
   59167 __ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
   59168   uint8x16_t __ret;
   59169   __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 48);
   59170   return __ret;
   59171 }
   59172 #else
   59173 __ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
   59174   uint8x16x3_t __rev0;
   59175   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59176   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59177   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59178   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59179   uint8x16_t __ret;
   59180   __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 48);
   59181   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59182   return __ret;
   59183 }
   59184 #endif
   59185 
   59186 #ifdef __LITTLE_ENDIAN__
   59187 __ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
   59188   int8x16_t __ret;
   59189   __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 32);
   59190   return __ret;
   59191 }
   59192 #else
   59193 __ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
   59194   int8x16x3_t __rev0;
   59195   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59196   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59197   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59198   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59199   int8x16_t __ret;
   59200   __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 32);
   59201   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59202   return __ret;
   59203 }
   59204 #endif
   59205 
   59206 #ifdef __LITTLE_ENDIAN__
   59207 __ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
   59208   uint8x8_t __ret;
   59209   __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 16);
   59210   return __ret;
   59211 }
   59212 #else
   59213 __ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
   59214   uint8x16x3_t __rev0;
   59215   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59216   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59217   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59218   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59219   uint8x8_t __ret;
   59220   __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 16);
   59221   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59222   return __ret;
   59223 }
   59224 #endif
   59225 
   59226 #ifdef __LITTLE_ENDIAN__
   59227 __ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
   59228   int8x8_t __ret;
   59229   __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 0);
   59230   return __ret;
   59231 }
   59232 #else
   59233 __ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
   59234   int8x16x3_t __rev0;
   59235   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59236   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59237   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59238   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59239   int8x8_t __ret;
   59240   __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 0);
   59241   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59242   return __ret;
   59243 }
   59244 #endif
   59245 
   59246 #ifdef __LITTLE_ENDIAN__
   59247 __ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
   59248   poly8x8_t __ret;
   59249   __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 4);
   59250   return __ret;
   59251 }
   59252 #else
   59253 __ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
   59254   poly8x16x4_t __rev0;
   59255   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59256   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59257   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59258   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59259   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59260   poly8x8_t __ret;
   59261   __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 4);
   59262   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59263   return __ret;
   59264 }
   59265 #endif
   59266 
   59267 #ifdef __LITTLE_ENDIAN__
   59268 __ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
   59269   poly8x16_t __ret;
   59270   __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 36);
   59271   return __ret;
   59272 }
   59273 #else
   59274 __ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
   59275   poly8x16x4_t __rev0;
   59276   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59277   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59278   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59279   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59280   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59281   poly8x16_t __ret;
   59282   __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 36);
   59283   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59284   return __ret;
   59285 }
   59286 #endif
   59287 
   59288 #ifdef __LITTLE_ENDIAN__
   59289 __ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
   59290   uint8x16_t __ret;
   59291   __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 48);
   59292   return __ret;
   59293 }
   59294 #else
   59295 __ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
   59296   uint8x16x4_t __rev0;
   59297   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59298   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59299   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59300   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59301   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59302   uint8x16_t __ret;
   59303   __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 48);
   59304   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59305   return __ret;
   59306 }
   59307 #endif
   59308 
   59309 #ifdef __LITTLE_ENDIAN__
   59310 __ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
   59311   int8x16_t __ret;
   59312   __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 32);
   59313   return __ret;
   59314 }
   59315 #else
   59316 __ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
   59317   int8x16x4_t __rev0;
   59318   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59319   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59320   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59321   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59322   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59323   int8x16_t __ret;
   59324   __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 32);
   59325   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59326   return __ret;
   59327 }
   59328 #endif
   59329 
   59330 #ifdef __LITTLE_ENDIAN__
   59331 __ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
   59332   uint8x8_t __ret;
   59333   __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 16);
   59334   return __ret;
   59335 }
   59336 #else
   59337 __ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
   59338   uint8x16x4_t __rev0;
   59339   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59340   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59341   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59342   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59343   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59344   uint8x8_t __ret;
   59345   __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 16);
   59346   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59347   return __ret;
   59348 }
   59349 #endif
   59350 
   59351 #ifdef __LITTLE_ENDIAN__
   59352 __ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
   59353   int8x8_t __ret;
   59354   __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 0);
   59355   return __ret;
   59356 }
   59357 #else
   59358 __ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
   59359   int8x16x4_t __rev0;
   59360   __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59361   __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59362   __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59363   __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59364   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59365   int8x8_t __ret;
   59366   __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 0);
   59367   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59368   return __ret;
   59369 }
   59370 #endif
   59371 
   59372 #ifdef __LITTLE_ENDIAN__
   59373 __ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
   59374   poly8x8_t __ret;
   59375   __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 4);
   59376   return __ret;
   59377 }
   59378 #else
   59379 __ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
   59380   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59381   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59382   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59383   poly8x8_t __ret;
   59384   __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 4);
   59385   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59386   return __ret;
   59387 }
   59388 #endif
   59389 
   59390 #ifdef __LITTLE_ENDIAN__
   59391 __ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
   59392   poly8x16_t __ret;
   59393   __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
   59394   return __ret;
   59395 }
   59396 #else
   59397 __ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
   59398   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59399   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59400   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59401   poly8x16_t __ret;
   59402   __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
   59403   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59404   return __ret;
   59405 }
   59406 #endif
   59407 
   59408 #ifdef __LITTLE_ENDIAN__
   59409 __ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   59410   uint8x16_t __ret;
   59411   __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
   59412   return __ret;
   59413 }
   59414 #else
   59415 __ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   59416   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59417   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59418   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59419   uint8x16_t __ret;
   59420   __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
   59421   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59422   return __ret;
   59423 }
   59424 #endif
   59425 
   59426 #ifdef __LITTLE_ENDIAN__
   59427 __ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   59428   int8x16_t __ret;
   59429   __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
   59430   return __ret;
   59431 }
   59432 #else
   59433 __ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   59434   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59435   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59436   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59437   int8x16_t __ret;
   59438   __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
   59439   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59440   return __ret;
   59441 }
   59442 #endif
   59443 
   59444 #ifdef __LITTLE_ENDIAN__
   59445 __ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
   59446   uint8x8_t __ret;
   59447   __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 16);
   59448   return __ret;
   59449 }
   59450 #else
   59451 __ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
   59452   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59453   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59454   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59455   uint8x8_t __ret;
   59456   __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 16);
   59457   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59458   return __ret;
   59459 }
   59460 #endif
   59461 
   59462 #ifdef __LITTLE_ENDIAN__
   59463 __ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
   59464   int8x8_t __ret;
   59465   __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 0);
   59466   return __ret;
   59467 }
   59468 #else
   59469 __ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
   59470   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59471   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59472   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59473   int8x8_t __ret;
   59474   __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 0);
   59475   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59476   return __ret;
   59477 }
   59478 #endif
   59479 
   59480 #ifdef __LITTLE_ENDIAN__
   59481 __ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
   59482   poly8x8_t __ret;
   59483   __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 4);
   59484   return __ret;
   59485 }
   59486 #else
   59487 __ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
   59488   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59489   poly8x16x2_t __rev1;
   59490   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59491   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59492   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59493   poly8x8_t __ret;
   59494   __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 4);
   59495   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59496   return __ret;
   59497 }
   59498 #endif
   59499 
   59500 #ifdef __LITTLE_ENDIAN__
   59501 __ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
   59502   poly8x16_t __ret;
   59503   __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 36);
   59504   return __ret;
   59505 }
   59506 #else
   59507 __ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
   59508   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59509   poly8x16x2_t __rev1;
   59510   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59511   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59512   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59513   poly8x16_t __ret;
   59514   __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 36);
   59515   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59516   return __ret;
   59517 }
   59518 #endif
   59519 
   59520 #ifdef __LITTLE_ENDIAN__
   59521 __ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
   59522   uint8x16_t __ret;
   59523   __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 48);
   59524   return __ret;
   59525 }
   59526 #else
   59527 __ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
   59528   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59529   uint8x16x2_t __rev1;
   59530   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59531   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59532   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59533   uint8x16_t __ret;
   59534   __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 48);
   59535   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59536   return __ret;
   59537 }
   59538 #endif
   59539 
   59540 #ifdef __LITTLE_ENDIAN__
   59541 __ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
   59542   int8x16_t __ret;
   59543   __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 32);
   59544   return __ret;
   59545 }
   59546 #else
   59547 __ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
   59548   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59549   int8x16x2_t __rev1;
   59550   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59551   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59552   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59553   int8x16_t __ret;
   59554   __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 32);
   59555   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59556   return __ret;
   59557 }
   59558 #endif
   59559 
   59560 #ifdef __LITTLE_ENDIAN__
   59561 __ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
   59562   uint8x8_t __ret;
   59563   __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 16);
   59564   return __ret;
   59565 }
   59566 #else
   59567 __ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
   59568   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59569   uint8x16x2_t __rev1;
   59570   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59571   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59572   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59573   uint8x8_t __ret;
   59574   __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 16);
   59575   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59576   return __ret;
   59577 }
   59578 #endif
   59579 
   59580 #ifdef __LITTLE_ENDIAN__
   59581 __ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
   59582   int8x8_t __ret;
   59583   __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 0);
   59584   return __ret;
   59585 }
   59586 #else
   59587 __ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
   59588   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59589   int8x16x2_t __rev1;
   59590   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59591   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59592   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59593   int8x8_t __ret;
   59594   __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 0);
   59595   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59596   return __ret;
   59597 }
   59598 #endif
   59599 
   59600 #ifdef __LITTLE_ENDIAN__
   59601 __ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
   59602   poly8x8_t __ret;
   59603   __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 4);
   59604   return __ret;
   59605 }
   59606 #else
   59607 __ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
   59608   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59609   poly8x16x3_t __rev1;
   59610   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59611   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59612   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59613   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59614   poly8x8_t __ret;
   59615   __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 4);
   59616   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59617   return __ret;
   59618 }
   59619 #endif
   59620 
   59621 #ifdef __LITTLE_ENDIAN__
   59622 __ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
   59623   poly8x16_t __ret;
   59624   __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 36);
   59625   return __ret;
   59626 }
   59627 #else
   59628 __ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
   59629   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59630   poly8x16x3_t __rev1;
   59631   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59632   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59633   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59634   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59635   poly8x16_t __ret;
   59636   __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 36);
   59637   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59638   return __ret;
   59639 }
   59640 #endif
   59641 
   59642 #ifdef __LITTLE_ENDIAN__
   59643 __ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
   59644   uint8x16_t __ret;
   59645   __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 48);
   59646   return __ret;
   59647 }
   59648 #else
   59649 __ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
   59650   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59651   uint8x16x3_t __rev1;
   59652   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59653   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59654   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59655   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59656   uint8x16_t __ret;
   59657   __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 48);
   59658   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59659   return __ret;
   59660 }
   59661 #endif
   59662 
   59663 #ifdef __LITTLE_ENDIAN__
   59664 __ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
   59665   int8x16_t __ret;
   59666   __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 32);
   59667   return __ret;
   59668 }
   59669 #else
   59670 __ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
   59671   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59672   int8x16x3_t __rev1;
   59673   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59674   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59675   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59676   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59677   int8x16_t __ret;
   59678   __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 32);
   59679   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59680   return __ret;
   59681 }
   59682 #endif
   59683 
   59684 #ifdef __LITTLE_ENDIAN__
   59685 __ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
   59686   uint8x8_t __ret;
   59687   __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 16);
   59688   return __ret;
   59689 }
   59690 #else
   59691 __ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
   59692   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59693   uint8x16x3_t __rev1;
   59694   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59695   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59696   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59697   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59698   uint8x8_t __ret;
   59699   __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 16);
   59700   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59701   return __ret;
   59702 }
   59703 #endif
   59704 
   59705 #ifdef __LITTLE_ENDIAN__
   59706 __ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
   59707   int8x8_t __ret;
   59708   __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 0);
   59709   return __ret;
   59710 }
   59711 #else
   59712 __ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
   59713   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59714   int8x16x3_t __rev1;
   59715   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59716   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59717   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59718   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59719   int8x8_t __ret;
   59720   __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 0);
   59721   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59722   return __ret;
   59723 }
   59724 #endif
   59725 
   59726 #ifdef __LITTLE_ENDIAN__
   59727 __ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
   59728   poly8x8_t __ret;
   59729   __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 4);
   59730   return __ret;
   59731 }
   59732 #else
   59733 __ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
   59734   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59735   poly8x16x4_t __rev1;
   59736   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59737   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59738   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59739   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59740   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59741   poly8x8_t __ret;
   59742   __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 4);
   59743   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59744   return __ret;
   59745 }
   59746 #endif
   59747 
   59748 #ifdef __LITTLE_ENDIAN__
   59749 __ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
   59750   poly8x16_t __ret;
   59751   __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 36);
   59752   return __ret;
   59753 }
   59754 #else
   59755 __ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
   59756   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59757   poly8x16x4_t __rev1;
   59758   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59759   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59760   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59761   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59762   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59763   poly8x16_t __ret;
   59764   __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 36);
   59765   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59766   return __ret;
   59767 }
   59768 #endif
   59769 
   59770 #ifdef __LITTLE_ENDIAN__
   59771 __ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
   59772   uint8x16_t __ret;
   59773   __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 48);
   59774   return __ret;
   59775 }
   59776 #else
   59777 __ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
   59778   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59779   uint8x16x4_t __rev1;
   59780   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59781   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59782   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59783   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59784   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59785   uint8x16_t __ret;
   59786   __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 48);
   59787   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59788   return __ret;
   59789 }
   59790 #endif
   59791 
   59792 #ifdef __LITTLE_ENDIAN__
   59793 __ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
   59794   int8x16_t __ret;
   59795   __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 32);
   59796   return __ret;
   59797 }
   59798 #else
   59799 __ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
   59800   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59801   int8x16x4_t __rev1;
   59802   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59803   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59804   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59805   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59806   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59807   int8x16_t __ret;
   59808   __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 32);
   59809   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59810   return __ret;
   59811 }
   59812 #endif
   59813 
   59814 #ifdef __LITTLE_ENDIAN__
   59815 __ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
   59816   uint8x8_t __ret;
   59817   __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 16);
   59818   return __ret;
   59819 }
   59820 #else
   59821 __ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
   59822   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59823   uint8x16x4_t __rev1;
   59824   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59825   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59826   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59827   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59828   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59829   uint8x8_t __ret;
   59830   __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 16);
   59831   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59832   return __ret;
   59833 }
   59834 #endif
   59835 
   59836 #ifdef __LITTLE_ENDIAN__
   59837 __ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
   59838   int8x8_t __ret;
   59839   __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 0);
   59840   return __ret;
   59841 }
   59842 #else
   59843 __ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
   59844   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59845   int8x16x4_t __rev1;
   59846   __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59847   __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59848   __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59849   __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59850   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59851   int8x8_t __ret;
   59852   __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 0);
   59853   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59854   return __ret;
   59855 }
   59856 #endif
   59857 
   59858 #ifdef __LITTLE_ENDIAN__
   59859 __ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   59860   uint16x8_t __ret;
   59861   __ret = vcombine_u16(__p0, vraddhn_u32(__p1, __p2));
   59862   return __ret;
   59863 }
   59864 #else
   59865 __ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   59866   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   59867   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   59868   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   59869   uint16x8_t __ret;
   59870   __ret = __noswap_vcombine_u16(__rev0, __noswap_vraddhn_u32(__rev1, __rev2));
   59871   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59872   return __ret;
   59873 }
   59874 #endif
   59875 
   59876 #ifdef __LITTLE_ENDIAN__
   59877 __ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   59878   uint32x4_t __ret;
   59879   __ret = vcombine_u32(__p0, vraddhn_u64(__p1, __p2));
   59880   return __ret;
   59881 }
   59882 #else
   59883 __ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   59884   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   59885   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   59886   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   59887   uint32x4_t __ret;
   59888   __ret = __noswap_vcombine_u32(__rev0, __noswap_vraddhn_u64(__rev1, __rev2));
   59889   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   59890   return __ret;
   59891 }
   59892 #endif
   59893 
   59894 #ifdef __LITTLE_ENDIAN__
   59895 __ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   59896   uint8x16_t __ret;
   59897   __ret = vcombine_u8(__p0, vraddhn_u16(__p1, __p2));
   59898   return __ret;
   59899 }
   59900 #else
   59901 __ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   59902   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59903   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59904   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59905   uint8x16_t __ret;
   59906   __ret = __noswap_vcombine_u8(__rev0, __noswap_vraddhn_u16(__rev1, __rev2));
   59907   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59908   return __ret;
   59909 }
   59910 #endif
   59911 
   59912 #ifdef __LITTLE_ENDIAN__
   59913 __ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   59914   int16x8_t __ret;
   59915   __ret = vcombine_s16(__p0, vraddhn_s32(__p1, __p2));
   59916   return __ret;
   59917 }
   59918 #else
   59919 __ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   59920   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   59921   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   59922   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   59923   int16x8_t __ret;
   59924   __ret = __noswap_vcombine_s16(__rev0, __noswap_vraddhn_s32(__rev1, __rev2));
   59925   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59926   return __ret;
   59927 }
   59928 #endif
   59929 
   59930 #ifdef __LITTLE_ENDIAN__
   59931 __ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   59932   int32x4_t __ret;
   59933   __ret = vcombine_s32(__p0, vraddhn_s64(__p1, __p2));
   59934   return __ret;
   59935 }
   59936 #else
   59937 __ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   59938   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   59939   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   59940   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   59941   int32x4_t __ret;
   59942   __ret = __noswap_vcombine_s32(__rev0, __noswap_vraddhn_s64(__rev1, __rev2));
   59943   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   59944   return __ret;
   59945 }
   59946 #endif
   59947 
   59948 #ifdef __LITTLE_ENDIAN__
   59949 __ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   59950   int8x16_t __ret;
   59951   __ret = vcombine_s8(__p0, vraddhn_s16(__p1, __p2));
   59952   return __ret;
   59953 }
   59954 #else
   59955 __ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   59956   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59957   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   59958   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   59959   int8x16_t __ret;
   59960   __ret = __noswap_vcombine_s8(__rev0, __noswap_vraddhn_s16(__rev1, __rev2));
   59961   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59962   return __ret;
   59963 }
   59964 #endif
   59965 
   59966 #ifdef __LITTLE_ENDIAN__
   59967 __ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
   59968   poly8x8_t __ret;
   59969   __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 4);
   59970   return __ret;
   59971 }
   59972 #else
   59973 __ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
   59974   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   59975   poly8x8_t __ret;
   59976   __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 4);
   59977   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   59978   return __ret;
   59979 }
   59980 #endif
   59981 
   59982 #ifdef __LITTLE_ENDIAN__
   59983 __ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
   59984   poly8x16_t __ret;
   59985   __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 36);
   59986   return __ret;
   59987 }
   59988 #else
   59989 __ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
   59990   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59991   poly8x16_t __ret;
   59992   __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 36);
   59993   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   59994   return __ret;
   59995 }
   59996 #endif
   59997 
   59998 #ifdef __LITTLE_ENDIAN__
   59999 __ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
   60000   uint8x16_t __ret;
   60001   __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 48);
   60002   return __ret;
   60003 }
   60004 #else
   60005 __ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
   60006   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   60007   uint8x16_t __ret;
   60008   __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 48);
   60009   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   60010   return __ret;
   60011 }
   60012 #endif
   60013 
   60014 #ifdef __LITTLE_ENDIAN__
   60015 __ai int8x16_t vrbitq_s8(int8x16_t __p0) {
   60016   int8x16_t __ret;
   60017   __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 32);
   60018   return __ret;
   60019 }
   60020 #else
   60021 __ai int8x16_t vrbitq_s8(int8x16_t __p0) {
   60022   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   60023   int8x16_t __ret;
   60024   __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 32);
   60025   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   60026   return __ret;
   60027 }
   60028 #endif
   60029 
   60030 #ifdef __LITTLE_ENDIAN__
   60031 __ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
   60032   uint8x8_t __ret;
   60033   __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 16);
   60034   return __ret;
   60035 }
   60036 #else
   60037 __ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
   60038   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   60039   uint8x8_t __ret;
   60040   __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 16);
   60041   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   60042   return __ret;
   60043 }
   60044 #endif
   60045 
   60046 #ifdef __LITTLE_ENDIAN__
   60047 __ai int8x8_t vrbit_s8(int8x8_t __p0) {
   60048   int8x8_t __ret;
   60049   __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 0);
   60050   return __ret;
   60051 }
   60052 #else
   60053 __ai int8x8_t vrbit_s8(int8x8_t __p0) {
   60054   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   60055   int8x8_t __ret;
   60056   __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 0);
   60057   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   60058   return __ret;
   60059 }
   60060 #endif
   60061 
   60062 #ifdef __LITTLE_ENDIAN__
   60063 __ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
   60064   float64x2_t __ret;
   60065   __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 42);
   60066   return __ret;
   60067 }
   60068 #else
   60069 __ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
   60070   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   60071   float64x2_t __ret;
   60072   __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 42);
   60073   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   60074   return __ret;
   60075 }
   60076 #endif
   60077 
   60078 #ifdef __LITTLE_ENDIAN__
   60079 __ai float64x1_t vrecpe_f64(float64x1_t __p0) {
   60080   float64x1_t __ret;
   60081   __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
   60082   return __ret;
   60083 }
   60084 #else
   60085 __ai float64x1_t vrecpe_f64(float64x1_t __p0) {
   60086   float64x1_t __ret;
   60087   __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
   60088   return __ret;
   60089 }
   60090 #endif
   60091 
   60092 #ifdef __LITTLE_ENDIAN__
   60093 __ai float64_t vrecped_f64(float64_t __p0) {
   60094   float64_t __ret;
   60095   __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
   60096   return __ret;
   60097 }
   60098 #else
   60099 __ai float64_t vrecped_f64(float64_t __p0) {
   60100   float64_t __ret;
   60101   __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
   60102   return __ret;
   60103 }
   60104 #endif
   60105 
   60106 #ifdef __LITTLE_ENDIAN__
   60107 __ai float32_t vrecpes_f32(float32_t __p0) {
   60108   float32_t __ret;
   60109   __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
   60110   return __ret;
   60111 }
   60112 #else
   60113 __ai float32_t vrecpes_f32(float32_t __p0) {
   60114   float32_t __ret;
   60115   __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
   60116   return __ret;
   60117 }
   60118 #endif
   60119 
   60120 #ifdef __LITTLE_ENDIAN__
   60121 __ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
   60122   float64x2_t __ret;
   60123   __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   60124   return __ret;
   60125 }
   60126 #else
   60127 __ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
   60128   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   60129   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   60130   float64x2_t __ret;
   60131   __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   60132   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   60133   return __ret;
   60134 }
   60135 #endif
   60136 
   60137 #ifdef __LITTLE_ENDIAN__
   60138 __ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
   60139   float64x1_t __ret;
   60140   __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   60141   return __ret;
   60142 }
   60143 #else
   60144 __ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
   60145   float64x1_t __ret;
   60146   __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   60147   return __ret;
   60148 }
   60149 #endif
   60150 
   60151 #ifdef __LITTLE_ENDIAN__
   60152 __ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
   60153   float64_t __ret;
   60154   __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
   60155   return __ret;
   60156 }
   60157 #else
   60158 __ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
   60159   float64_t __ret;
   60160   __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
   60161   return __ret;
   60162 }
   60163 #endif
   60164 
   60165 #ifdef __LITTLE_ENDIAN__
   60166 __ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
   60167   float32_t __ret;
   60168   __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
   60169   return __ret;
   60170 }
   60171 #else
   60172 __ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
   60173   float32_t __ret;
   60174   __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
   60175   return __ret;
   60176 }
   60177 #endif
   60178 
   60179 #ifdef __LITTLE_ENDIAN__
   60180 __ai float64_t vrecpxd_f64(float64_t __p0) {
   60181   float64_t __ret;
   60182   __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
   60183   return __ret;
   60184 }
   60185 #else
   60186 __ai float64_t vrecpxd_f64(float64_t __p0) {
   60187   float64_t __ret;
   60188   __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
   60189   return __ret;
   60190 }
   60191 #endif
   60192 
   60193 #ifdef __LITTLE_ENDIAN__
   60194 __ai float32_t vrecpxs_f32(float32_t __p0) {
   60195   float32_t __ret;
   60196   __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
   60197   return __ret;
   60198 }
   60199 #else
   60200 __ai float32_t vrecpxs_f32(float32_t __p0) {
   60201   float32_t __ret;
   60202   __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
   60203   return __ret;
   60204 }
   60205 #endif
   60206 
   60207 #ifdef __LITTLE_ENDIAN__
   60208 __ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
   60209   uint64_t __ret;
   60210   __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
   60211   return __ret;
   60212 }
   60213 #else
   60214 __ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
   60215   uint64_t __ret;
   60216   __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
   60217   return __ret;
   60218 }
   60219 #endif
   60220 
   60221 #ifdef __LITTLE_ENDIAN__
   60222 __ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
   60223   int64_t __ret;
   60224   __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
   60225   return __ret;
   60226 }
   60227 #else
   60228 __ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
   60229   int64_t __ret;
   60230   __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
   60231   return __ret;
   60232 }
   60233 #endif
   60234 
   60235 #ifdef __LITTLE_ENDIAN__
   60236 #define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
   60237   uint64_t __s0 = __p0; \
   60238   uint64_t __ret; \
   60239   __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
   60240   __ret; \
   60241 })
   60242 #else
   60243 #define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
   60244   uint64_t __s0 = __p0; \
   60245   uint64_t __ret; \
   60246   __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
   60247   __ret; \
   60248 })
   60249 #endif
   60250 
   60251 #ifdef __LITTLE_ENDIAN__
   60252 #define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
   60253   int64_t __s0 = __p0; \
   60254   int64_t __ret; \
   60255   __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
   60256   __ret; \
   60257 })
   60258 #else
   60259 #define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
   60260   int64_t __s0 = __p0; \
   60261   int64_t __ret; \
   60262   __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
   60263   __ret; \
   60264 })
   60265 #endif
   60266 
   60267 #ifdef __LITTLE_ENDIAN__
   60268 #define vrshrn_high_n_u32(__p0_206, __p1_206, __p2_206) __extension__ ({ \
   60269   uint16x4_t __s0_206 = __p0_206; \
   60270   uint32x4_t __s1_206 = __p1_206; \
   60271   uint16x8_t __ret_206; \
   60272   __ret_206 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_206), (uint16x4_t)(vrshrn_n_u32(__s1_206, __p2_206)))); \
   60273   __ret_206; \
   60274 })
   60275 #else
   60276 #define vrshrn_high_n_u32(__p0_207, __p1_207, __p2_207) __extension__ ({ \
   60277   uint16x4_t __s0_207 = __p0_207; \
   60278   uint32x4_t __s1_207 = __p1_207; \
   60279   uint16x4_t __rev0_207;  __rev0_207 = __builtin_shufflevector(__s0_207, __s0_207, 3, 2, 1, 0); \
   60280   uint32x4_t __rev1_207;  __rev1_207 = __builtin_shufflevector(__s1_207, __s1_207, 3, 2, 1, 0); \
   60281   uint16x8_t __ret_207; \
   60282   __ret_207 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_207), (uint16x4_t)(__noswap_vrshrn_n_u32(__rev1_207, __p2_207)))); \
   60283   __ret_207 = __builtin_shufflevector(__ret_207, __ret_207, 7, 6, 5, 4, 3, 2, 1, 0); \
   60284   __ret_207; \
   60285 })
   60286 #endif
   60287 
   60288 #ifdef __LITTLE_ENDIAN__
   60289 #define vrshrn_high_n_u64(__p0_208, __p1_208, __p2_208) __extension__ ({ \
   60290   uint32x2_t __s0_208 = __p0_208; \
   60291   uint64x2_t __s1_208 = __p1_208; \
   60292   uint32x4_t __ret_208; \
   60293   __ret_208 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_208), (uint32x2_t)(vrshrn_n_u64(__s1_208, __p2_208)))); \
   60294   __ret_208; \
   60295 })
   60296 #else
   60297 #define vrshrn_high_n_u64(__p0_209, __p1_209, __p2_209) __extension__ ({ \
   60298   uint32x2_t __s0_209 = __p0_209; \
   60299   uint64x2_t __s1_209 = __p1_209; \
   60300   uint32x2_t __rev0_209;  __rev0_209 = __builtin_shufflevector(__s0_209, __s0_209, 1, 0); \
   60301   uint64x2_t __rev1_209;  __rev1_209 = __builtin_shufflevector(__s1_209, __s1_209, 1, 0); \
   60302   uint32x4_t __ret_209; \
   60303   __ret_209 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_209), (uint32x2_t)(__noswap_vrshrn_n_u64(__rev1_209, __p2_209)))); \
   60304   __ret_209 = __builtin_shufflevector(__ret_209, __ret_209, 3, 2, 1, 0); \
   60305   __ret_209; \
   60306 })
   60307 #endif
   60308 
   60309 #ifdef __LITTLE_ENDIAN__
   60310 #define vrshrn_high_n_u16(__p0_210, __p1_210, __p2_210) __extension__ ({ \
   60311   uint8x8_t __s0_210 = __p0_210; \
   60312   uint16x8_t __s1_210 = __p1_210; \
   60313   uint8x16_t __ret_210; \
   60314   __ret_210 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_210), (uint8x8_t)(vrshrn_n_u16(__s1_210, __p2_210)))); \
   60315   __ret_210; \
   60316 })
   60317 #else
   60318 #define vrshrn_high_n_u16(__p0_211, __p1_211, __p2_211) __extension__ ({ \
   60319   uint8x8_t __s0_211 = __p0_211; \
   60320   uint16x8_t __s1_211 = __p1_211; \
   60321   uint8x8_t __rev0_211;  __rev0_211 = __builtin_shufflevector(__s0_211, __s0_211, 7, 6, 5, 4, 3, 2, 1, 0); \
   60322   uint16x8_t __rev1_211;  __rev1_211 = __builtin_shufflevector(__s1_211, __s1_211, 7, 6, 5, 4, 3, 2, 1, 0); \
   60323   uint8x16_t __ret_211; \
   60324   __ret_211 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_211), (uint8x8_t)(__noswap_vrshrn_n_u16(__rev1_211, __p2_211)))); \
   60325   __ret_211 = __builtin_shufflevector(__ret_211, __ret_211, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   60326   __ret_211; \
   60327 })
   60328 #endif
   60329 
   60330 #ifdef __LITTLE_ENDIAN__
   60331 #define vrshrn_high_n_s32(__p0_212, __p1_212, __p2_212) __extension__ ({ \
   60332   int16x4_t __s0_212 = __p0_212; \
   60333   int32x4_t __s1_212 = __p1_212; \
   60334   int16x8_t __ret_212; \
   60335   __ret_212 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_212), (int16x4_t)(vrshrn_n_s32(__s1_212, __p2_212)))); \
   60336   __ret_212; \
   60337 })
   60338 #else
   60339 #define vrshrn_high_n_s32(__p0_213, __p1_213, __p2_213) __extension__ ({ \
   60340   int16x4_t __s0_213 = __p0_213; \
   60341   int32x4_t __s1_213 = __p1_213; \
   60342   int16x4_t __rev0_213;  __rev0_213 = __builtin_shufflevector(__s0_213, __s0_213, 3, 2, 1, 0); \
   60343   int32x4_t __rev1_213;  __rev1_213 = __builtin_shufflevector(__s1_213, __s1_213, 3, 2, 1, 0); \
   60344   int16x8_t __ret_213; \
   60345   __ret_213 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_213), (int16x4_t)(__noswap_vrshrn_n_s32(__rev1_213, __p2_213)))); \
   60346   __ret_213 = __builtin_shufflevector(__ret_213, __ret_213, 7, 6, 5, 4, 3, 2, 1, 0); \
   60347   __ret_213; \
   60348 })
   60349 #endif
   60350 
   60351 #ifdef __LITTLE_ENDIAN__
   60352 #define vrshrn_high_n_s64(__p0_214, __p1_214, __p2_214) __extension__ ({ \
   60353   int32x2_t __s0_214 = __p0_214; \
   60354   int64x2_t __s1_214 = __p1_214; \
   60355   int32x4_t __ret_214; \
   60356   __ret_214 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_214), (int32x2_t)(vrshrn_n_s64(__s1_214, __p2_214)))); \
   60357   __ret_214; \
   60358 })
   60359 #else
   60360 #define vrshrn_high_n_s64(__p0_215, __p1_215, __p2_215) __extension__ ({ \
   60361   int32x2_t __s0_215 = __p0_215; \
   60362   int64x2_t __s1_215 = __p1_215; \
   60363   int32x2_t __rev0_215;  __rev0_215 = __builtin_shufflevector(__s0_215, __s0_215, 1, 0); \
   60364   int64x2_t __rev1_215;  __rev1_215 = __builtin_shufflevector(__s1_215, __s1_215, 1, 0); \
   60365   int32x4_t __ret_215; \
   60366   __ret_215 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_215), (int32x2_t)(__noswap_vrshrn_n_s64(__rev1_215, __p2_215)))); \
   60367   __ret_215 = __builtin_shufflevector(__ret_215, __ret_215, 3, 2, 1, 0); \
   60368   __ret_215; \
   60369 })
   60370 #endif
   60371 
   60372 #ifdef __LITTLE_ENDIAN__
   60373 #define vrshrn_high_n_s16(__p0_216, __p1_216, __p2_216) __extension__ ({ \
   60374   int8x8_t __s0_216 = __p0_216; \
   60375   int16x8_t __s1_216 = __p1_216; \
   60376   int8x16_t __ret_216; \
   60377   __ret_216 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_216), (int8x8_t)(vrshrn_n_s16(__s1_216, __p2_216)))); \
   60378   __ret_216; \
   60379 })
   60380 #else
   60381 #define vrshrn_high_n_s16(__p0_217, __p1_217, __p2_217) __extension__ ({ \
   60382   int8x8_t __s0_217 = __p0_217; \
   60383   int16x8_t __s1_217 = __p1_217; \
   60384   int8x8_t __rev0_217;  __rev0_217 = __builtin_shufflevector(__s0_217, __s0_217, 7, 6, 5, 4, 3, 2, 1, 0); \
   60385   int16x8_t __rev1_217;  __rev1_217 = __builtin_shufflevector(__s1_217, __s1_217, 7, 6, 5, 4, 3, 2, 1, 0); \
   60386   int8x16_t __ret_217; \
   60387   __ret_217 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_217), (int8x8_t)(__noswap_vrshrn_n_s16(__rev1_217, __p2_217)))); \
   60388   __ret_217 = __builtin_shufflevector(__ret_217, __ret_217, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   60389   __ret_217; \
   60390 })
   60391 #endif
   60392 
   60393 #ifdef __LITTLE_ENDIAN__
   60394 __ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
   60395   float64x2_t __ret;
   60396   __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 42);
   60397   return __ret;
   60398 }
   60399 #else
   60400 __ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
   60401   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   60402   float64x2_t __ret;
   60403   __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 42);
   60404   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   60405   return __ret;
   60406 }
   60407 #endif
   60408 
   60409 #ifdef __LITTLE_ENDIAN__
   60410 __ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
   60411   float64x1_t __ret;
   60412   __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
   60413   return __ret;
   60414 }
   60415 #else
   60416 __ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
   60417   float64x1_t __ret;
   60418   __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
   60419   return __ret;
   60420 }
   60421 #endif
   60422 
   60423 #ifdef __LITTLE_ENDIAN__
   60424 __ai float64_t vrsqrted_f64(float64_t __p0) {
   60425   float64_t __ret;
   60426   __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
   60427   return __ret;
   60428 }
   60429 #else
   60430 __ai float64_t vrsqrted_f64(float64_t __p0) {
   60431   float64_t __ret;
   60432   __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
   60433   return __ret;
   60434 }
   60435 #endif
   60436 
   60437 #ifdef __LITTLE_ENDIAN__
   60438 __ai float32_t vrsqrtes_f32(float32_t __p0) {
   60439   float32_t __ret;
   60440   __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
   60441   return __ret;
   60442 }
   60443 #else
   60444 __ai float32_t vrsqrtes_f32(float32_t __p0) {
   60445   float32_t __ret;
   60446   __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
   60447   return __ret;
   60448 }
   60449 #endif
   60450 
   60451 #ifdef __LITTLE_ENDIAN__
   60452 __ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
   60453   float64x2_t __ret;
   60454   __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
   60455   return __ret;
   60456 }
   60457 #else
   60458 __ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
   60459   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   60460   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   60461   float64x2_t __ret;
   60462   __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
   60463   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   60464   return __ret;
   60465 }
   60466 #endif
   60467 
   60468 #ifdef __LITTLE_ENDIAN__
   60469 __ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
   60470   float64x1_t __ret;
   60471   __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   60472   return __ret;
   60473 }
   60474 #else
   60475 __ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
   60476   float64x1_t __ret;
   60477   __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
   60478   return __ret;
   60479 }
   60480 #endif
   60481 
   60482 #ifdef __LITTLE_ENDIAN__
   60483 __ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
   60484   float64_t __ret;
   60485   __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
   60486   return __ret;
   60487 }
   60488 #else
   60489 __ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
   60490   float64_t __ret;
   60491   __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
   60492   return __ret;
   60493 }
   60494 #endif
   60495 
   60496 #ifdef __LITTLE_ENDIAN__
   60497 __ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
   60498   float32_t __ret;
   60499   __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
   60500   return __ret;
   60501 }
   60502 #else
   60503 __ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
   60504   float32_t __ret;
   60505   __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
   60506   return __ret;
   60507 }
   60508 #endif
   60509 
   60510 #ifdef __LITTLE_ENDIAN__
   60511 #define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
   60512   uint64_t __s0 = __p0; \
   60513   uint64_t __s1 = __p1; \
   60514   uint64_t __ret; \
   60515   __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
   60516   __ret; \
   60517 })
   60518 #else
   60519 #define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
   60520   uint64_t __s0 = __p0; \
   60521   uint64_t __s1 = __p1; \
   60522   uint64_t __ret; \
   60523   __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
   60524   __ret; \
   60525 })
   60526 #endif
   60527 
   60528 #ifdef __LITTLE_ENDIAN__
   60529 #define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
   60530   int64_t __s0 = __p0; \
   60531   int64_t __s1 = __p1; \
   60532   int64_t __ret; \
   60533   __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
   60534   __ret; \
   60535 })
   60536 #else
   60537 #define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
   60538   int64_t __s0 = __p0; \
   60539   int64_t __s1 = __p1; \
   60540   int64_t __ret; \
   60541   __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
   60542   __ret; \
   60543 })
   60544 #endif
   60545 
   60546 #ifdef __LITTLE_ENDIAN__
   60547 __ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   60548   uint16x8_t __ret;
   60549   __ret = vcombine_u16(__p0, vrsubhn_u32(__p1, __p2));
   60550   return __ret;
   60551 }
   60552 #else
   60553 __ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   60554   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   60555   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   60556   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   60557   uint16x8_t __ret;
   60558   __ret = __noswap_vcombine_u16(__rev0, __noswap_vrsubhn_u32(__rev1, __rev2));
   60559   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   60560   return __ret;
   60561 }
   60562 #endif
   60563 
   60564 #ifdef __LITTLE_ENDIAN__
   60565 __ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   60566   uint32x4_t __ret;
   60567   __ret = vcombine_u32(__p0, vrsubhn_u64(__p1, __p2));
   60568   return __ret;
   60569 }
   60570 #else
   60571 __ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   60572   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   60573   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   60574   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   60575   uint32x4_t __ret;
   60576   __ret = __noswap_vcombine_u32(__rev0, __noswap_vrsubhn_u64(__rev1, __rev2));
   60577   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   60578   return __ret;
   60579 }
   60580 #endif
   60581 
   60582 #ifdef __LITTLE_ENDIAN__
   60583 __ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   60584   uint8x16_t __ret;
   60585   __ret = vcombine_u8(__p0, vrsubhn_u16(__p1, __p2));
   60586   return __ret;
   60587 }
   60588 #else
   60589 __ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   60590   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   60591   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   60592   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   60593   uint8x16_t __ret;
   60594   __ret = __noswap_vcombine_u8(__rev0, __noswap_vrsubhn_u16(__rev1, __rev2));
   60595   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   60596   return __ret;
   60597 }
   60598 #endif
   60599 
   60600 #ifdef __LITTLE_ENDIAN__
   60601 __ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   60602   int16x8_t __ret;
   60603   __ret = vcombine_s16(__p0, vrsubhn_s32(__p1, __p2));
   60604   return __ret;
   60605 }
   60606 #else
   60607 __ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   60608   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   60609   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   60610   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   60611   int16x8_t __ret;
   60612   __ret = __noswap_vcombine_s16(__rev0, __noswap_vrsubhn_s32(__rev1, __rev2));
   60613   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   60614   return __ret;
   60615 }
   60616 #endif
   60617 
   60618 #ifdef __LITTLE_ENDIAN__
   60619 __ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   60620   int32x4_t __ret;
   60621   __ret = vcombine_s32(__p0, vrsubhn_s64(__p1, __p2));
   60622   return __ret;
   60623 }
   60624 #else
   60625 __ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   60626   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   60627   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   60628   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   60629   int32x4_t __ret;
   60630   __ret = __noswap_vcombine_s32(__rev0, __noswap_vrsubhn_s64(__rev1, __rev2));
   60631   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   60632   return __ret;
   60633 }
   60634 #endif
   60635 
   60636 #ifdef __LITTLE_ENDIAN__
   60637 __ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   60638   int8x16_t __ret;
   60639   __ret = vcombine_s8(__p0, vrsubhn_s16(__p1, __p2));
   60640   return __ret;
   60641 }
   60642 #else
   60643 __ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   60644   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   60645   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   60646   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   60647   int8x16_t __ret;
   60648   __ret = __noswap_vcombine_s8(__rev0, __noswap_vrsubhn_s16(__rev1, __rev2));
   60649   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   60650   return __ret;
   60651 }
   60652 #endif
   60653 
   60654 #ifdef __LITTLE_ENDIAN__
   60655 #define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   60656   poly64_t __s0 = __p0; \
   60657   poly64x1_t __s1 = __p1; \
   60658   poly64x1_t __ret; \
   60659   __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
   60660   __ret; \
   60661 })
   60662 #else
   60663 #define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   60664   poly64_t __s0 = __p0; \
   60665   poly64x1_t __s1 = __p1; \
   60666   poly64x1_t __ret; \
   60667   __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
   60668   __ret; \
   60669 })
   60670 #define __noswap_vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   60671   poly64_t __s0 = __p0; \
   60672   poly64x1_t __s1 = __p1; \
   60673   poly64x1_t __ret; \
   60674   __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
   60675   __ret; \
   60676 })
   60677 #endif
   60678 
   60679 #ifdef __LITTLE_ENDIAN__
   60680 #define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   60681   poly64_t __s0 = __p0; \
   60682   poly64x2_t __s1 = __p1; \
   60683   poly64x2_t __ret; \
   60684   __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
   60685   __ret; \
   60686 })
   60687 #else
   60688 #define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   60689   poly64_t __s0 = __p0; \
   60690   poly64x2_t __s1 = __p1; \
   60691   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   60692   poly64x2_t __ret; \
   60693   __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
   60694   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   60695   __ret; \
   60696 })
   60697 #define __noswap_vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   60698   poly64_t __s0 = __p0; \
   60699   poly64x2_t __s1 = __p1; \
   60700   poly64x2_t __ret; \
   60701   __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
   60702   __ret; \
   60703 })
   60704 #endif
   60705 
   60706 #ifdef __LITTLE_ENDIAN__
   60707 #define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   60708   float64_t __s0 = __p0; \
   60709   float64x2_t __s1 = __p1; \
   60710   float64x2_t __ret; \
   60711   __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
   60712   __ret; \
   60713 })
   60714 #else
   60715 #define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   60716   float64_t __s0 = __p0; \
   60717   float64x2_t __s1 = __p1; \
   60718   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   60719   float64x2_t __ret; \
   60720   __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__rev1, __p2); \
   60721   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   60722   __ret; \
   60723 })
   60724 #define __noswap_vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   60725   float64_t __s0 = __p0; \
   60726   float64x2_t __s1 = __p1; \
   60727   float64x2_t __ret; \
   60728   __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
   60729   __ret; \
   60730 })
   60731 #endif
   60732 
   60733 #ifdef __LITTLE_ENDIAN__
   60734 #define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   60735   float64_t __s0 = __p0; \
   60736   float64x1_t __s1 = __p1; \
   60737   float64x1_t __ret; \
   60738   __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
   60739   __ret; \
   60740 })
   60741 #else
   60742 #define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   60743   float64_t __s0 = __p0; \
   60744   float64x1_t __s1 = __p1; \
   60745   float64x1_t __ret; \
   60746   __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
   60747   __ret; \
   60748 })
   60749 #define __noswap_vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   60750   float64_t __s0 = __p0; \
   60751   float64x1_t __s1 = __p1; \
   60752   float64x1_t __ret; \
   60753   __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
   60754   __ret; \
   60755 })
   60756 #endif
   60757 
   60758 #ifdef __LITTLE_ENDIAN__
   60759 __ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
   60760   uint64_t __ret;
   60761   __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
   60762   return __ret;
   60763 }
   60764 #else
   60765 __ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
   60766   uint64_t __ret;
   60767   __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
   60768   return __ret;
   60769 }
   60770 #endif
   60771 
   60772 #ifdef __LITTLE_ENDIAN__
   60773 __ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
   60774   int64_t __ret;
   60775   __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
   60776   return __ret;
   60777 }
   60778 #else
   60779 __ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
   60780   int64_t __ret;
   60781   __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
   60782   return __ret;
   60783 }
   60784 #endif
   60785 
   60786 #ifdef __LITTLE_ENDIAN__
   60787 #define vshld_n_u64(__p0, __p1) __extension__ ({ \
   60788   uint64_t __s0 = __p0; \
   60789   uint64_t __ret; \
   60790   __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
   60791   __ret; \
   60792 })
   60793 #else
   60794 #define vshld_n_u64(__p0, __p1) __extension__ ({ \
   60795   uint64_t __s0 = __p0; \
   60796   uint64_t __ret; \
   60797   __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
   60798   __ret; \
   60799 })
   60800 #endif
   60801 
   60802 #ifdef __LITTLE_ENDIAN__
   60803 #define vshld_n_s64(__p0, __p1) __extension__ ({ \
   60804   int64_t __s0 = __p0; \
   60805   int64_t __ret; \
   60806   __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
   60807   __ret; \
   60808 })
   60809 #else
   60810 #define vshld_n_s64(__p0, __p1) __extension__ ({ \
   60811   int64_t __s0 = __p0; \
   60812   int64_t __ret; \
   60813   __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
   60814   __ret; \
   60815 })
   60816 #endif
   60817 
   60818 #ifdef __LITTLE_ENDIAN__
   60819 #define vshll_high_n_u8(__p0_218, __p1_218) __extension__ ({ \
   60820   uint8x16_t __s0_218 = __p0_218; \
   60821   uint16x8_t __ret_218; \
   60822   __ret_218 = (uint16x8_t)(vshll_n_u8(vget_high_u8(__s0_218), __p1_218)); \
   60823   __ret_218; \
   60824 })
   60825 #else
   60826 #define vshll_high_n_u8(__p0_219, __p1_219) __extension__ ({ \
   60827   uint8x16_t __s0_219 = __p0_219; \
   60828   uint8x16_t __rev0_219;  __rev0_219 = __builtin_shufflevector(__s0_219, __s0_219, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   60829   uint16x8_t __ret_219; \
   60830   __ret_219 = (uint16x8_t)(__noswap_vshll_n_u8(__noswap_vget_high_u8(__rev0_219), __p1_219)); \
   60831   __ret_219 = __builtin_shufflevector(__ret_219, __ret_219, 7, 6, 5, 4, 3, 2, 1, 0); \
   60832   __ret_219; \
   60833 })
   60834 #endif
   60835 
   60836 #ifdef __LITTLE_ENDIAN__
   60837 #define vshll_high_n_u32(__p0_220, __p1_220) __extension__ ({ \
   60838   uint32x4_t __s0_220 = __p0_220; \
   60839   uint64x2_t __ret_220; \
   60840   __ret_220 = (uint64x2_t)(vshll_n_u32(vget_high_u32(__s0_220), __p1_220)); \
   60841   __ret_220; \
   60842 })
   60843 #else
   60844 #define vshll_high_n_u32(__p0_221, __p1_221) __extension__ ({ \
   60845   uint32x4_t __s0_221 = __p0_221; \
   60846   uint32x4_t __rev0_221;  __rev0_221 = __builtin_shufflevector(__s0_221, __s0_221, 3, 2, 1, 0); \
   60847   uint64x2_t __ret_221; \
   60848   __ret_221 = (uint64x2_t)(__noswap_vshll_n_u32(__noswap_vget_high_u32(__rev0_221), __p1_221)); \
   60849   __ret_221 = __builtin_shufflevector(__ret_221, __ret_221, 1, 0); \
   60850   __ret_221; \
   60851 })
   60852 #endif
   60853 
   60854 #ifdef __LITTLE_ENDIAN__
   60855 #define vshll_high_n_u16(__p0_222, __p1_222) __extension__ ({ \
   60856   uint16x8_t __s0_222 = __p0_222; \
   60857   uint32x4_t __ret_222; \
   60858   __ret_222 = (uint32x4_t)(vshll_n_u16(vget_high_u16(__s0_222), __p1_222)); \
   60859   __ret_222; \
   60860 })
   60861 #else
   60862 #define vshll_high_n_u16(__p0_223, __p1_223) __extension__ ({ \
   60863   uint16x8_t __s0_223 = __p0_223; \
   60864   uint16x8_t __rev0_223;  __rev0_223 = __builtin_shufflevector(__s0_223, __s0_223, 7, 6, 5, 4, 3, 2, 1, 0); \
   60865   uint32x4_t __ret_223; \
   60866   __ret_223 = (uint32x4_t)(__noswap_vshll_n_u16(__noswap_vget_high_u16(__rev0_223), __p1_223)); \
   60867   __ret_223 = __builtin_shufflevector(__ret_223, __ret_223, 3, 2, 1, 0); \
   60868   __ret_223; \
   60869 })
   60870 #endif
   60871 
   60872 #ifdef __LITTLE_ENDIAN__
   60873 #define vshll_high_n_s8(__p0_224, __p1_224) __extension__ ({ \
   60874   int8x16_t __s0_224 = __p0_224; \
   60875   int16x8_t __ret_224; \
   60876   __ret_224 = (int16x8_t)(vshll_n_s8(vget_high_s8(__s0_224), __p1_224)); \
   60877   __ret_224; \
   60878 })
   60879 #else
   60880 #define vshll_high_n_s8(__p0_225, __p1_225) __extension__ ({ \
   60881   int8x16_t __s0_225 = __p0_225; \
   60882   int8x16_t __rev0_225;  __rev0_225 = __builtin_shufflevector(__s0_225, __s0_225, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   60883   int16x8_t __ret_225; \
   60884   __ret_225 = (int16x8_t)(__noswap_vshll_n_s8(__noswap_vget_high_s8(__rev0_225), __p1_225)); \
   60885   __ret_225 = __builtin_shufflevector(__ret_225, __ret_225, 7, 6, 5, 4, 3, 2, 1, 0); \
   60886   __ret_225; \
   60887 })
   60888 #endif
   60889 
   60890 #ifdef __LITTLE_ENDIAN__
   60891 #define vshll_high_n_s32(__p0_226, __p1_226) __extension__ ({ \
   60892   int32x4_t __s0_226 = __p0_226; \
   60893   int64x2_t __ret_226; \
   60894   __ret_226 = (int64x2_t)(vshll_n_s32(vget_high_s32(__s0_226), __p1_226)); \
   60895   __ret_226; \
   60896 })
   60897 #else
   60898 #define vshll_high_n_s32(__p0_227, __p1_227) __extension__ ({ \
   60899   int32x4_t __s0_227 = __p0_227; \
   60900   int32x4_t __rev0_227;  __rev0_227 = __builtin_shufflevector(__s0_227, __s0_227, 3, 2, 1, 0); \
   60901   int64x2_t __ret_227; \
   60902   __ret_227 = (int64x2_t)(__noswap_vshll_n_s32(__noswap_vget_high_s32(__rev0_227), __p1_227)); \
   60903   __ret_227 = __builtin_shufflevector(__ret_227, __ret_227, 1, 0); \
   60904   __ret_227; \
   60905 })
   60906 #endif
   60907 
   60908 #ifdef __LITTLE_ENDIAN__
   60909 #define vshll_high_n_s16(__p0_228, __p1_228) __extension__ ({ \
   60910   int16x8_t __s0_228 = __p0_228; \
   60911   int32x4_t __ret_228; \
   60912   __ret_228 = (int32x4_t)(vshll_n_s16(vget_high_s16(__s0_228), __p1_228)); \
   60913   __ret_228; \
   60914 })
   60915 #else
   60916 #define vshll_high_n_s16(__p0_229, __p1_229) __extension__ ({ \
   60917   int16x8_t __s0_229 = __p0_229; \
   60918   int16x8_t __rev0_229;  __rev0_229 = __builtin_shufflevector(__s0_229, __s0_229, 7, 6, 5, 4, 3, 2, 1, 0); \
   60919   int32x4_t __ret_229; \
   60920   __ret_229 = (int32x4_t)(__noswap_vshll_n_s16(__noswap_vget_high_s16(__rev0_229), __p1_229)); \
   60921   __ret_229 = __builtin_shufflevector(__ret_229, __ret_229, 3, 2, 1, 0); \
   60922   __ret_229; \
   60923 })
   60924 #endif
   60925 
   60926 #ifdef __LITTLE_ENDIAN__
   60927 #define vshrd_n_u64(__p0, __p1) __extension__ ({ \
   60928   uint64_t __s0 = __p0; \
   60929   uint64_t __ret; \
   60930   __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
   60931   __ret; \
   60932 })
   60933 #else
   60934 #define vshrd_n_u64(__p0, __p1) __extension__ ({ \
   60935   uint64_t __s0 = __p0; \
   60936   uint64_t __ret; \
   60937   __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
   60938   __ret; \
   60939 })
   60940 #endif
   60941 
   60942 #ifdef __LITTLE_ENDIAN__
   60943 #define vshrd_n_s64(__p0, __p1) __extension__ ({ \
   60944   int64_t __s0 = __p0; \
   60945   int64_t __ret; \
   60946   __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
   60947   __ret; \
   60948 })
   60949 #else
   60950 #define vshrd_n_s64(__p0, __p1) __extension__ ({ \
   60951   int64_t __s0 = __p0; \
   60952   int64_t __ret; \
   60953   __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
   60954   __ret; \
   60955 })
   60956 #endif
   60957 
   60958 #ifdef __LITTLE_ENDIAN__
   60959 #define vshrn_high_n_u32(__p0_230, __p1_230, __p2_230) __extension__ ({ \
   60960   uint16x4_t __s0_230 = __p0_230; \
   60961   uint32x4_t __s1_230 = __p1_230; \
   60962   uint16x8_t __ret_230; \
   60963   __ret_230 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_230), (uint16x4_t)(vshrn_n_u32(__s1_230, __p2_230)))); \
   60964   __ret_230; \
   60965 })
   60966 #else
   60967 #define vshrn_high_n_u32(__p0_231, __p1_231, __p2_231) __extension__ ({ \
   60968   uint16x4_t __s0_231 = __p0_231; \
   60969   uint32x4_t __s1_231 = __p1_231; \
   60970   uint16x4_t __rev0_231;  __rev0_231 = __builtin_shufflevector(__s0_231, __s0_231, 3, 2, 1, 0); \
   60971   uint32x4_t __rev1_231;  __rev1_231 = __builtin_shufflevector(__s1_231, __s1_231, 3, 2, 1, 0); \
   60972   uint16x8_t __ret_231; \
   60973   __ret_231 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_231), (uint16x4_t)(__noswap_vshrn_n_u32(__rev1_231, __p2_231)))); \
   60974   __ret_231 = __builtin_shufflevector(__ret_231, __ret_231, 7, 6, 5, 4, 3, 2, 1, 0); \
   60975   __ret_231; \
   60976 })
   60977 #endif
   60978 
   60979 #ifdef __LITTLE_ENDIAN__
   60980 #define vshrn_high_n_u64(__p0_232, __p1_232, __p2_232) __extension__ ({ \
   60981   uint32x2_t __s0_232 = __p0_232; \
   60982   uint64x2_t __s1_232 = __p1_232; \
   60983   uint32x4_t __ret_232; \
   60984   __ret_232 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_232), (uint32x2_t)(vshrn_n_u64(__s1_232, __p2_232)))); \
   60985   __ret_232; \
   60986 })
   60987 #else
   60988 #define vshrn_high_n_u64(__p0_233, __p1_233, __p2_233) __extension__ ({ \
   60989   uint32x2_t __s0_233 = __p0_233; \
   60990   uint64x2_t __s1_233 = __p1_233; \
   60991   uint32x2_t __rev0_233;  __rev0_233 = __builtin_shufflevector(__s0_233, __s0_233, 1, 0); \
   60992   uint64x2_t __rev1_233;  __rev1_233 = __builtin_shufflevector(__s1_233, __s1_233, 1, 0); \
   60993   uint32x4_t __ret_233; \
   60994   __ret_233 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_233), (uint32x2_t)(__noswap_vshrn_n_u64(__rev1_233, __p2_233)))); \
   60995   __ret_233 = __builtin_shufflevector(__ret_233, __ret_233, 3, 2, 1, 0); \
   60996   __ret_233; \
   60997 })
   60998 #endif
   60999 
   61000 #ifdef __LITTLE_ENDIAN__
   61001 #define vshrn_high_n_u16(__p0_234, __p1_234, __p2_234) __extension__ ({ \
   61002   uint8x8_t __s0_234 = __p0_234; \
   61003   uint16x8_t __s1_234 = __p1_234; \
   61004   uint8x16_t __ret_234; \
   61005   __ret_234 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_234), (uint8x8_t)(vshrn_n_u16(__s1_234, __p2_234)))); \
   61006   __ret_234; \
   61007 })
   61008 #else
   61009 #define vshrn_high_n_u16(__p0_235, __p1_235, __p2_235) __extension__ ({ \
   61010   uint8x8_t __s0_235 = __p0_235; \
   61011   uint16x8_t __s1_235 = __p1_235; \
   61012   uint8x8_t __rev0_235;  __rev0_235 = __builtin_shufflevector(__s0_235, __s0_235, 7, 6, 5, 4, 3, 2, 1, 0); \
   61013   uint16x8_t __rev1_235;  __rev1_235 = __builtin_shufflevector(__s1_235, __s1_235, 7, 6, 5, 4, 3, 2, 1, 0); \
   61014   uint8x16_t __ret_235; \
   61015   __ret_235 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_235), (uint8x8_t)(__noswap_vshrn_n_u16(__rev1_235, __p2_235)))); \
   61016   __ret_235 = __builtin_shufflevector(__ret_235, __ret_235, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   61017   __ret_235; \
   61018 })
   61019 #endif
   61020 
   61021 #ifdef __LITTLE_ENDIAN__
   61022 #define vshrn_high_n_s32(__p0_236, __p1_236, __p2_236) __extension__ ({ \
   61023   int16x4_t __s0_236 = __p0_236; \
   61024   int32x4_t __s1_236 = __p1_236; \
   61025   int16x8_t __ret_236; \
   61026   __ret_236 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_236), (int16x4_t)(vshrn_n_s32(__s1_236, __p2_236)))); \
   61027   __ret_236; \
   61028 })
   61029 #else
   61030 #define vshrn_high_n_s32(__p0_237, __p1_237, __p2_237) __extension__ ({ \
   61031   int16x4_t __s0_237 = __p0_237; \
   61032   int32x4_t __s1_237 = __p1_237; \
   61033   int16x4_t __rev0_237;  __rev0_237 = __builtin_shufflevector(__s0_237, __s0_237, 3, 2, 1, 0); \
   61034   int32x4_t __rev1_237;  __rev1_237 = __builtin_shufflevector(__s1_237, __s1_237, 3, 2, 1, 0); \
   61035   int16x8_t __ret_237; \
   61036   __ret_237 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_237), (int16x4_t)(__noswap_vshrn_n_s32(__rev1_237, __p2_237)))); \
   61037   __ret_237 = __builtin_shufflevector(__ret_237, __ret_237, 7, 6, 5, 4, 3, 2, 1, 0); \
   61038   __ret_237; \
   61039 })
   61040 #endif
   61041 
   61042 #ifdef __LITTLE_ENDIAN__
   61043 #define vshrn_high_n_s64(__p0_238, __p1_238, __p2_238) __extension__ ({ \
   61044   int32x2_t __s0_238 = __p0_238; \
   61045   int64x2_t __s1_238 = __p1_238; \
   61046   int32x4_t __ret_238; \
   61047   __ret_238 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_238), (int32x2_t)(vshrn_n_s64(__s1_238, __p2_238)))); \
   61048   __ret_238; \
   61049 })
   61050 #else
   61051 #define vshrn_high_n_s64(__p0_239, __p1_239, __p2_239) __extension__ ({ \
   61052   int32x2_t __s0_239 = __p0_239; \
   61053   int64x2_t __s1_239 = __p1_239; \
   61054   int32x2_t __rev0_239;  __rev0_239 = __builtin_shufflevector(__s0_239, __s0_239, 1, 0); \
   61055   int64x2_t __rev1_239;  __rev1_239 = __builtin_shufflevector(__s1_239, __s1_239, 1, 0); \
   61056   int32x4_t __ret_239; \
   61057   __ret_239 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_239), (int32x2_t)(__noswap_vshrn_n_s64(__rev1_239, __p2_239)))); \
   61058   __ret_239 = __builtin_shufflevector(__ret_239, __ret_239, 3, 2, 1, 0); \
   61059   __ret_239; \
   61060 })
   61061 #endif
   61062 
   61063 #ifdef __LITTLE_ENDIAN__
   61064 #define vshrn_high_n_s16(__p0_240, __p1_240, __p2_240) __extension__ ({ \
   61065   int8x8_t __s0_240 = __p0_240; \
   61066   int16x8_t __s1_240 = __p1_240; \
   61067   int8x16_t __ret_240; \
   61068   __ret_240 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_240), (int8x8_t)(vshrn_n_s16(__s1_240, __p2_240)))); \
   61069   __ret_240; \
   61070 })
   61071 #else
   61072 #define vshrn_high_n_s16(__p0_241, __p1_241, __p2_241) __extension__ ({ \
   61073   int8x8_t __s0_241 = __p0_241; \
   61074   int16x8_t __s1_241 = __p1_241; \
   61075   int8x8_t __rev0_241;  __rev0_241 = __builtin_shufflevector(__s0_241, __s0_241, 7, 6, 5, 4, 3, 2, 1, 0); \
   61076   int16x8_t __rev1_241;  __rev1_241 = __builtin_shufflevector(__s1_241, __s1_241, 7, 6, 5, 4, 3, 2, 1, 0); \
   61077   int8x16_t __ret_241; \
   61078   __ret_241 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_241), (int8x8_t)(__noswap_vshrn_n_s16(__rev1_241, __p2_241)))); \
   61079   __ret_241 = __builtin_shufflevector(__ret_241, __ret_241, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   61080   __ret_241; \
   61081 })
   61082 #endif
   61083 
   61084 #ifdef __LITTLE_ENDIAN__
   61085 #define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
   61086   uint64_t __s0 = __p0; \
   61087   uint64_t __s1 = __p1; \
   61088   uint64_t __ret; \
   61089   __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
   61090   __ret; \
   61091 })
   61092 #else
   61093 #define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
   61094   uint64_t __s0 = __p0; \
   61095   uint64_t __s1 = __p1; \
   61096   uint64_t __ret; \
   61097   __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
   61098   __ret; \
   61099 })
   61100 #endif
   61101 
   61102 #ifdef __LITTLE_ENDIAN__
   61103 #define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
   61104   int64_t __s0 = __p0; \
   61105   int64_t __s1 = __p1; \
   61106   int64_t __ret; \
   61107   __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
   61108   __ret; \
   61109 })
   61110 #else
   61111 #define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
   61112   int64_t __s0 = __p0; \
   61113   int64_t __s1 = __p1; \
   61114   int64_t __ret; \
   61115   __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
   61116   __ret; \
   61117 })
   61118 #endif
   61119 
   61120 #ifdef __LITTLE_ENDIAN__
   61121 #define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
   61122   poly64x1_t __s0 = __p0; \
   61123   poly64x1_t __s1 = __p1; \
   61124   poly64x1_t __ret; \
   61125   __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
   61126   __ret; \
   61127 })
   61128 #else
   61129 #define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
   61130   poly64x1_t __s0 = __p0; \
   61131   poly64x1_t __s1 = __p1; \
   61132   poly64x1_t __ret; \
   61133   __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
   61134   __ret; \
   61135 })
   61136 #endif
   61137 
   61138 #ifdef __LITTLE_ENDIAN__
   61139 #define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
   61140   poly64x2_t __s0 = __p0; \
   61141   poly64x2_t __s1 = __p1; \
   61142   poly64x2_t __ret; \
   61143   __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
   61144   __ret; \
   61145 })
   61146 #else
   61147 #define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
   61148   poly64x2_t __s0 = __p0; \
   61149   poly64x2_t __s1 = __p1; \
   61150   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   61151   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   61152   poly64x2_t __ret; \
   61153   __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
   61154   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   61155   __ret; \
   61156 })
   61157 #endif
   61158 
   61159 #ifdef __LITTLE_ENDIAN__
   61160 __ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
   61161   uint8_t __ret;
   61162   __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
   61163   return __ret;
   61164 }
   61165 #else
   61166 __ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
   61167   uint8_t __ret;
   61168   __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
   61169   return __ret;
   61170 }
   61171 #endif
   61172 
   61173 #ifdef __LITTLE_ENDIAN__
   61174 __ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
   61175   uint32_t __ret;
   61176   __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
   61177   return __ret;
   61178 }
   61179 #else
   61180 __ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
   61181   uint32_t __ret;
   61182   __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
   61183   return __ret;
   61184 }
   61185 #endif
   61186 
   61187 #ifdef __LITTLE_ENDIAN__
   61188 __ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
   61189   uint64_t __ret;
   61190   __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
   61191   return __ret;
   61192 }
   61193 #else
   61194 __ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
   61195   uint64_t __ret;
   61196   __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
   61197   return __ret;
   61198 }
   61199 #endif
   61200 
   61201 #ifdef __LITTLE_ENDIAN__
   61202 __ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
   61203   uint16_t __ret;
   61204   __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
   61205   return __ret;
   61206 }
   61207 #else
   61208 __ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
   61209   uint16_t __ret;
   61210   __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
   61211   return __ret;
   61212 }
   61213 #endif
   61214 
   61215 #ifdef __LITTLE_ENDIAN__
   61216 __ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   61217   uint8x16_t __ret;
   61218   __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
   61219   return __ret;
   61220 }
   61221 #else
   61222 __ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
   61223   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   61224   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   61225   uint8x16_t __ret;
   61226   __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
   61227   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   61228   return __ret;
   61229 }
   61230 #endif
   61231 
   61232 #ifdef __LITTLE_ENDIAN__
   61233 __ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   61234   uint32x4_t __ret;
   61235   __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
   61236   return __ret;
   61237 }
   61238 #else
   61239 __ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
   61240   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   61241   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   61242   uint32x4_t __ret;
   61243   __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
   61244   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   61245   return __ret;
   61246 }
   61247 #endif
   61248 
   61249 #ifdef __LITTLE_ENDIAN__
   61250 __ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   61251   uint64x2_t __ret;
   61252   __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   61253   return __ret;
   61254 }
   61255 #else
   61256 __ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   61257   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   61258   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   61259   uint64x2_t __ret;
   61260   __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   61261   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   61262   return __ret;
   61263 }
   61264 #endif
   61265 
   61266 #ifdef __LITTLE_ENDIAN__
   61267 __ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   61268   uint16x8_t __ret;
   61269   __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
   61270   return __ret;
   61271 }
   61272 #else
   61273 __ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
   61274   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   61275   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   61276   uint16x8_t __ret;
   61277   __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
   61278   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   61279   return __ret;
   61280 }
   61281 #endif
   61282 
   61283 #ifdef __LITTLE_ENDIAN__
   61284 __ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   61285   uint8x8_t __ret;
   61286   __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
   61287   return __ret;
   61288 }
   61289 #else
   61290 __ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
   61291   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   61292   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   61293   uint8x8_t __ret;
   61294   __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
   61295   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   61296   return __ret;
   61297 }
   61298 #endif
   61299 
   61300 #ifdef __LITTLE_ENDIAN__
   61301 __ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   61302   uint32x2_t __ret;
   61303   __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
   61304   return __ret;
   61305 }
   61306 #else
   61307 __ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
   61308   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   61309   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   61310   uint32x2_t __ret;
   61311   __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
   61312   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   61313   return __ret;
   61314 }
   61315 #endif
   61316 
   61317 #ifdef __LITTLE_ENDIAN__
   61318 __ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
   61319   uint64x1_t __ret;
   61320   __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   61321   return __ret;
   61322 }
   61323 #else
   61324 __ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
   61325   uint64x1_t __ret;
   61326   __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   61327   return __ret;
   61328 }
   61329 #endif
   61330 
   61331 #ifdef __LITTLE_ENDIAN__
   61332 __ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   61333   uint16x4_t __ret;
   61334   __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
   61335   return __ret;
   61336 }
   61337 #else
   61338 __ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
   61339   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   61340   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   61341   uint16x4_t __ret;
   61342   __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
   61343   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   61344   return __ret;
   61345 }
   61346 #endif
   61347 
   61348 #ifdef __LITTLE_ENDIAN__
   61349 __ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
   61350   float64x2_t __ret;
   61351   __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 42);
   61352   return __ret;
   61353 }
   61354 #else
   61355 __ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
   61356   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   61357   float64x2_t __ret;
   61358   __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 42);
   61359   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   61360   return __ret;
   61361 }
   61362 #endif
   61363 
   61364 #ifdef __LITTLE_ENDIAN__
   61365 __ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
   61366   float32x4_t __ret;
   61367   __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 41);
   61368   return __ret;
   61369 }
   61370 #else
   61371 __ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
   61372   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   61373   float32x4_t __ret;
   61374   __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 41);
   61375   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   61376   return __ret;
   61377 }
   61378 #endif
   61379 
   61380 #ifdef __LITTLE_ENDIAN__
   61381 __ai float64x1_t vsqrt_f64(float64x1_t __p0) {
   61382   float64x1_t __ret;
   61383   __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
   61384   return __ret;
   61385 }
   61386 #else
   61387 __ai float64x1_t vsqrt_f64(float64x1_t __p0) {
   61388   float64x1_t __ret;
   61389   __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
   61390   return __ret;
   61391 }
   61392 #endif
   61393 
   61394 #ifdef __LITTLE_ENDIAN__
   61395 __ai float32x2_t vsqrt_f32(float32x2_t __p0) {
   61396   float32x2_t __ret;
   61397   __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 9);
   61398   return __ret;
   61399 }
   61400 #else
   61401 __ai float32x2_t vsqrt_f32(float32x2_t __p0) {
   61402   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   61403   float32x2_t __ret;
   61404   __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__rev0, 9);
   61405   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   61406   return __ret;
   61407 }
   61408 #endif
   61409 
   61410 #ifdef __LITTLE_ENDIAN__
   61411 #define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
   61412   uint64_t __s0 = __p0; \
   61413   uint64_t __s1 = __p1; \
   61414   uint64_t __ret; \
   61415   __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
   61416   __ret; \
   61417 })
   61418 #else
   61419 #define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
   61420   uint64_t __s0 = __p0; \
   61421   uint64_t __s1 = __p1; \
   61422   uint64_t __ret; \
   61423   __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
   61424   __ret; \
   61425 })
   61426 #endif
   61427 
   61428 #ifdef __LITTLE_ENDIAN__
   61429 #define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
   61430   int64_t __s0 = __p0; \
   61431   int64_t __s1 = __p1; \
   61432   int64_t __ret; \
   61433   __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
   61434   __ret; \
   61435 })
   61436 #else
   61437 #define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
   61438   int64_t __s0 = __p0; \
   61439   int64_t __s1 = __p1; \
   61440   int64_t __ret; \
   61441   __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
   61442   __ret; \
   61443 })
   61444 #endif
   61445 
   61446 #ifdef __LITTLE_ENDIAN__
   61447 #define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
   61448   uint64_t __s0 = __p0; \
   61449   uint64_t __s1 = __p1; \
   61450   uint64_t __ret; \
   61451   __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
   61452   __ret; \
   61453 })
   61454 #else
   61455 #define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
   61456   uint64_t __s0 = __p0; \
   61457   uint64_t __s1 = __p1; \
   61458   uint64_t __ret; \
   61459   __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
   61460   __ret; \
   61461 })
   61462 #endif
   61463 
   61464 #ifdef __LITTLE_ENDIAN__
   61465 #define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
   61466   int64_t __s0 = __p0; \
   61467   int64_t __s1 = __p1; \
   61468   int64_t __ret; \
   61469   __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
   61470   __ret; \
   61471 })
   61472 #else
   61473 #define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
   61474   int64_t __s0 = __p0; \
   61475   int64_t __s1 = __p1; \
   61476   int64_t __ret; \
   61477   __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
   61478   __ret; \
   61479 })
   61480 #endif
   61481 
   61482 #ifdef __LITTLE_ENDIAN__
   61483 #define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
   61484   poly64x1_t __s0 = __p0; \
   61485   poly64x1_t __s1 = __p1; \
   61486   poly64x1_t __ret; \
   61487   __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
   61488   __ret; \
   61489 })
   61490 #else
   61491 #define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
   61492   poly64x1_t __s0 = __p0; \
   61493   poly64x1_t __s1 = __p1; \
   61494   poly64x1_t __ret; \
   61495   __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
   61496   __ret; \
   61497 })
   61498 #endif
   61499 
   61500 #ifdef __LITTLE_ENDIAN__
   61501 #define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
   61502   poly64x2_t __s0 = __p0; \
   61503   poly64x2_t __s1 = __p1; \
   61504   poly64x2_t __ret; \
   61505   __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
   61506   __ret; \
   61507 })
   61508 #else
   61509 #define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
   61510   poly64x2_t __s0 = __p0; \
   61511   poly64x2_t __s1 = __p1; \
   61512   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   61513   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   61514   poly64x2_t __ret; \
   61515   __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
   61516   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   61517   __ret; \
   61518 })
   61519 #endif
   61520 
   61521 #ifdef __LITTLE_ENDIAN__
   61522 #define vst1_p64(__p0, __p1) __extension__ ({ \
   61523   poly64x1_t __s1 = __p1; \
   61524   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
   61525 })
   61526 #else
   61527 #define vst1_p64(__p0, __p1) __extension__ ({ \
   61528   poly64x1_t __s1 = __p1; \
   61529   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
   61530 })
   61531 #endif
   61532 
   61533 #ifdef __LITTLE_ENDIAN__
   61534 #define vst1q_p64(__p0, __p1) __extension__ ({ \
   61535   poly64x2_t __s1 = __p1; \
   61536   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 38); \
   61537 })
   61538 #else
   61539 #define vst1q_p64(__p0, __p1) __extension__ ({ \
   61540   poly64x2_t __s1 = __p1; \
   61541   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   61542   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 38); \
   61543 })
   61544 #endif
   61545 
   61546 #ifdef __LITTLE_ENDIAN__
   61547 #define vst1q_f64(__p0, __p1) __extension__ ({ \
   61548   float64x2_t __s1 = __p1; \
   61549   __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 42); \
   61550 })
   61551 #else
   61552 #define vst1q_f64(__p0, __p1) __extension__ ({ \
   61553   float64x2_t __s1 = __p1; \
   61554   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   61555   __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 42); \
   61556 })
   61557 #endif
   61558 
   61559 #ifdef __LITTLE_ENDIAN__
   61560 #define vst1_f64(__p0, __p1) __extension__ ({ \
   61561   float64x1_t __s1 = __p1; \
   61562   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
   61563 })
   61564 #else
   61565 #define vst1_f64(__p0, __p1) __extension__ ({ \
   61566   float64x1_t __s1 = __p1; \
   61567   __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
   61568 })
   61569 #endif
   61570 
   61571 #ifdef __LITTLE_ENDIAN__
   61572 #define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   61573   poly64x1_t __s1 = __p1; \
   61574   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
   61575 })
   61576 #else
   61577 #define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   61578   poly64x1_t __s1 = __p1; \
   61579   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
   61580 })
   61581 #endif
   61582 
   61583 #ifdef __LITTLE_ENDIAN__
   61584 #define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   61585   poly64x2_t __s1 = __p1; \
   61586   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
   61587 })
   61588 #else
   61589 #define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   61590   poly64x2_t __s1 = __p1; \
   61591   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   61592   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
   61593 })
   61594 #endif
   61595 
   61596 #ifdef __LITTLE_ENDIAN__
   61597 #define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   61598   float64x2_t __s1 = __p1; \
   61599   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
   61600 })
   61601 #else
   61602 #define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   61603   float64x2_t __s1 = __p1; \
   61604   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   61605   __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
   61606 })
   61607 #endif
   61608 
   61609 #ifdef __LITTLE_ENDIAN__
   61610 #define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   61611   float64x1_t __s1 = __p1; \
   61612   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
   61613 })
   61614 #else
   61615 #define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   61616   float64x1_t __s1 = __p1; \
   61617   __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
   61618 })
   61619 #endif
   61620 
   61621 #ifdef __LITTLE_ENDIAN__
   61622 #define vst1_p8_x2(__p0, __p1) __extension__ ({ \
   61623   poly8x8x2_t __s1 = __p1; \
   61624   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
   61625 })
   61626 #else
   61627 #define vst1_p8_x2(__p0, __p1) __extension__ ({ \
   61628   poly8x8x2_t __s1 = __p1; \
   61629   poly8x8x2_t __rev1; \
   61630   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   61631   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   61632   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
   61633 })
   61634 #endif
   61635 
   61636 #ifdef __LITTLE_ENDIAN__
   61637 #define vst1_p64_x2(__p0, __p1) __extension__ ({ \
   61638   poly64x1x2_t __s1 = __p1; \
   61639   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
   61640 })
   61641 #else
   61642 #define vst1_p64_x2(__p0, __p1) __extension__ ({ \
   61643   poly64x1x2_t __s1 = __p1; \
   61644   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
   61645 })
   61646 #endif
   61647 
   61648 #ifdef __LITTLE_ENDIAN__
   61649 #define vst1_p16_x2(__p0, __p1) __extension__ ({ \
   61650   poly16x4x2_t __s1 = __p1; \
   61651   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
   61652 })
   61653 #else
   61654 #define vst1_p16_x2(__p0, __p1) __extension__ ({ \
   61655   poly16x4x2_t __s1 = __p1; \
   61656   poly16x4x2_t __rev1; \
   61657   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   61658   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   61659   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
   61660 })
   61661 #endif
   61662 
   61663 #ifdef __LITTLE_ENDIAN__
   61664 #define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
   61665   poly8x16x2_t __s1 = __p1; \
   61666   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
   61667 })
   61668 #else
   61669 #define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
   61670   poly8x16x2_t __s1 = __p1; \
   61671   poly8x16x2_t __rev1; \
   61672   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   61673   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   61674   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
   61675 })
   61676 #endif
   61677 
   61678 #ifdef __LITTLE_ENDIAN__
   61679 #define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
   61680   poly64x2x2_t __s1 = __p1; \
   61681   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
   61682 })
   61683 #else
   61684 #define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
   61685   poly64x2x2_t __s1 = __p1; \
   61686   poly64x2x2_t __rev1; \
   61687   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   61688   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   61689   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
   61690 })
   61691 #endif
   61692 
   61693 #ifdef __LITTLE_ENDIAN__
   61694 #define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
   61695   poly16x8x2_t __s1 = __p1; \
   61696   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
   61697 })
   61698 #else
   61699 #define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
   61700   poly16x8x2_t __s1 = __p1; \
   61701   poly16x8x2_t __rev1; \
   61702   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   61703   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   61704   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
   61705 })
   61706 #endif
   61707 
   61708 #ifdef __LITTLE_ENDIAN__
   61709 #define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
   61710   uint8x16x2_t __s1 = __p1; \
   61711   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
   61712 })
   61713 #else
   61714 #define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
   61715   uint8x16x2_t __s1 = __p1; \
   61716   uint8x16x2_t __rev1; \
   61717   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   61718   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   61719   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
   61720 })
   61721 #endif
   61722 
   61723 #ifdef __LITTLE_ENDIAN__
   61724 #define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
   61725   uint32x4x2_t __s1 = __p1; \
   61726   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
   61727 })
   61728 #else
   61729 #define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
   61730   uint32x4x2_t __s1 = __p1; \
   61731   uint32x4x2_t __rev1; \
   61732   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   61733   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   61734   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
   61735 })
   61736 #endif
   61737 
   61738 #ifdef __LITTLE_ENDIAN__
   61739 #define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
   61740   uint64x2x2_t __s1 = __p1; \
   61741   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
   61742 })
   61743 #else
   61744 #define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
   61745   uint64x2x2_t __s1 = __p1; \
   61746   uint64x2x2_t __rev1; \
   61747   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   61748   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   61749   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
   61750 })
   61751 #endif
   61752 
   61753 #ifdef __LITTLE_ENDIAN__
   61754 #define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
   61755   uint16x8x2_t __s1 = __p1; \
   61756   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
   61757 })
   61758 #else
   61759 #define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
   61760   uint16x8x2_t __s1 = __p1; \
   61761   uint16x8x2_t __rev1; \
   61762   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   61763   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   61764   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
   61765 })
   61766 #endif
   61767 
   61768 #ifdef __LITTLE_ENDIAN__
   61769 #define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
   61770   int8x16x2_t __s1 = __p1; \
   61771   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
   61772 })
   61773 #else
   61774 #define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
   61775   int8x16x2_t __s1 = __p1; \
   61776   int8x16x2_t __rev1; \
   61777   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   61778   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   61779   __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
   61780 })
   61781 #endif
   61782 
   61783 #ifdef __LITTLE_ENDIAN__
   61784 #define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
   61785   float64x2x2_t __s1 = __p1; \
   61786   __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 42); \
   61787 })
   61788 #else
   61789 #define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
   61790   float64x2x2_t __s1 = __p1; \
   61791   float64x2x2_t __rev1; \
   61792   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   61793   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   61794   __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
   61795 })
   61796 #endif
   61797 
   61798 #ifdef __LITTLE_ENDIAN__
   61799 #define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
   61800   float32x4x2_t __s1 = __p1; \
   61801   __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 41); \
   61802 })
   61803 #else
   61804 #define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
   61805   float32x4x2_t __s1 = __p1; \
   61806   float32x4x2_t __rev1; \
   61807   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   61808   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   61809   __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
   61810 })
   61811 #endif
   61812 
   61813 #ifdef __LITTLE_ENDIAN__
   61814 #define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
   61815   float16x8x2_t __s1 = __p1; \
   61816   __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 40); \
   61817 })
   61818 #else
   61819 #define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
   61820   float16x8x2_t __s1 = __p1; \
   61821   float16x8x2_t __rev1; \
   61822   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   61823   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   61824   __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
   61825 })
   61826 #endif
   61827 
   61828 #ifdef __LITTLE_ENDIAN__
   61829 #define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
   61830   int32x4x2_t __s1 = __p1; \
   61831   __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 34); \
   61832 })
   61833 #else
   61834 #define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
   61835   int32x4x2_t __s1 = __p1; \
   61836   int32x4x2_t __rev1; \
   61837   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   61838   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   61839   __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
   61840 })
   61841 #endif
   61842 
   61843 #ifdef __LITTLE_ENDIAN__
   61844 #define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
   61845   int64x2x2_t __s1 = __p1; \
   61846   __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 35); \
   61847 })
   61848 #else
   61849 #define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
   61850   int64x2x2_t __s1 = __p1; \
   61851   int64x2x2_t __rev1; \
   61852   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   61853   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   61854   __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
   61855 })
   61856 #endif
   61857 
   61858 #ifdef __LITTLE_ENDIAN__
   61859 #define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
   61860   int16x8x2_t __s1 = __p1; \
   61861   __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 33); \
   61862 })
   61863 #else
   61864 #define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
   61865   int16x8x2_t __s1 = __p1; \
   61866   int16x8x2_t __rev1; \
   61867   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   61868   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   61869   __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
   61870 })
   61871 #endif
   61872 
   61873 #ifdef __LITTLE_ENDIAN__
   61874 #define vst1_u8_x2(__p0, __p1) __extension__ ({ \
   61875   uint8x8x2_t __s1 = __p1; \
   61876   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
   61877 })
   61878 #else
   61879 #define vst1_u8_x2(__p0, __p1) __extension__ ({ \
   61880   uint8x8x2_t __s1 = __p1; \
   61881   uint8x8x2_t __rev1; \
   61882   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   61883   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   61884   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
   61885 })
   61886 #endif
   61887 
   61888 #ifdef __LITTLE_ENDIAN__
   61889 #define vst1_u32_x2(__p0, __p1) __extension__ ({ \
   61890   uint32x2x2_t __s1 = __p1; \
   61891   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
   61892 })
   61893 #else
   61894 #define vst1_u32_x2(__p0, __p1) __extension__ ({ \
   61895   uint32x2x2_t __s1 = __p1; \
   61896   uint32x2x2_t __rev1; \
   61897   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   61898   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   61899   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
   61900 })
   61901 #endif
   61902 
   61903 #ifdef __LITTLE_ENDIAN__
   61904 #define vst1_u64_x2(__p0, __p1) __extension__ ({ \
   61905   uint64x1x2_t __s1 = __p1; \
   61906   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
   61907 })
   61908 #else
   61909 #define vst1_u64_x2(__p0, __p1) __extension__ ({ \
   61910   uint64x1x2_t __s1 = __p1; \
   61911   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
   61912 })
   61913 #endif
   61914 
   61915 #ifdef __LITTLE_ENDIAN__
   61916 #define vst1_u16_x2(__p0, __p1) __extension__ ({ \
   61917   uint16x4x2_t __s1 = __p1; \
   61918   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
   61919 })
   61920 #else
   61921 #define vst1_u16_x2(__p0, __p1) __extension__ ({ \
   61922   uint16x4x2_t __s1 = __p1; \
   61923   uint16x4x2_t __rev1; \
   61924   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   61925   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   61926   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
   61927 })
   61928 #endif
   61929 
   61930 #ifdef __LITTLE_ENDIAN__
   61931 #define vst1_s8_x2(__p0, __p1) __extension__ ({ \
   61932   int8x8x2_t __s1 = __p1; \
   61933   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
   61934 })
   61935 #else
   61936 #define vst1_s8_x2(__p0, __p1) __extension__ ({ \
   61937   int8x8x2_t __s1 = __p1; \
   61938   int8x8x2_t __rev1; \
   61939   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   61940   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   61941   __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
   61942 })
   61943 #endif
   61944 
   61945 #ifdef __LITTLE_ENDIAN__
   61946 #define vst1_f64_x2(__p0, __p1) __extension__ ({ \
   61947   float64x1x2_t __s1 = __p1; \
   61948   __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
   61949 })
   61950 #else
   61951 #define vst1_f64_x2(__p0, __p1) __extension__ ({ \
   61952   float64x1x2_t __s1 = __p1; \
   61953   __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
   61954 })
   61955 #endif
   61956 
   61957 #ifdef __LITTLE_ENDIAN__
   61958 #define vst1_f32_x2(__p0, __p1) __extension__ ({ \
   61959   float32x2x2_t __s1 = __p1; \
   61960   __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 9); \
   61961 })
   61962 #else
   61963 #define vst1_f32_x2(__p0, __p1) __extension__ ({ \
   61964   float32x2x2_t __s1 = __p1; \
   61965   float32x2x2_t __rev1; \
   61966   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   61967   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   61968   __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
   61969 })
   61970 #endif
   61971 
   61972 #ifdef __LITTLE_ENDIAN__
   61973 #define vst1_f16_x2(__p0, __p1) __extension__ ({ \
   61974   float16x4x2_t __s1 = __p1; \
   61975   __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 8); \
   61976 })
   61977 #else
   61978 #define vst1_f16_x2(__p0, __p1) __extension__ ({ \
   61979   float16x4x2_t __s1 = __p1; \
   61980   float16x4x2_t __rev1; \
   61981   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   61982   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   61983   __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
   61984 })
   61985 #endif
   61986 
   61987 #ifdef __LITTLE_ENDIAN__
   61988 #define vst1_s32_x2(__p0, __p1) __extension__ ({ \
   61989   int32x2x2_t __s1 = __p1; \
   61990   __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 2); \
   61991 })
   61992 #else
   61993 #define vst1_s32_x2(__p0, __p1) __extension__ ({ \
   61994   int32x2x2_t __s1 = __p1; \
   61995   int32x2x2_t __rev1; \
   61996   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   61997   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   61998   __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
   61999 })
   62000 #endif
   62001 
   62002 #ifdef __LITTLE_ENDIAN__
   62003 #define vst1_s64_x2(__p0, __p1) __extension__ ({ \
   62004   int64x1x2_t __s1 = __p1; \
   62005   __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
   62006 })
   62007 #else
   62008 #define vst1_s64_x2(__p0, __p1) __extension__ ({ \
   62009   int64x1x2_t __s1 = __p1; \
   62010   __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
   62011 })
   62012 #endif
   62013 
   62014 #ifdef __LITTLE_ENDIAN__
   62015 #define vst1_s16_x2(__p0, __p1) __extension__ ({ \
   62016   int16x4x2_t __s1 = __p1; \
   62017   __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 1); \
   62018 })
   62019 #else
   62020 #define vst1_s16_x2(__p0, __p1) __extension__ ({ \
   62021   int16x4x2_t __s1 = __p1; \
   62022   int16x4x2_t __rev1; \
   62023   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62024   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62025   __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
   62026 })
   62027 #endif
   62028 
   62029 #ifdef __LITTLE_ENDIAN__
   62030 #define vst1_p8_x3(__p0, __p1) __extension__ ({ \
   62031   poly8x8x3_t __s1 = __p1; \
   62032   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
   62033 })
   62034 #else
   62035 #define vst1_p8_x3(__p0, __p1) __extension__ ({ \
   62036   poly8x8x3_t __s1 = __p1; \
   62037   poly8x8x3_t __rev1; \
   62038   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62039   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62040   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62041   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
   62042 })
   62043 #endif
   62044 
   62045 #ifdef __LITTLE_ENDIAN__
   62046 #define vst1_p64_x3(__p0, __p1) __extension__ ({ \
   62047   poly64x1x3_t __s1 = __p1; \
   62048   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
   62049 })
   62050 #else
   62051 #define vst1_p64_x3(__p0, __p1) __extension__ ({ \
   62052   poly64x1x3_t __s1 = __p1; \
   62053   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
   62054 })
   62055 #endif
   62056 
   62057 #ifdef __LITTLE_ENDIAN__
   62058 #define vst1_p16_x3(__p0, __p1) __extension__ ({ \
   62059   poly16x4x3_t __s1 = __p1; \
   62060   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
   62061 })
   62062 #else
   62063 #define vst1_p16_x3(__p0, __p1) __extension__ ({ \
   62064   poly16x4x3_t __s1 = __p1; \
   62065   poly16x4x3_t __rev1; \
   62066   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62067   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62068   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62069   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
   62070 })
   62071 #endif
   62072 
   62073 #ifdef __LITTLE_ENDIAN__
   62074 #define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
   62075   poly8x16x3_t __s1 = __p1; \
   62076   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
   62077 })
   62078 #else
   62079 #define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
   62080   poly8x16x3_t __s1 = __p1; \
   62081   poly8x16x3_t __rev1; \
   62082   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62083   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62084   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62085   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
   62086 })
   62087 #endif
   62088 
   62089 #ifdef __LITTLE_ENDIAN__
   62090 #define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
   62091   poly64x2x3_t __s1 = __p1; \
   62092   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
   62093 })
   62094 #else
   62095 #define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
   62096   poly64x2x3_t __s1 = __p1; \
   62097   poly64x2x3_t __rev1; \
   62098   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62099   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62100   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62101   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
   62102 })
   62103 #endif
   62104 
   62105 #ifdef __LITTLE_ENDIAN__
   62106 #define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
   62107   poly16x8x3_t __s1 = __p1; \
   62108   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
   62109 })
   62110 #else
   62111 #define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
   62112   poly16x8x3_t __s1 = __p1; \
   62113   poly16x8x3_t __rev1; \
   62114   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62115   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62116   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62117   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
   62118 })
   62119 #endif
   62120 
   62121 #ifdef __LITTLE_ENDIAN__
   62122 #define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
   62123   uint8x16x3_t __s1 = __p1; \
   62124   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
   62125 })
   62126 #else
   62127 #define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
   62128   uint8x16x3_t __s1 = __p1; \
   62129   uint8x16x3_t __rev1; \
   62130   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62131   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62132   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62133   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
   62134 })
   62135 #endif
   62136 
   62137 #ifdef __LITTLE_ENDIAN__
   62138 #define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
   62139   uint32x4x3_t __s1 = __p1; \
   62140   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
   62141 })
   62142 #else
   62143 #define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
   62144   uint32x4x3_t __s1 = __p1; \
   62145   uint32x4x3_t __rev1; \
   62146   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62147   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62148   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62149   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
   62150 })
   62151 #endif
   62152 
   62153 #ifdef __LITTLE_ENDIAN__
   62154 #define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
   62155   uint64x2x3_t __s1 = __p1; \
   62156   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
   62157 })
   62158 #else
   62159 #define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
   62160   uint64x2x3_t __s1 = __p1; \
   62161   uint64x2x3_t __rev1; \
   62162   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62163   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62164   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62165   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
   62166 })
   62167 #endif
   62168 
   62169 #ifdef __LITTLE_ENDIAN__
   62170 #define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
   62171   uint16x8x3_t __s1 = __p1; \
   62172   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
   62173 })
   62174 #else
   62175 #define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
   62176   uint16x8x3_t __s1 = __p1; \
   62177   uint16x8x3_t __rev1; \
   62178   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62179   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62180   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62181   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
   62182 })
   62183 #endif
   62184 
   62185 #ifdef __LITTLE_ENDIAN__
   62186 #define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
   62187   int8x16x3_t __s1 = __p1; \
   62188   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
   62189 })
   62190 #else
   62191 #define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
   62192   int8x16x3_t __s1 = __p1; \
   62193   int8x16x3_t __rev1; \
   62194   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62195   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62196   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62197   __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
   62198 })
   62199 #endif
   62200 
   62201 #ifdef __LITTLE_ENDIAN__
   62202 #define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
   62203   float64x2x3_t __s1 = __p1; \
   62204   __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
   62205 })
   62206 #else
   62207 #define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
   62208   float64x2x3_t __s1 = __p1; \
   62209   float64x2x3_t __rev1; \
   62210   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62211   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62212   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62213   __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
   62214 })
   62215 #endif
   62216 
   62217 #ifdef __LITTLE_ENDIAN__
   62218 #define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
   62219   float32x4x3_t __s1 = __p1; \
   62220   __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
   62221 })
   62222 #else
   62223 #define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
   62224   float32x4x3_t __s1 = __p1; \
   62225   float32x4x3_t __rev1; \
   62226   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62227   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62228   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62229   __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
   62230 })
   62231 #endif
   62232 
   62233 #ifdef __LITTLE_ENDIAN__
   62234 #define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
   62235   float16x8x3_t __s1 = __p1; \
   62236   __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
   62237 })
   62238 #else
   62239 #define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
   62240   float16x8x3_t __s1 = __p1; \
   62241   float16x8x3_t __rev1; \
   62242   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62243   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62244   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62245   __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
   62246 })
   62247 #endif
   62248 
   62249 #ifdef __LITTLE_ENDIAN__
   62250 #define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
   62251   int32x4x3_t __s1 = __p1; \
   62252   __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
   62253 })
   62254 #else
   62255 #define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
   62256   int32x4x3_t __s1 = __p1; \
   62257   int32x4x3_t __rev1; \
   62258   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62259   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62260   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62261   __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
   62262 })
   62263 #endif
   62264 
   62265 #ifdef __LITTLE_ENDIAN__
   62266 #define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
   62267   int64x2x3_t __s1 = __p1; \
   62268   __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
   62269 })
   62270 #else
   62271 #define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
   62272   int64x2x3_t __s1 = __p1; \
   62273   int64x2x3_t __rev1; \
   62274   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62275   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62276   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62277   __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
   62278 })
   62279 #endif
   62280 
   62281 #ifdef __LITTLE_ENDIAN__
   62282 #define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
   62283   int16x8x3_t __s1 = __p1; \
   62284   __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
   62285 })
   62286 #else
   62287 #define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
   62288   int16x8x3_t __s1 = __p1; \
   62289   int16x8x3_t __rev1; \
   62290   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62291   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62292   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62293   __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
   62294 })
   62295 #endif
   62296 
   62297 #ifdef __LITTLE_ENDIAN__
   62298 #define vst1_u8_x3(__p0, __p1) __extension__ ({ \
   62299   uint8x8x3_t __s1 = __p1; \
   62300   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
   62301 })
   62302 #else
   62303 #define vst1_u8_x3(__p0, __p1) __extension__ ({ \
   62304   uint8x8x3_t __s1 = __p1; \
   62305   uint8x8x3_t __rev1; \
   62306   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62307   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62308   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62309   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
   62310 })
   62311 #endif
   62312 
   62313 #ifdef __LITTLE_ENDIAN__
   62314 #define vst1_u32_x3(__p0, __p1) __extension__ ({ \
   62315   uint32x2x3_t __s1 = __p1; \
   62316   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
   62317 })
   62318 #else
   62319 #define vst1_u32_x3(__p0, __p1) __extension__ ({ \
   62320   uint32x2x3_t __s1 = __p1; \
   62321   uint32x2x3_t __rev1; \
   62322   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62323   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62324   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62325   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
   62326 })
   62327 #endif
   62328 
   62329 #ifdef __LITTLE_ENDIAN__
   62330 #define vst1_u64_x3(__p0, __p1) __extension__ ({ \
   62331   uint64x1x3_t __s1 = __p1; \
   62332   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
   62333 })
   62334 #else
   62335 #define vst1_u64_x3(__p0, __p1) __extension__ ({ \
   62336   uint64x1x3_t __s1 = __p1; \
   62337   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
   62338 })
   62339 #endif
   62340 
   62341 #ifdef __LITTLE_ENDIAN__
   62342 #define vst1_u16_x3(__p0, __p1) __extension__ ({ \
   62343   uint16x4x3_t __s1 = __p1; \
   62344   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
   62345 })
   62346 #else
   62347 #define vst1_u16_x3(__p0, __p1) __extension__ ({ \
   62348   uint16x4x3_t __s1 = __p1; \
   62349   uint16x4x3_t __rev1; \
   62350   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62351   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62352   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62353   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
   62354 })
   62355 #endif
   62356 
   62357 #ifdef __LITTLE_ENDIAN__
   62358 #define vst1_s8_x3(__p0, __p1) __extension__ ({ \
   62359   int8x8x3_t __s1 = __p1; \
   62360   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
   62361 })
   62362 #else
   62363 #define vst1_s8_x3(__p0, __p1) __extension__ ({ \
   62364   int8x8x3_t __s1 = __p1; \
   62365   int8x8x3_t __rev1; \
   62366   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62367   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62368   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62369   __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
   62370 })
   62371 #endif
   62372 
   62373 #ifdef __LITTLE_ENDIAN__
   62374 #define vst1_f64_x3(__p0, __p1) __extension__ ({ \
   62375   float64x1x3_t __s1 = __p1; \
   62376   __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
   62377 })
   62378 #else
   62379 #define vst1_f64_x3(__p0, __p1) __extension__ ({ \
   62380   float64x1x3_t __s1 = __p1; \
   62381   __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
   62382 })
   62383 #endif
   62384 
   62385 #ifdef __LITTLE_ENDIAN__
   62386 #define vst1_f32_x3(__p0, __p1) __extension__ ({ \
   62387   float32x2x3_t __s1 = __p1; \
   62388   __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
   62389 })
   62390 #else
   62391 #define vst1_f32_x3(__p0, __p1) __extension__ ({ \
   62392   float32x2x3_t __s1 = __p1; \
   62393   float32x2x3_t __rev1; \
   62394   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62395   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62396   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62397   __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
   62398 })
   62399 #endif
   62400 
   62401 #ifdef __LITTLE_ENDIAN__
   62402 #define vst1_f16_x3(__p0, __p1) __extension__ ({ \
   62403   float16x4x3_t __s1 = __p1; \
   62404   __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
   62405 })
   62406 #else
   62407 #define vst1_f16_x3(__p0, __p1) __extension__ ({ \
   62408   float16x4x3_t __s1 = __p1; \
   62409   float16x4x3_t __rev1; \
   62410   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62411   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62412   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62413   __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
   62414 })
   62415 #endif
   62416 
   62417 #ifdef __LITTLE_ENDIAN__
   62418 #define vst1_s32_x3(__p0, __p1) __extension__ ({ \
   62419   int32x2x3_t __s1 = __p1; \
   62420   __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
   62421 })
   62422 #else
   62423 #define vst1_s32_x3(__p0, __p1) __extension__ ({ \
   62424   int32x2x3_t __s1 = __p1; \
   62425   int32x2x3_t __rev1; \
   62426   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62427   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62428   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62429   __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
   62430 })
   62431 #endif
   62432 
   62433 #ifdef __LITTLE_ENDIAN__
   62434 #define vst1_s64_x3(__p0, __p1) __extension__ ({ \
   62435   int64x1x3_t __s1 = __p1; \
   62436   __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
   62437 })
   62438 #else
   62439 #define vst1_s64_x3(__p0, __p1) __extension__ ({ \
   62440   int64x1x3_t __s1 = __p1; \
   62441   __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
   62442 })
   62443 #endif
   62444 
   62445 #ifdef __LITTLE_ENDIAN__
   62446 #define vst1_s16_x3(__p0, __p1) __extension__ ({ \
   62447   int16x4x3_t __s1 = __p1; \
   62448   __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
   62449 })
   62450 #else
   62451 #define vst1_s16_x3(__p0, __p1) __extension__ ({ \
   62452   int16x4x3_t __s1 = __p1; \
   62453   int16x4x3_t __rev1; \
   62454   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62455   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62456   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62457   __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
   62458 })
   62459 #endif
   62460 
   62461 #ifdef __LITTLE_ENDIAN__
   62462 #define vst1_p8_x4(__p0, __p1) __extension__ ({ \
   62463   poly8x8x4_t __s1 = __p1; \
   62464   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
   62465 })
   62466 #else
   62467 #define vst1_p8_x4(__p0, __p1) __extension__ ({ \
   62468   poly8x8x4_t __s1 = __p1; \
   62469   poly8x8x4_t __rev1; \
   62470   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62471   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62472   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62473   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   62474   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
   62475 })
   62476 #endif
   62477 
   62478 #ifdef __LITTLE_ENDIAN__
   62479 #define vst1_p64_x4(__p0, __p1) __extension__ ({ \
   62480   poly64x1x4_t __s1 = __p1; \
   62481   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
   62482 })
   62483 #else
   62484 #define vst1_p64_x4(__p0, __p1) __extension__ ({ \
   62485   poly64x1x4_t __s1 = __p1; \
   62486   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
   62487 })
   62488 #endif
   62489 
   62490 #ifdef __LITTLE_ENDIAN__
   62491 #define vst1_p16_x4(__p0, __p1) __extension__ ({ \
   62492   poly16x4x4_t __s1 = __p1; \
   62493   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
   62494 })
   62495 #else
   62496 #define vst1_p16_x4(__p0, __p1) __extension__ ({ \
   62497   poly16x4x4_t __s1 = __p1; \
   62498   poly16x4x4_t __rev1; \
   62499   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62500   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62501   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62502   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   62503   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
   62504 })
   62505 #endif
   62506 
   62507 #ifdef __LITTLE_ENDIAN__
   62508 #define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
   62509   poly8x16x4_t __s1 = __p1; \
   62510   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
   62511 })
   62512 #else
   62513 #define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
   62514   poly8x16x4_t __s1 = __p1; \
   62515   poly8x16x4_t __rev1; \
   62516   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62517   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62518   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62519   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62520   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
   62521 })
   62522 #endif
   62523 
   62524 #ifdef __LITTLE_ENDIAN__
   62525 #define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
   62526   poly64x2x4_t __s1 = __p1; \
   62527   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
   62528 })
   62529 #else
   62530 #define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
   62531   poly64x2x4_t __s1 = __p1; \
   62532   poly64x2x4_t __rev1; \
   62533   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62534   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62535   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62536   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   62537   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
   62538 })
   62539 #endif
   62540 
   62541 #ifdef __LITTLE_ENDIAN__
   62542 #define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
   62543   poly16x8x4_t __s1 = __p1; \
   62544   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
   62545 })
   62546 #else
   62547 #define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
   62548   poly16x8x4_t __s1 = __p1; \
   62549   poly16x8x4_t __rev1; \
   62550   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62551   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62552   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62553   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   62554   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
   62555 })
   62556 #endif
   62557 
   62558 #ifdef __LITTLE_ENDIAN__
   62559 #define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
   62560   uint8x16x4_t __s1 = __p1; \
   62561   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
   62562 })
   62563 #else
   62564 #define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
   62565   uint8x16x4_t __s1 = __p1; \
   62566   uint8x16x4_t __rev1; \
   62567   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62568   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62569   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62570   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62571   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
   62572 })
   62573 #endif
   62574 
   62575 #ifdef __LITTLE_ENDIAN__
   62576 #define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
   62577   uint32x4x4_t __s1 = __p1; \
   62578   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
   62579 })
   62580 #else
   62581 #define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
   62582   uint32x4x4_t __s1 = __p1; \
   62583   uint32x4x4_t __rev1; \
   62584   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62585   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62586   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62587   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   62588   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
   62589 })
   62590 #endif
   62591 
   62592 #ifdef __LITTLE_ENDIAN__
   62593 #define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
   62594   uint64x2x4_t __s1 = __p1; \
   62595   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
   62596 })
   62597 #else
   62598 #define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
   62599   uint64x2x4_t __s1 = __p1; \
   62600   uint64x2x4_t __rev1; \
   62601   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62602   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62603   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62604   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   62605   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
   62606 })
   62607 #endif
   62608 
   62609 #ifdef __LITTLE_ENDIAN__
   62610 #define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
   62611   uint16x8x4_t __s1 = __p1; \
   62612   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
   62613 })
   62614 #else
   62615 #define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
   62616   uint16x8x4_t __s1 = __p1; \
   62617   uint16x8x4_t __rev1; \
   62618   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62619   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62620   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62621   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   62622   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
   62623 })
   62624 #endif
   62625 
   62626 #ifdef __LITTLE_ENDIAN__
   62627 #define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
   62628   int8x16x4_t __s1 = __p1; \
   62629   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
   62630 })
   62631 #else
   62632 #define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
   62633   int8x16x4_t __s1 = __p1; \
   62634   int8x16x4_t __rev1; \
   62635   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62636   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62637   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62638   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   62639   __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
   62640 })
   62641 #endif
   62642 
   62643 #ifdef __LITTLE_ENDIAN__
   62644 #define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
   62645   float64x2x4_t __s1 = __p1; \
   62646   __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
   62647 })
   62648 #else
   62649 #define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
   62650   float64x2x4_t __s1 = __p1; \
   62651   float64x2x4_t __rev1; \
   62652   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62653   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62654   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62655   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   62656   __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
   62657 })
   62658 #endif
   62659 
   62660 #ifdef __LITTLE_ENDIAN__
   62661 #define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
   62662   float32x4x4_t __s1 = __p1; \
   62663   __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
   62664 })
   62665 #else
   62666 #define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
   62667   float32x4x4_t __s1 = __p1; \
   62668   float32x4x4_t __rev1; \
   62669   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62670   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62671   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62672   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   62673   __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
   62674 })
   62675 #endif
   62676 
   62677 #ifdef __LITTLE_ENDIAN__
   62678 #define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
   62679   float16x8x4_t __s1 = __p1; \
   62680   __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
   62681 })
   62682 #else
   62683 #define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
   62684   float16x8x4_t __s1 = __p1; \
   62685   float16x8x4_t __rev1; \
   62686   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62687   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62688   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62689   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   62690   __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
   62691 })
   62692 #endif
   62693 
   62694 #ifdef __LITTLE_ENDIAN__
   62695 #define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
   62696   int32x4x4_t __s1 = __p1; \
   62697   __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
   62698 })
   62699 #else
   62700 #define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
   62701   int32x4x4_t __s1 = __p1; \
   62702   int32x4x4_t __rev1; \
   62703   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62704   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62705   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62706   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   62707   __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
   62708 })
   62709 #endif
   62710 
   62711 #ifdef __LITTLE_ENDIAN__
   62712 #define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
   62713   int64x2x4_t __s1 = __p1; \
   62714   __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
   62715 })
   62716 #else
   62717 #define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
   62718   int64x2x4_t __s1 = __p1; \
   62719   int64x2x4_t __rev1; \
   62720   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62721   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62722   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62723   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   62724   __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
   62725 })
   62726 #endif
   62727 
   62728 #ifdef __LITTLE_ENDIAN__
   62729 #define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
   62730   int16x8x4_t __s1 = __p1; \
   62731   __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
   62732 })
   62733 #else
   62734 #define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
   62735   int16x8x4_t __s1 = __p1; \
   62736   int16x8x4_t __rev1; \
   62737   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62738   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62739   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62740   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   62741   __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
   62742 })
   62743 #endif
   62744 
   62745 #ifdef __LITTLE_ENDIAN__
   62746 #define vst1_u8_x4(__p0, __p1) __extension__ ({ \
   62747   uint8x8x4_t __s1 = __p1; \
   62748   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
   62749 })
   62750 #else
   62751 #define vst1_u8_x4(__p0, __p1) __extension__ ({ \
   62752   uint8x8x4_t __s1 = __p1; \
   62753   uint8x8x4_t __rev1; \
   62754   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62755   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62756   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62757   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   62758   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
   62759 })
   62760 #endif
   62761 
   62762 #ifdef __LITTLE_ENDIAN__
   62763 #define vst1_u32_x4(__p0, __p1) __extension__ ({ \
   62764   uint32x2x4_t __s1 = __p1; \
   62765   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
   62766 })
   62767 #else
   62768 #define vst1_u32_x4(__p0, __p1) __extension__ ({ \
   62769   uint32x2x4_t __s1 = __p1; \
   62770   uint32x2x4_t __rev1; \
   62771   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62772   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62773   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62774   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   62775   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
   62776 })
   62777 #endif
   62778 
   62779 #ifdef __LITTLE_ENDIAN__
   62780 #define vst1_u64_x4(__p0, __p1) __extension__ ({ \
   62781   uint64x1x4_t __s1 = __p1; \
   62782   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
   62783 })
   62784 #else
   62785 #define vst1_u64_x4(__p0, __p1) __extension__ ({ \
   62786   uint64x1x4_t __s1 = __p1; \
   62787   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
   62788 })
   62789 #endif
   62790 
   62791 #ifdef __LITTLE_ENDIAN__
   62792 #define vst1_u16_x4(__p0, __p1) __extension__ ({ \
   62793   uint16x4x4_t __s1 = __p1; \
   62794   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
   62795 })
   62796 #else
   62797 #define vst1_u16_x4(__p0, __p1) __extension__ ({ \
   62798   uint16x4x4_t __s1 = __p1; \
   62799   uint16x4x4_t __rev1; \
   62800   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62801   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62802   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62803   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   62804   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
   62805 })
   62806 #endif
   62807 
   62808 #ifdef __LITTLE_ENDIAN__
   62809 #define vst1_s8_x4(__p0, __p1) __extension__ ({ \
   62810   int8x8x4_t __s1 = __p1; \
   62811   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
   62812 })
   62813 #else
   62814 #define vst1_s8_x4(__p0, __p1) __extension__ ({ \
   62815   int8x8x4_t __s1 = __p1; \
   62816   int8x8x4_t __rev1; \
   62817   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
   62818   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
   62819   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
   62820   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
   62821   __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
   62822 })
   62823 #endif
   62824 
   62825 #ifdef __LITTLE_ENDIAN__
   62826 #define vst1_f64_x4(__p0, __p1) __extension__ ({ \
   62827   float64x1x4_t __s1 = __p1; \
   62828   __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
   62829 })
   62830 #else
   62831 #define vst1_f64_x4(__p0, __p1) __extension__ ({ \
   62832   float64x1x4_t __s1 = __p1; \
   62833   __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
   62834 })
   62835 #endif
   62836 
   62837 #ifdef __LITTLE_ENDIAN__
   62838 #define vst1_f32_x4(__p0, __p1) __extension__ ({ \
   62839   float32x2x4_t __s1 = __p1; \
   62840   __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
   62841 })
   62842 #else
   62843 #define vst1_f32_x4(__p0, __p1) __extension__ ({ \
   62844   float32x2x4_t __s1 = __p1; \
   62845   float32x2x4_t __rev1; \
   62846   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62847   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62848   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62849   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   62850   __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
   62851 })
   62852 #endif
   62853 
   62854 #ifdef __LITTLE_ENDIAN__
   62855 #define vst1_f16_x4(__p0, __p1) __extension__ ({ \
   62856   float16x4x4_t __s1 = __p1; \
   62857   __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
   62858 })
   62859 #else
   62860 #define vst1_f16_x4(__p0, __p1) __extension__ ({ \
   62861   float16x4x4_t __s1 = __p1; \
   62862   float16x4x4_t __rev1; \
   62863   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62864   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62865   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62866   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   62867   __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
   62868 })
   62869 #endif
   62870 
   62871 #ifdef __LITTLE_ENDIAN__
   62872 #define vst1_s32_x4(__p0, __p1) __extension__ ({ \
   62873   int32x2x4_t __s1 = __p1; \
   62874   __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
   62875 })
   62876 #else
   62877 #define vst1_s32_x4(__p0, __p1) __extension__ ({ \
   62878   int32x2x4_t __s1 = __p1; \
   62879   int32x2x4_t __rev1; \
   62880   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62881   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62882   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   62883   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   62884   __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
   62885 })
   62886 #endif
   62887 
   62888 #ifdef __LITTLE_ENDIAN__
   62889 #define vst1_s64_x4(__p0, __p1) __extension__ ({ \
   62890   int64x1x4_t __s1 = __p1; \
   62891   __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
   62892 })
   62893 #else
   62894 #define vst1_s64_x4(__p0, __p1) __extension__ ({ \
   62895   int64x1x4_t __s1 = __p1; \
   62896   __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
   62897 })
   62898 #endif
   62899 
   62900 #ifdef __LITTLE_ENDIAN__
   62901 #define vst1_s16_x4(__p0, __p1) __extension__ ({ \
   62902   int16x4x4_t __s1 = __p1; \
   62903   __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
   62904 })
   62905 #else
   62906 #define vst1_s16_x4(__p0, __p1) __extension__ ({ \
   62907   int16x4x4_t __s1 = __p1; \
   62908   int16x4x4_t __rev1; \
   62909   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
   62910   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
   62911   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
   62912   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
   62913   __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
   62914 })
   62915 #endif
   62916 
   62917 #ifdef __LITTLE_ENDIAN__
   62918 #define vst2_p64(__p0, __p1) __extension__ ({ \
   62919   poly64x1x2_t __s1 = __p1; \
   62920   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
   62921 })
   62922 #else
   62923 #define vst2_p64(__p0, __p1) __extension__ ({ \
   62924   poly64x1x2_t __s1 = __p1; \
   62925   __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
   62926 })
   62927 #endif
   62928 
   62929 #ifdef __LITTLE_ENDIAN__
   62930 #define vst2q_p64(__p0, __p1) __extension__ ({ \
   62931   poly64x2x2_t __s1 = __p1; \
   62932   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
   62933 })
   62934 #else
   62935 #define vst2q_p64(__p0, __p1) __extension__ ({ \
   62936   poly64x2x2_t __s1 = __p1; \
   62937   poly64x2x2_t __rev1; \
   62938   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62939   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62940   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
   62941 })
   62942 #endif
   62943 
   62944 #ifdef __LITTLE_ENDIAN__
   62945 #define vst2q_u64(__p0, __p1) __extension__ ({ \
   62946   uint64x2x2_t __s1 = __p1; \
   62947   __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
   62948 })
   62949 #else
   62950 #define vst2q_u64(__p0, __p1) __extension__ ({ \
   62951   uint64x2x2_t __s1 = __p1; \
   62952   uint64x2x2_t __rev1; \
   62953   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62954   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62955   __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
   62956 })
   62957 #endif
   62958 
   62959 #ifdef __LITTLE_ENDIAN__
   62960 #define vst2q_f64(__p0, __p1) __extension__ ({ \
   62961   float64x2x2_t __s1 = __p1; \
   62962   __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 42); \
   62963 })
   62964 #else
   62965 #define vst2q_f64(__p0, __p1) __extension__ ({ \
   62966   float64x2x2_t __s1 = __p1; \
   62967   float64x2x2_t __rev1; \
   62968   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62969   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62970   __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
   62971 })
   62972 #endif
   62973 
   62974 #ifdef __LITTLE_ENDIAN__
   62975 #define vst2q_s64(__p0, __p1) __extension__ ({ \
   62976   int64x2x2_t __s1 = __p1; \
   62977   __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 35); \
   62978 })
   62979 #else
   62980 #define vst2q_s64(__p0, __p1) __extension__ ({ \
   62981   int64x2x2_t __s1 = __p1; \
   62982   int64x2x2_t __rev1; \
   62983   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   62984   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   62985   __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
   62986 })
   62987 #endif
   62988 
   62989 #ifdef __LITTLE_ENDIAN__
   62990 #define vst2_f64(__p0, __p1) __extension__ ({ \
   62991   float64x1x2_t __s1 = __p1; \
   62992   __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
   62993 })
   62994 #else
   62995 #define vst2_f64(__p0, __p1) __extension__ ({ \
   62996   float64x1x2_t __s1 = __p1; \
   62997   __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
   62998 })
   62999 #endif
   63000 
   63001 #ifdef __LITTLE_ENDIAN__
   63002 #define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63003   poly64x1x2_t __s1 = __p1; \
   63004   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
   63005 })
   63006 #else
   63007 #define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63008   poly64x1x2_t __s1 = __p1; \
   63009   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
   63010 })
   63011 #endif
   63012 
   63013 #ifdef __LITTLE_ENDIAN__
   63014 #define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   63015   poly8x16x2_t __s1 = __p1; \
   63016   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
   63017 })
   63018 #else
   63019 #define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   63020   poly8x16x2_t __s1 = __p1; \
   63021   poly8x16x2_t __rev1; \
   63022   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63023   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63024   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
   63025 })
   63026 #endif
   63027 
   63028 #ifdef __LITTLE_ENDIAN__
   63029 #define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63030   poly64x2x2_t __s1 = __p1; \
   63031   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
   63032 })
   63033 #else
   63034 #define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63035   poly64x2x2_t __s1 = __p1; \
   63036   poly64x2x2_t __rev1; \
   63037   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63038   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63039   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
   63040 })
   63041 #endif
   63042 
   63043 #ifdef __LITTLE_ENDIAN__
   63044 #define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   63045   uint8x16x2_t __s1 = __p1; \
   63046   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
   63047 })
   63048 #else
   63049 #define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   63050   uint8x16x2_t __s1 = __p1; \
   63051   uint8x16x2_t __rev1; \
   63052   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63053   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63054   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
   63055 })
   63056 #endif
   63057 
   63058 #ifdef __LITTLE_ENDIAN__
   63059 #define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63060   uint64x2x2_t __s1 = __p1; \
   63061   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
   63062 })
   63063 #else
   63064 #define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63065   uint64x2x2_t __s1 = __p1; \
   63066   uint64x2x2_t __rev1; \
   63067   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63068   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63069   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
   63070 })
   63071 #endif
   63072 
   63073 #ifdef __LITTLE_ENDIAN__
   63074 #define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   63075   int8x16x2_t __s1 = __p1; \
   63076   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
   63077 })
   63078 #else
   63079 #define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   63080   int8x16x2_t __s1 = __p1; \
   63081   int8x16x2_t __rev1; \
   63082   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63083   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63084   __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
   63085 })
   63086 #endif
   63087 
   63088 #ifdef __LITTLE_ENDIAN__
   63089 #define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63090   float64x2x2_t __s1 = __p1; \
   63091   __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 42); \
   63092 })
   63093 #else
   63094 #define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63095   float64x2x2_t __s1 = __p1; \
   63096   float64x2x2_t __rev1; \
   63097   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63098   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63099   __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
   63100 })
   63101 #endif
   63102 
   63103 #ifdef __LITTLE_ENDIAN__
   63104 #define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63105   int64x2x2_t __s1 = __p1; \
   63106   __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 35); \
   63107 })
   63108 #else
   63109 #define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63110   int64x2x2_t __s1 = __p1; \
   63111   int64x2x2_t __rev1; \
   63112   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63113   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63114   __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
   63115 })
   63116 #endif
   63117 
   63118 #ifdef __LITTLE_ENDIAN__
   63119 #define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63120   uint64x1x2_t __s1 = __p1; \
   63121   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
   63122 })
   63123 #else
   63124 #define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63125   uint64x1x2_t __s1 = __p1; \
   63126   __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
   63127 })
   63128 #endif
   63129 
   63130 #ifdef __LITTLE_ENDIAN__
   63131 #define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63132   float64x1x2_t __s1 = __p1; \
   63133   __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
   63134 })
   63135 #else
   63136 #define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63137   float64x1x2_t __s1 = __p1; \
   63138   __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
   63139 })
   63140 #endif
   63141 
   63142 #ifdef __LITTLE_ENDIAN__
   63143 #define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63144   int64x1x2_t __s1 = __p1; \
   63145   __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
   63146 })
   63147 #else
   63148 #define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63149   int64x1x2_t __s1 = __p1; \
   63150   __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
   63151 })
   63152 #endif
   63153 
   63154 #ifdef __LITTLE_ENDIAN__
   63155 #define vst3_p64(__p0, __p1) __extension__ ({ \
   63156   poly64x1x3_t __s1 = __p1; \
   63157   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
   63158 })
   63159 #else
   63160 #define vst3_p64(__p0, __p1) __extension__ ({ \
   63161   poly64x1x3_t __s1 = __p1; \
   63162   __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
   63163 })
   63164 #endif
   63165 
   63166 #ifdef __LITTLE_ENDIAN__
   63167 #define vst3q_p64(__p0, __p1) __extension__ ({ \
   63168   poly64x2x3_t __s1 = __p1; \
   63169   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
   63170 })
   63171 #else
   63172 #define vst3q_p64(__p0, __p1) __extension__ ({ \
   63173   poly64x2x3_t __s1 = __p1; \
   63174   poly64x2x3_t __rev1; \
   63175   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63176   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63177   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63178   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
   63179 })
   63180 #endif
   63181 
   63182 #ifdef __LITTLE_ENDIAN__
   63183 #define vst3q_u64(__p0, __p1) __extension__ ({ \
   63184   uint64x2x3_t __s1 = __p1; \
   63185   __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
   63186 })
   63187 #else
   63188 #define vst3q_u64(__p0, __p1) __extension__ ({ \
   63189   uint64x2x3_t __s1 = __p1; \
   63190   uint64x2x3_t __rev1; \
   63191   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63192   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63193   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63194   __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
   63195 })
   63196 #endif
   63197 
   63198 #ifdef __LITTLE_ENDIAN__
   63199 #define vst3q_f64(__p0, __p1) __extension__ ({ \
   63200   float64x2x3_t __s1 = __p1; \
   63201   __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
   63202 })
   63203 #else
   63204 #define vst3q_f64(__p0, __p1) __extension__ ({ \
   63205   float64x2x3_t __s1 = __p1; \
   63206   float64x2x3_t __rev1; \
   63207   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63208   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63209   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63210   __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
   63211 })
   63212 #endif
   63213 
   63214 #ifdef __LITTLE_ENDIAN__
   63215 #define vst3q_s64(__p0, __p1) __extension__ ({ \
   63216   int64x2x3_t __s1 = __p1; \
   63217   __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
   63218 })
   63219 #else
   63220 #define vst3q_s64(__p0, __p1) __extension__ ({ \
   63221   int64x2x3_t __s1 = __p1; \
   63222   int64x2x3_t __rev1; \
   63223   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63224   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63225   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63226   __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
   63227 })
   63228 #endif
   63229 
   63230 #ifdef __LITTLE_ENDIAN__
   63231 #define vst3_f64(__p0, __p1) __extension__ ({ \
   63232   float64x1x3_t __s1 = __p1; \
   63233   __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
   63234 })
   63235 #else
   63236 #define vst3_f64(__p0, __p1) __extension__ ({ \
   63237   float64x1x3_t __s1 = __p1; \
   63238   __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
   63239 })
   63240 #endif
   63241 
   63242 #ifdef __LITTLE_ENDIAN__
   63243 #define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63244   poly64x1x3_t __s1 = __p1; \
   63245   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
   63246 })
   63247 #else
   63248 #define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63249   poly64x1x3_t __s1 = __p1; \
   63250   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
   63251 })
   63252 #endif
   63253 
   63254 #ifdef __LITTLE_ENDIAN__
   63255 #define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   63256   poly8x16x3_t __s1 = __p1; \
   63257   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
   63258 })
   63259 #else
   63260 #define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   63261   poly8x16x3_t __s1 = __p1; \
   63262   poly8x16x3_t __rev1; \
   63263   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63264   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63265   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63266   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
   63267 })
   63268 #endif
   63269 
   63270 #ifdef __LITTLE_ENDIAN__
   63271 #define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63272   poly64x2x3_t __s1 = __p1; \
   63273   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
   63274 })
   63275 #else
   63276 #define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63277   poly64x2x3_t __s1 = __p1; \
   63278   poly64x2x3_t __rev1; \
   63279   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63280   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63281   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63282   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
   63283 })
   63284 #endif
   63285 
   63286 #ifdef __LITTLE_ENDIAN__
   63287 #define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   63288   uint8x16x3_t __s1 = __p1; \
   63289   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
   63290 })
   63291 #else
   63292 #define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   63293   uint8x16x3_t __s1 = __p1; \
   63294   uint8x16x3_t __rev1; \
   63295   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63296   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63297   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63298   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
   63299 })
   63300 #endif
   63301 
   63302 #ifdef __LITTLE_ENDIAN__
   63303 #define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63304   uint64x2x3_t __s1 = __p1; \
   63305   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
   63306 })
   63307 #else
   63308 #define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63309   uint64x2x3_t __s1 = __p1; \
   63310   uint64x2x3_t __rev1; \
   63311   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63312   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63313   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63314   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
   63315 })
   63316 #endif
   63317 
   63318 #ifdef __LITTLE_ENDIAN__
   63319 #define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   63320   int8x16x3_t __s1 = __p1; \
   63321   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
   63322 })
   63323 #else
   63324 #define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   63325   int8x16x3_t __s1 = __p1; \
   63326   int8x16x3_t __rev1; \
   63327   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63328   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63329   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63330   __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
   63331 })
   63332 #endif
   63333 
   63334 #ifdef __LITTLE_ENDIAN__
   63335 #define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63336   float64x2x3_t __s1 = __p1; \
   63337   __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
   63338 })
   63339 #else
   63340 #define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63341   float64x2x3_t __s1 = __p1; \
   63342   float64x2x3_t __rev1; \
   63343   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63344   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63345   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63346   __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
   63347 })
   63348 #endif
   63349 
   63350 #ifdef __LITTLE_ENDIAN__
   63351 #define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63352   int64x2x3_t __s1 = __p1; \
   63353   __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
   63354 })
   63355 #else
   63356 #define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63357   int64x2x3_t __s1 = __p1; \
   63358   int64x2x3_t __rev1; \
   63359   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63360   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63361   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63362   __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
   63363 })
   63364 #endif
   63365 
   63366 #ifdef __LITTLE_ENDIAN__
   63367 #define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63368   uint64x1x3_t __s1 = __p1; \
   63369   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
   63370 })
   63371 #else
   63372 #define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63373   uint64x1x3_t __s1 = __p1; \
   63374   __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
   63375 })
   63376 #endif
   63377 
   63378 #ifdef __LITTLE_ENDIAN__
   63379 #define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63380   float64x1x3_t __s1 = __p1; \
   63381   __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
   63382 })
   63383 #else
   63384 #define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63385   float64x1x3_t __s1 = __p1; \
   63386   __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
   63387 })
   63388 #endif
   63389 
   63390 #ifdef __LITTLE_ENDIAN__
   63391 #define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63392   int64x1x3_t __s1 = __p1; \
   63393   __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
   63394 })
   63395 #else
   63396 #define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63397   int64x1x3_t __s1 = __p1; \
   63398   __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
   63399 })
   63400 #endif
   63401 
   63402 #ifdef __LITTLE_ENDIAN__
   63403 #define vst4_p64(__p0, __p1) __extension__ ({ \
   63404   poly64x1x4_t __s1 = __p1; \
   63405   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
   63406 })
   63407 #else
   63408 #define vst4_p64(__p0, __p1) __extension__ ({ \
   63409   poly64x1x4_t __s1 = __p1; \
   63410   __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
   63411 })
   63412 #endif
   63413 
   63414 #ifdef __LITTLE_ENDIAN__
   63415 #define vst4q_p64(__p0, __p1) __extension__ ({ \
   63416   poly64x2x4_t __s1 = __p1; \
   63417   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
   63418 })
   63419 #else
   63420 #define vst4q_p64(__p0, __p1) __extension__ ({ \
   63421   poly64x2x4_t __s1 = __p1; \
   63422   poly64x2x4_t __rev1; \
   63423   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63424   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63425   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63426   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   63427   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
   63428 })
   63429 #endif
   63430 
   63431 #ifdef __LITTLE_ENDIAN__
   63432 #define vst4q_u64(__p0, __p1) __extension__ ({ \
   63433   uint64x2x4_t __s1 = __p1; \
   63434   __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
   63435 })
   63436 #else
   63437 #define vst4q_u64(__p0, __p1) __extension__ ({ \
   63438   uint64x2x4_t __s1 = __p1; \
   63439   uint64x2x4_t __rev1; \
   63440   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63441   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63442   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63443   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   63444   __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
   63445 })
   63446 #endif
   63447 
   63448 #ifdef __LITTLE_ENDIAN__
   63449 #define vst4q_f64(__p0, __p1) __extension__ ({ \
   63450   float64x2x4_t __s1 = __p1; \
   63451   __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
   63452 })
   63453 #else
   63454 #define vst4q_f64(__p0, __p1) __extension__ ({ \
   63455   float64x2x4_t __s1 = __p1; \
   63456   float64x2x4_t __rev1; \
   63457   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63458   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63459   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63460   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   63461   __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
   63462 })
   63463 #endif
   63464 
   63465 #ifdef __LITTLE_ENDIAN__
   63466 #define vst4q_s64(__p0, __p1) __extension__ ({ \
   63467   int64x2x4_t __s1 = __p1; \
   63468   __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
   63469 })
   63470 #else
   63471 #define vst4q_s64(__p0, __p1) __extension__ ({ \
   63472   int64x2x4_t __s1 = __p1; \
   63473   int64x2x4_t __rev1; \
   63474   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63475   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63476   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63477   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   63478   __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
   63479 })
   63480 #endif
   63481 
   63482 #ifdef __LITTLE_ENDIAN__
   63483 #define vst4_f64(__p0, __p1) __extension__ ({ \
   63484   float64x1x4_t __s1 = __p1; \
   63485   __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
   63486 })
   63487 #else
   63488 #define vst4_f64(__p0, __p1) __extension__ ({ \
   63489   float64x1x4_t __s1 = __p1; \
   63490   __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
   63491 })
   63492 #endif
   63493 
   63494 #ifdef __LITTLE_ENDIAN__
   63495 #define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63496   poly64x1x4_t __s1 = __p1; \
   63497   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
   63498 })
   63499 #else
   63500 #define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63501   poly64x1x4_t __s1 = __p1; \
   63502   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
   63503 })
   63504 #endif
   63505 
   63506 #ifdef __LITTLE_ENDIAN__
   63507 #define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   63508   poly8x16x4_t __s1 = __p1; \
   63509   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
   63510 })
   63511 #else
   63512 #define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
   63513   poly8x16x4_t __s1 = __p1; \
   63514   poly8x16x4_t __rev1; \
   63515   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63516   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63517   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63518   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63519   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
   63520 })
   63521 #endif
   63522 
   63523 #ifdef __LITTLE_ENDIAN__
   63524 #define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63525   poly64x2x4_t __s1 = __p1; \
   63526   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
   63527 })
   63528 #else
   63529 #define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
   63530   poly64x2x4_t __s1 = __p1; \
   63531   poly64x2x4_t __rev1; \
   63532   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63533   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63534   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63535   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   63536   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
   63537 })
   63538 #endif
   63539 
   63540 #ifdef __LITTLE_ENDIAN__
   63541 #define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   63542   uint8x16x4_t __s1 = __p1; \
   63543   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
   63544 })
   63545 #else
   63546 #define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
   63547   uint8x16x4_t __s1 = __p1; \
   63548   uint8x16x4_t __rev1; \
   63549   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63550   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63551   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63552   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63553   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
   63554 })
   63555 #endif
   63556 
   63557 #ifdef __LITTLE_ENDIAN__
   63558 #define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63559   uint64x2x4_t __s1 = __p1; \
   63560   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
   63561 })
   63562 #else
   63563 #define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63564   uint64x2x4_t __s1 = __p1; \
   63565   uint64x2x4_t __rev1; \
   63566   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63567   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63568   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63569   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   63570   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
   63571 })
   63572 #endif
   63573 
   63574 #ifdef __LITTLE_ENDIAN__
   63575 #define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   63576   int8x16x4_t __s1 = __p1; \
   63577   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
   63578 })
   63579 #else
   63580 #define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
   63581   int8x16x4_t __s1 = __p1; \
   63582   int8x16x4_t __rev1; \
   63583   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63584   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63585   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63586   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
   63587   __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
   63588 })
   63589 #endif
   63590 
   63591 #ifdef __LITTLE_ENDIAN__
   63592 #define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63593   float64x2x4_t __s1 = __p1; \
   63594   __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
   63595 })
   63596 #else
   63597 #define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63598   float64x2x4_t __s1 = __p1; \
   63599   float64x2x4_t __rev1; \
   63600   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63601   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63602   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63603   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   63604   __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
   63605 })
   63606 #endif
   63607 
   63608 #ifdef __LITTLE_ENDIAN__
   63609 #define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63610   int64x2x4_t __s1 = __p1; \
   63611   __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
   63612 })
   63613 #else
   63614 #define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63615   int64x2x4_t __s1 = __p1; \
   63616   int64x2x4_t __rev1; \
   63617   __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
   63618   __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
   63619   __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
   63620   __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
   63621   __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
   63622 })
   63623 #endif
   63624 
   63625 #ifdef __LITTLE_ENDIAN__
   63626 #define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63627   uint64x1x4_t __s1 = __p1; \
   63628   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
   63629 })
   63630 #else
   63631 #define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
   63632   uint64x1x4_t __s1 = __p1; \
   63633   __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
   63634 })
   63635 #endif
   63636 
   63637 #ifdef __LITTLE_ENDIAN__
   63638 #define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63639   float64x1x4_t __s1 = __p1; \
   63640   __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
   63641 })
   63642 #else
   63643 #define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
   63644   float64x1x4_t __s1 = __p1; \
   63645   __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
   63646 })
   63647 #endif
   63648 
   63649 #ifdef __LITTLE_ENDIAN__
   63650 #define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63651   int64x1x4_t __s1 = __p1; \
   63652   __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
   63653 })
   63654 #else
   63655 #define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
   63656   int64x1x4_t __s1 = __p1; \
   63657   __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
   63658 })
   63659 #endif
   63660 
   63661 #ifdef __LITTLE_ENDIAN__
   63662 #define vstrq_p128(__p0, __p1) __extension__ ({ \
   63663   poly128_t __s1 = __p1; \
   63664   __builtin_neon_vstrq_p128(__p0, __s1); \
   63665 })
   63666 #else
   63667 #define vstrq_p128(__p0, __p1) __extension__ ({ \
   63668   poly128_t __s1 = __p1; \
   63669   __builtin_neon_vstrq_p128(__p0, __s1); \
   63670 })
   63671 #endif
   63672 
   63673 #ifdef __LITTLE_ENDIAN__
   63674 __ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
   63675   uint64_t __ret;
   63676   __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
   63677   return __ret;
   63678 }
   63679 #else
   63680 __ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
   63681   uint64_t __ret;
   63682   __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
   63683   return __ret;
   63684 }
   63685 #endif
   63686 
   63687 #ifdef __LITTLE_ENDIAN__
   63688 __ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
   63689   int64_t __ret;
   63690   __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
   63691   return __ret;
   63692 }
   63693 #else
   63694 __ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
   63695   int64_t __ret;
   63696   __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
   63697   return __ret;
   63698 }
   63699 #endif
   63700 
   63701 #ifdef __LITTLE_ENDIAN__
   63702 __ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
   63703   float64x2_t __ret;
   63704   __ret = __p0 - __p1;
   63705   return __ret;
   63706 }
   63707 #else
   63708 __ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
   63709   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   63710   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   63711   float64x2_t __ret;
   63712   __ret = __rev0 - __rev1;
   63713   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   63714   return __ret;
   63715 }
   63716 #endif
   63717 
   63718 #ifdef __LITTLE_ENDIAN__
   63719 __ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
   63720   float64x1_t __ret;
   63721   __ret = __p0 - __p1;
   63722   return __ret;
   63723 }
   63724 #else
   63725 __ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
   63726   float64x1_t __ret;
   63727   __ret = __p0 - __p1;
   63728   return __ret;
   63729 }
   63730 #endif
   63731 
   63732 #ifdef __LITTLE_ENDIAN__
   63733 __ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   63734   uint16x8_t __ret;
   63735   __ret = vcombine_u16(__p0, vsubhn_u32(__p1, __p2));
   63736   return __ret;
   63737 }
   63738 #else
   63739 __ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   63740   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   63741   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   63742   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   63743   uint16x8_t __ret;
   63744   __ret = __noswap_vcombine_u16(__rev0, __noswap_vsubhn_u32(__rev1, __rev2));
   63745   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   63746   return __ret;
   63747 }
   63748 #endif
   63749 
   63750 #ifdef __LITTLE_ENDIAN__
   63751 __ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   63752   uint32x4_t __ret;
   63753   __ret = vcombine_u32(__p0, vsubhn_u64(__p1, __p2));
   63754   return __ret;
   63755 }
   63756 #else
   63757 __ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
   63758   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   63759   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   63760   uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   63761   uint32x4_t __ret;
   63762   __ret = __noswap_vcombine_u32(__rev0, __noswap_vsubhn_u64(__rev1, __rev2));
   63763   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   63764   return __ret;
   63765 }
   63766 #endif
   63767 
   63768 #ifdef __LITTLE_ENDIAN__
   63769 __ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   63770   uint8x16_t __ret;
   63771   __ret = vcombine_u8(__p0, vsubhn_u16(__p1, __p2));
   63772   return __ret;
   63773 }
   63774 #else
   63775 __ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   63776   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   63777   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   63778   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   63779   uint8x16_t __ret;
   63780   __ret = __noswap_vcombine_u8(__rev0, __noswap_vsubhn_u16(__rev1, __rev2));
   63781   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   63782   return __ret;
   63783 }
   63784 #endif
   63785 
   63786 #ifdef __LITTLE_ENDIAN__
   63787 __ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   63788   int16x8_t __ret;
   63789   __ret = vcombine_s16(__p0, vsubhn_s32(__p1, __p2));
   63790   return __ret;
   63791 }
   63792 #else
   63793 __ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   63794   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   63795   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   63796   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   63797   int16x8_t __ret;
   63798   __ret = __noswap_vcombine_s16(__rev0, __noswap_vsubhn_s32(__rev1, __rev2));
   63799   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   63800   return __ret;
   63801 }
   63802 #endif
   63803 
   63804 #ifdef __LITTLE_ENDIAN__
   63805 __ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   63806   int32x4_t __ret;
   63807   __ret = vcombine_s32(__p0, vsubhn_s64(__p1, __p2));
   63808   return __ret;
   63809 }
   63810 #else
   63811 __ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
   63812   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   63813   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   63814   int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   63815   int32x4_t __ret;
   63816   __ret = __noswap_vcombine_s32(__rev0, __noswap_vsubhn_s64(__rev1, __rev2));
   63817   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   63818   return __ret;
   63819 }
   63820 #endif
   63821 
   63822 #ifdef __LITTLE_ENDIAN__
   63823 __ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   63824   int8x16_t __ret;
   63825   __ret = vcombine_s8(__p0, vsubhn_s16(__p1, __p2));
   63826   return __ret;
   63827 }
   63828 #else
   63829 __ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   63830   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   63831   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   63832   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   63833   int8x16_t __ret;
   63834   __ret = __noswap_vcombine_s8(__rev0, __noswap_vsubhn_s16(__rev1, __rev2));
   63835   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   63836   return __ret;
   63837 }
   63838 #endif
   63839 
   63840 #ifdef __LITTLE_ENDIAN__
   63841 __ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
   63842   uint16x8_t __ret;
   63843   __ret = vmovl_high_u8(__p0) - vmovl_high_u8(__p1);
   63844   return __ret;
   63845 }
   63846 #else
   63847 __ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
   63848   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   63849   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   63850   uint16x8_t __ret;
   63851   __ret = __noswap_vmovl_high_u8(__rev0) - __noswap_vmovl_high_u8(__rev1);
   63852   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   63853   return __ret;
   63854 }
   63855 #endif
   63856 
   63857 #ifdef __LITTLE_ENDIAN__
   63858 __ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
   63859   uint64x2_t __ret;
   63860   __ret = vmovl_high_u32(__p0) - vmovl_high_u32(__p1);
   63861   return __ret;
   63862 }
   63863 #else
   63864 __ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
   63865   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   63866   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   63867   uint64x2_t __ret;
   63868   __ret = __noswap_vmovl_high_u32(__rev0) - __noswap_vmovl_high_u32(__rev1);
   63869   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   63870   return __ret;
   63871 }
   63872 #endif
   63873 
   63874 #ifdef __LITTLE_ENDIAN__
   63875 __ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
   63876   uint32x4_t __ret;
   63877   __ret = vmovl_high_u16(__p0) - vmovl_high_u16(__p1);
   63878   return __ret;
   63879 }
   63880 #else
   63881 __ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
   63882   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   63883   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   63884   uint32x4_t __ret;
   63885   __ret = __noswap_vmovl_high_u16(__rev0) - __noswap_vmovl_high_u16(__rev1);
   63886   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   63887   return __ret;
   63888 }
   63889 #endif
   63890 
   63891 #ifdef __LITTLE_ENDIAN__
   63892 __ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
   63893   int16x8_t __ret;
   63894   __ret = vmovl_high_s8(__p0) - vmovl_high_s8(__p1);
   63895   return __ret;
   63896 }
   63897 #else
   63898 __ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
   63899   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   63900   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   63901   int16x8_t __ret;
   63902   __ret = __noswap_vmovl_high_s8(__rev0) - __noswap_vmovl_high_s8(__rev1);
   63903   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   63904   return __ret;
   63905 }
   63906 #endif
   63907 
   63908 #ifdef __LITTLE_ENDIAN__
   63909 __ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
   63910   int64x2_t __ret;
   63911   __ret = vmovl_high_s32(__p0) - vmovl_high_s32(__p1);
   63912   return __ret;
   63913 }
   63914 #else
   63915 __ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
   63916   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   63917   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   63918   int64x2_t __ret;
   63919   __ret = __noswap_vmovl_high_s32(__rev0) - __noswap_vmovl_high_s32(__rev1);
   63920   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   63921   return __ret;
   63922 }
   63923 #endif
   63924 
   63925 #ifdef __LITTLE_ENDIAN__
   63926 __ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
   63927   int32x4_t __ret;
   63928   __ret = vmovl_high_s16(__p0) - vmovl_high_s16(__p1);
   63929   return __ret;
   63930 }
   63931 #else
   63932 __ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
   63933   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   63934   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   63935   int32x4_t __ret;
   63936   __ret = __noswap_vmovl_high_s16(__rev0) - __noswap_vmovl_high_s16(__rev1);
   63937   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   63938   return __ret;
   63939 }
   63940 #endif
   63941 
   63942 #ifdef __LITTLE_ENDIAN__
   63943 __ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
   63944   uint16x8_t __ret;
   63945   __ret = __p0 - vmovl_high_u8(__p1);
   63946   return __ret;
   63947 }
   63948 #else
   63949 __ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
   63950   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   63951   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   63952   uint16x8_t __ret;
   63953   __ret = __rev0 - __noswap_vmovl_high_u8(__rev1);
   63954   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   63955   return __ret;
   63956 }
   63957 #endif
   63958 
   63959 #ifdef __LITTLE_ENDIAN__
   63960 __ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
   63961   uint64x2_t __ret;
   63962   __ret = __p0 - vmovl_high_u32(__p1);
   63963   return __ret;
   63964 }
   63965 #else
   63966 __ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
   63967   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   63968   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   63969   uint64x2_t __ret;
   63970   __ret = __rev0 - __noswap_vmovl_high_u32(__rev1);
   63971   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   63972   return __ret;
   63973 }
   63974 #endif
   63975 
   63976 #ifdef __LITTLE_ENDIAN__
   63977 __ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
   63978   uint32x4_t __ret;
   63979   __ret = __p0 - vmovl_high_u16(__p1);
   63980   return __ret;
   63981 }
   63982 #else
   63983 __ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
   63984   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   63985   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   63986   uint32x4_t __ret;
   63987   __ret = __rev0 - __noswap_vmovl_high_u16(__rev1);
   63988   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   63989   return __ret;
   63990 }
   63991 #endif
   63992 
   63993 #ifdef __LITTLE_ENDIAN__
   63994 __ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
   63995   int16x8_t __ret;
   63996   __ret = __p0 - vmovl_high_s8(__p1);
   63997   return __ret;
   63998 }
   63999 #else
   64000 __ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
   64001   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64002   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64003   int16x8_t __ret;
   64004   __ret = __rev0 - __noswap_vmovl_high_s8(__rev1);
   64005   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64006   return __ret;
   64007 }
   64008 #endif
   64009 
   64010 #ifdef __LITTLE_ENDIAN__
   64011 __ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
   64012   int64x2_t __ret;
   64013   __ret = __p0 - vmovl_high_s32(__p1);
   64014   return __ret;
   64015 }
   64016 #else
   64017 __ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
   64018   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64019   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64020   int64x2_t __ret;
   64021   __ret = __rev0 - __noswap_vmovl_high_s32(__rev1);
   64022   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64023   return __ret;
   64024 }
   64025 #endif
   64026 
   64027 #ifdef __LITTLE_ENDIAN__
   64028 __ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
   64029   int32x4_t __ret;
   64030   __ret = __p0 - vmovl_high_s16(__p1);
   64031   return __ret;
   64032 }
   64033 #else
   64034 __ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
   64035   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64036   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64037   int32x4_t __ret;
   64038   __ret = __rev0 - __noswap_vmovl_high_s16(__rev1);
   64039   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64040   return __ret;
   64041 }
   64042 #endif
   64043 
   64044 #ifdef __LITTLE_ENDIAN__
   64045 __ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
   64046   poly8x8_t __ret;
   64047   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
   64048   return __ret;
   64049 }
   64050 #else
   64051 __ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
   64052   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64053   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64054   poly8x8_t __ret;
   64055   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
   64056   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64057   return __ret;
   64058 }
   64059 #endif
   64060 
   64061 #ifdef __LITTLE_ENDIAN__
   64062 __ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
   64063   poly16x4_t __ret;
   64064   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
   64065   return __ret;
   64066 }
   64067 #else
   64068 __ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
   64069   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64070   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64071   poly16x4_t __ret;
   64072   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
   64073   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64074   return __ret;
   64075 }
   64076 #endif
   64077 
   64078 #ifdef __LITTLE_ENDIAN__
   64079 __ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   64080   poly8x16_t __ret;
   64081   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
   64082   return __ret;
   64083 }
   64084 #else
   64085 __ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   64086   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64087   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64088   poly8x16_t __ret;
   64089   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
   64090   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64091   return __ret;
   64092 }
   64093 #endif
   64094 
   64095 #ifdef __LITTLE_ENDIAN__
   64096 __ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   64097   poly64x2_t __ret;
   64098   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   64099   return __ret;
   64100 }
   64101 #else
   64102 __ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   64103   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64104   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64105   poly64x2_t __ret;
   64106   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   64107   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64108   return __ret;
   64109 }
   64110 #endif
   64111 
   64112 #ifdef __LITTLE_ENDIAN__
   64113 __ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   64114   poly16x8_t __ret;
   64115   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
   64116   return __ret;
   64117 }
   64118 #else
   64119 __ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   64120   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64121   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64122   poly16x8_t __ret;
   64123   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
   64124   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64125   return __ret;
   64126 }
   64127 #endif
   64128 
   64129 #ifdef __LITTLE_ENDIAN__
   64130 __ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   64131   uint8x16_t __ret;
   64132   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
   64133   return __ret;
   64134 }
   64135 #else
   64136 __ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   64137   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64138   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64139   uint8x16_t __ret;
   64140   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
   64141   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64142   return __ret;
   64143 }
   64144 #endif
   64145 
   64146 #ifdef __LITTLE_ENDIAN__
   64147 __ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   64148   uint32x4_t __ret;
   64149   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
   64150   return __ret;
   64151 }
   64152 #else
   64153 __ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   64154   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64155   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64156   uint32x4_t __ret;
   64157   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
   64158   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64159   return __ret;
   64160 }
   64161 #endif
   64162 
   64163 #ifdef __LITTLE_ENDIAN__
   64164 __ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   64165   uint64x2_t __ret;
   64166   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   64167   return __ret;
   64168 }
   64169 #else
   64170 __ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   64171   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64172   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64173   uint64x2_t __ret;
   64174   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   64175   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64176   return __ret;
   64177 }
   64178 #endif
   64179 
   64180 #ifdef __LITTLE_ENDIAN__
   64181 __ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   64182   uint16x8_t __ret;
   64183   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
   64184   return __ret;
   64185 }
   64186 #else
   64187 __ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   64188   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64189   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64190   uint16x8_t __ret;
   64191   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
   64192   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64193   return __ret;
   64194 }
   64195 #endif
   64196 
   64197 #ifdef __LITTLE_ENDIAN__
   64198 __ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
   64199   int8x16_t __ret;
   64200   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
   64201   return __ret;
   64202 }
   64203 #else
   64204 __ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
   64205   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64206   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64207   int8x16_t __ret;
   64208   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
   64209   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64210   return __ret;
   64211 }
   64212 #endif
   64213 
   64214 #ifdef __LITTLE_ENDIAN__
   64215 __ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
   64216   float64x2_t __ret;
   64217   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   64218   return __ret;
   64219 }
   64220 #else
   64221 __ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
   64222   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64223   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64224   float64x2_t __ret;
   64225   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   64226   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64227   return __ret;
   64228 }
   64229 #endif
   64230 
   64231 #ifdef __LITTLE_ENDIAN__
   64232 __ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
   64233   float32x4_t __ret;
   64234   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
   64235   return __ret;
   64236 }
   64237 #else
   64238 __ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
   64239   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64240   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64241   float32x4_t __ret;
   64242   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
   64243   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64244   return __ret;
   64245 }
   64246 #endif
   64247 
   64248 #ifdef __LITTLE_ENDIAN__
   64249 __ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
   64250   int32x4_t __ret;
   64251   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
   64252   return __ret;
   64253 }
   64254 #else
   64255 __ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
   64256   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64257   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64258   int32x4_t __ret;
   64259   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
   64260   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64261   return __ret;
   64262 }
   64263 #endif
   64264 
   64265 #ifdef __LITTLE_ENDIAN__
   64266 __ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
   64267   int64x2_t __ret;
   64268   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   64269   return __ret;
   64270 }
   64271 #else
   64272 __ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
   64273   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64274   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64275   int64x2_t __ret;
   64276   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   64277   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64278   return __ret;
   64279 }
   64280 #endif
   64281 
   64282 #ifdef __LITTLE_ENDIAN__
   64283 __ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
   64284   int16x8_t __ret;
   64285   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
   64286   return __ret;
   64287 }
   64288 #else
   64289 __ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
   64290   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64291   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64292   int16x8_t __ret;
   64293   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
   64294   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64295   return __ret;
   64296 }
   64297 #endif
   64298 
   64299 #ifdef __LITTLE_ENDIAN__
   64300 __ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
   64301   uint8x8_t __ret;
   64302   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
   64303   return __ret;
   64304 }
   64305 #else
   64306 __ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
   64307   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64308   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64309   uint8x8_t __ret;
   64310   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
   64311   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64312   return __ret;
   64313 }
   64314 #endif
   64315 
   64316 #ifdef __LITTLE_ENDIAN__
   64317 __ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
   64318   uint32x2_t __ret;
   64319   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   64320   return __ret;
   64321 }
   64322 #else
   64323 __ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
   64324   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64325   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64326   uint32x2_t __ret;
   64327   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   64328   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64329   return __ret;
   64330 }
   64331 #endif
   64332 
   64333 #ifdef __LITTLE_ENDIAN__
   64334 __ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
   64335   uint16x4_t __ret;
   64336   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
   64337   return __ret;
   64338 }
   64339 #else
   64340 __ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
   64341   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64342   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64343   uint16x4_t __ret;
   64344   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
   64345   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64346   return __ret;
   64347 }
   64348 #endif
   64349 
   64350 #ifdef __LITTLE_ENDIAN__
   64351 __ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
   64352   int8x8_t __ret;
   64353   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
   64354   return __ret;
   64355 }
   64356 #else
   64357 __ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
   64358   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64359   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64360   int8x8_t __ret;
   64361   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
   64362   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64363   return __ret;
   64364 }
   64365 #endif
   64366 
   64367 #ifdef __LITTLE_ENDIAN__
   64368 __ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
   64369   float32x2_t __ret;
   64370   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   64371   return __ret;
   64372 }
   64373 #else
   64374 __ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
   64375   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64376   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64377   float32x2_t __ret;
   64378   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   64379   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64380   return __ret;
   64381 }
   64382 #endif
   64383 
   64384 #ifdef __LITTLE_ENDIAN__
   64385 __ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
   64386   int32x2_t __ret;
   64387   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   64388   return __ret;
   64389 }
   64390 #else
   64391 __ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
   64392   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64393   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64394   int32x2_t __ret;
   64395   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   64396   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64397   return __ret;
   64398 }
   64399 #endif
   64400 
   64401 #ifdef __LITTLE_ENDIAN__
   64402 __ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
   64403   int16x4_t __ret;
   64404   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
   64405   return __ret;
   64406 }
   64407 #else
   64408 __ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
   64409   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64410   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64411   int16x4_t __ret;
   64412   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
   64413   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64414   return __ret;
   64415 }
   64416 #endif
   64417 
   64418 #ifdef __LITTLE_ENDIAN__
   64419 __ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
   64420   poly8x8_t __ret;
   64421   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
   64422   return __ret;
   64423 }
   64424 #else
   64425 __ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
   64426   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64427   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64428   poly8x8_t __ret;
   64429   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
   64430   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64431   return __ret;
   64432 }
   64433 #endif
   64434 
   64435 #ifdef __LITTLE_ENDIAN__
   64436 __ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
   64437   poly16x4_t __ret;
   64438   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
   64439   return __ret;
   64440 }
   64441 #else
   64442 __ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
   64443   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64444   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64445   poly16x4_t __ret;
   64446   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
   64447   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64448   return __ret;
   64449 }
   64450 #endif
   64451 
   64452 #ifdef __LITTLE_ENDIAN__
   64453 __ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   64454   poly8x16_t __ret;
   64455   __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
   64456   return __ret;
   64457 }
   64458 #else
   64459 __ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   64460   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64461   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64462   poly8x16_t __ret;
   64463   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
   64464   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64465   return __ret;
   64466 }
   64467 #endif
   64468 
   64469 #ifdef __LITTLE_ENDIAN__
   64470 __ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   64471   poly64x2_t __ret;
   64472   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   64473   return __ret;
   64474 }
   64475 #else
   64476 __ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   64477   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64478   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64479   poly64x2_t __ret;
   64480   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   64481   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64482   return __ret;
   64483 }
   64484 #endif
   64485 
   64486 #ifdef __LITTLE_ENDIAN__
   64487 __ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   64488   poly16x8_t __ret;
   64489   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
   64490   return __ret;
   64491 }
   64492 #else
   64493 __ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   64494   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64495   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64496   poly16x8_t __ret;
   64497   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
   64498   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64499   return __ret;
   64500 }
   64501 #endif
   64502 
   64503 #ifdef __LITTLE_ENDIAN__
   64504 __ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   64505   uint8x16_t __ret;
   64506   __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
   64507   return __ret;
   64508 }
   64509 #else
   64510 __ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   64511   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64512   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64513   uint8x16_t __ret;
   64514   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
   64515   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64516   return __ret;
   64517 }
   64518 #endif
   64519 
   64520 #ifdef __LITTLE_ENDIAN__
   64521 __ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   64522   uint32x4_t __ret;
   64523   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
   64524   return __ret;
   64525 }
   64526 #else
   64527 __ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   64528   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64529   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64530   uint32x4_t __ret;
   64531   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
   64532   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64533   return __ret;
   64534 }
   64535 #endif
   64536 
   64537 #ifdef __LITTLE_ENDIAN__
   64538 __ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   64539   uint64x2_t __ret;
   64540   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   64541   return __ret;
   64542 }
   64543 #else
   64544 __ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   64545   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64546   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64547   uint64x2_t __ret;
   64548   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   64549   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64550   return __ret;
   64551 }
   64552 #endif
   64553 
   64554 #ifdef __LITTLE_ENDIAN__
   64555 __ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   64556   uint16x8_t __ret;
   64557   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
   64558   return __ret;
   64559 }
   64560 #else
   64561 __ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   64562   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64563   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64564   uint16x8_t __ret;
   64565   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
   64566   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64567   return __ret;
   64568 }
   64569 #endif
   64570 
   64571 #ifdef __LITTLE_ENDIAN__
   64572 __ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
   64573   int8x16_t __ret;
   64574   __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
   64575   return __ret;
   64576 }
   64577 #else
   64578 __ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
   64579   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64580   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64581   int8x16_t __ret;
   64582   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
   64583   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64584   return __ret;
   64585 }
   64586 #endif
   64587 
   64588 #ifdef __LITTLE_ENDIAN__
   64589 __ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
   64590   float64x2_t __ret;
   64591   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   64592   return __ret;
   64593 }
   64594 #else
   64595 __ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
   64596   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64597   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64598   float64x2_t __ret;
   64599   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   64600   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64601   return __ret;
   64602 }
   64603 #endif
   64604 
   64605 #ifdef __LITTLE_ENDIAN__
   64606 __ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
   64607   float32x4_t __ret;
   64608   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
   64609   return __ret;
   64610 }
   64611 #else
   64612 __ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
   64613   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64614   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64615   float32x4_t __ret;
   64616   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
   64617   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64618   return __ret;
   64619 }
   64620 #endif
   64621 
   64622 #ifdef __LITTLE_ENDIAN__
   64623 __ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
   64624   int32x4_t __ret;
   64625   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
   64626   return __ret;
   64627 }
   64628 #else
   64629 __ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
   64630   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64631   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64632   int32x4_t __ret;
   64633   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
   64634   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64635   return __ret;
   64636 }
   64637 #endif
   64638 
   64639 #ifdef __LITTLE_ENDIAN__
   64640 __ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
   64641   int64x2_t __ret;
   64642   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   64643   return __ret;
   64644 }
   64645 #else
   64646 __ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
   64647   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64648   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64649   int64x2_t __ret;
   64650   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   64651   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64652   return __ret;
   64653 }
   64654 #endif
   64655 
   64656 #ifdef __LITTLE_ENDIAN__
   64657 __ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
   64658   int16x8_t __ret;
   64659   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
   64660   return __ret;
   64661 }
   64662 #else
   64663 __ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
   64664   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64665   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64666   int16x8_t __ret;
   64667   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
   64668   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64669   return __ret;
   64670 }
   64671 #endif
   64672 
   64673 #ifdef __LITTLE_ENDIAN__
   64674 __ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
   64675   uint8x8_t __ret;
   64676   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
   64677   return __ret;
   64678 }
   64679 #else
   64680 __ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
   64681   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64682   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64683   uint8x8_t __ret;
   64684   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
   64685   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64686   return __ret;
   64687 }
   64688 #endif
   64689 
   64690 #ifdef __LITTLE_ENDIAN__
   64691 __ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
   64692   uint32x2_t __ret;
   64693   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   64694   return __ret;
   64695 }
   64696 #else
   64697 __ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
   64698   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64699   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64700   uint32x2_t __ret;
   64701   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   64702   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64703   return __ret;
   64704 }
   64705 #endif
   64706 
   64707 #ifdef __LITTLE_ENDIAN__
   64708 __ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
   64709   uint16x4_t __ret;
   64710   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
   64711   return __ret;
   64712 }
   64713 #else
   64714 __ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
   64715   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64716   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64717   uint16x4_t __ret;
   64718   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
   64719   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64720   return __ret;
   64721 }
   64722 #endif
   64723 
   64724 #ifdef __LITTLE_ENDIAN__
   64725 __ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
   64726   int8x8_t __ret;
   64727   __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
   64728   return __ret;
   64729 }
   64730 #else
   64731 __ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
   64732   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   64733   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   64734   int8x8_t __ret;
   64735   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
   64736   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   64737   return __ret;
   64738 }
   64739 #endif
   64740 
   64741 #ifdef __LITTLE_ENDIAN__
   64742 __ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
   64743   float32x2_t __ret;
   64744   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   64745   return __ret;
   64746 }
   64747 #else
   64748 __ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
   64749   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64750   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64751   float32x2_t __ret;
   64752   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   64753   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64754   return __ret;
   64755 }
   64756 #endif
   64757 
   64758 #ifdef __LITTLE_ENDIAN__
   64759 __ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
   64760   int32x2_t __ret;
   64761   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   64762   return __ret;
   64763 }
   64764 #else
   64765 __ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
   64766   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64767   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64768   int32x2_t __ret;
   64769   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   64770   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64771   return __ret;
   64772 }
   64773 #endif
   64774 
   64775 #ifdef __LITTLE_ENDIAN__
   64776 __ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
   64777   int16x4_t __ret;
   64778   __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
   64779   return __ret;
   64780 }
   64781 #else
   64782 __ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
   64783   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64784   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64785   int16x4_t __ret;
   64786   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
   64787   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64788   return __ret;
   64789 }
   64790 #endif
   64791 
   64792 #ifdef __LITTLE_ENDIAN__
   64793 __ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
   64794   uint64x1_t __ret;
   64795   __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   64796   return __ret;
   64797 }
   64798 #else
   64799 __ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
   64800   uint64x1_t __ret;
   64801   __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   64802   return __ret;
   64803 }
   64804 #endif
   64805 
   64806 #ifdef __LITTLE_ENDIAN__
   64807 __ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
   64808   uint64x2_t __ret;
   64809   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   64810   return __ret;
   64811 }
   64812 #else
   64813 __ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
   64814   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64815   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64816   uint64x2_t __ret;
   64817   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   64818   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64819   return __ret;
   64820 }
   64821 #endif
   64822 
   64823 #ifdef __LITTLE_ENDIAN__
   64824 __ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   64825   uint64x2_t __ret;
   64826   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   64827   return __ret;
   64828 }
   64829 #else
   64830 __ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
   64831   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64832   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64833   uint64x2_t __ret;
   64834   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   64835   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64836   return __ret;
   64837 }
   64838 #endif
   64839 
   64840 #ifdef __LITTLE_ENDIAN__
   64841 __ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
   64842   uint64x2_t __ret;
   64843   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
   64844   return __ret;
   64845 }
   64846 #else
   64847 __ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
   64848   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   64849   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   64850   uint64x2_t __ret;
   64851   __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
   64852   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   64853   return __ret;
   64854 }
   64855 #endif
   64856 
   64857 #ifdef __LITTLE_ENDIAN__
   64858 __ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
   64859   uint64x1_t __ret;
   64860   __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   64861   return __ret;
   64862 }
   64863 #else
   64864 __ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
   64865   uint64x1_t __ret;
   64866   __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   64867   return __ret;
   64868 }
   64869 #endif
   64870 
   64871 #ifdef __LITTLE_ENDIAN__
   64872 __ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
   64873   uint64x1_t __ret;
   64874   __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   64875   return __ret;
   64876 }
   64877 #else
   64878 __ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
   64879   uint64x1_t __ret;
   64880   __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
   64881   return __ret;
   64882 }
   64883 #endif
   64884 
   64885 #ifdef __LITTLE_ENDIAN__
   64886 __ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
   64887   uint64_t __ret;
   64888   __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
   64889   return __ret;
   64890 }
   64891 #else
   64892 __ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
   64893   uint64_t __ret;
   64894   __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
   64895   return __ret;
   64896 }
   64897 #endif
   64898 
   64899 #ifdef __LITTLE_ENDIAN__
   64900 __ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
   64901   int64_t __ret;
   64902   __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
   64903   return __ret;
   64904 }
   64905 #else
   64906 __ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
   64907   int64_t __ret;
   64908   __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
   64909   return __ret;
   64910 }
   64911 #endif
   64912 
   64913 #ifdef __LITTLE_ENDIAN__
   64914 __ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
   64915   int8_t __ret;
   64916   __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
   64917   return __ret;
   64918 }
   64919 #else
   64920 __ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
   64921   int8_t __ret;
   64922   __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
   64923   return __ret;
   64924 }
   64925 #endif
   64926 
   64927 #ifdef __LITTLE_ENDIAN__
   64928 __ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
   64929   int32_t __ret;
   64930   __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
   64931   return __ret;
   64932 }
   64933 #else
   64934 __ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
   64935   int32_t __ret;
   64936   __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
   64937   return __ret;
   64938 }
   64939 #endif
   64940 
   64941 #ifdef __LITTLE_ENDIAN__
   64942 __ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
   64943   int64_t __ret;
   64944   __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
   64945   return __ret;
   64946 }
   64947 #else
   64948 __ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
   64949   int64_t __ret;
   64950   __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
   64951   return __ret;
   64952 }
   64953 #endif
   64954 
   64955 #ifdef __LITTLE_ENDIAN__
   64956 __ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
   64957   int16_t __ret;
   64958   __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
   64959   return __ret;
   64960 }
   64961 #else
   64962 __ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
   64963   int16_t __ret;
   64964   __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
   64965   return __ret;
   64966 }
   64967 #endif
   64968 
   64969 #ifdef __LITTLE_ENDIAN__
   64970 __ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   64971   int8x16_t __ret;
   64972   __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
   64973   return __ret;
   64974 }
   64975 #else
   64976 __ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
   64977   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64978   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64979   int8x16_t __ret;
   64980   __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
   64981   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   64982   return __ret;
   64983 }
   64984 #endif
   64985 
   64986 #ifdef __LITTLE_ENDIAN__
   64987 __ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   64988   int32x4_t __ret;
   64989   __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
   64990   return __ret;
   64991 }
   64992 #else
   64993 __ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
   64994   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   64995   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   64996   int32x4_t __ret;
   64997   __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
   64998   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   64999   return __ret;
   65000 }
   65001 #endif
   65002 
   65003 #ifdef __LITTLE_ENDIAN__
   65004 __ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
   65005   int64x2_t __ret;
   65006   __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
   65007   return __ret;
   65008 }
   65009 #else
   65010 __ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
   65011   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65012   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65013   int64x2_t __ret;
   65014   __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
   65015   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65016   return __ret;
   65017 }
   65018 #endif
   65019 
   65020 #ifdef __LITTLE_ENDIAN__
   65021 __ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   65022   int16x8_t __ret;
   65023   __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
   65024   return __ret;
   65025 }
   65026 #else
   65027 __ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
   65028   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65029   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65030   int16x8_t __ret;
   65031   __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
   65032   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65033   return __ret;
   65034 }
   65035 #endif
   65036 
   65037 #ifdef __LITTLE_ENDIAN__
   65038 __ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
   65039   int8x8_t __ret;
   65040   __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
   65041   return __ret;
   65042 }
   65043 #else
   65044 __ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
   65045   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65046   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65047   int8x8_t __ret;
   65048   __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
   65049   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65050   return __ret;
   65051 }
   65052 #endif
   65053 
   65054 #ifdef __LITTLE_ENDIAN__
   65055 __ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
   65056   int32x2_t __ret;
   65057   __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
   65058   return __ret;
   65059 }
   65060 #else
   65061 __ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
   65062   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65063   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65064   int32x2_t __ret;
   65065   __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
   65066   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65067   return __ret;
   65068 }
   65069 #endif
   65070 
   65071 #ifdef __LITTLE_ENDIAN__
   65072 __ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
   65073   int64x1_t __ret;
   65074   __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   65075   return __ret;
   65076 }
   65077 #else
   65078 __ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
   65079   int64x1_t __ret;
   65080   __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
   65081   return __ret;
   65082 }
   65083 #endif
   65084 
   65085 #ifdef __LITTLE_ENDIAN__
   65086 __ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
   65087   int16x4_t __ret;
   65088   __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
   65089   return __ret;
   65090 }
   65091 #else
   65092 __ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
   65093   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65094   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65095   int16x4_t __ret;
   65096   __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
   65097   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65098   return __ret;
   65099 }
   65100 #endif
   65101 
   65102 #ifdef __LITTLE_ENDIAN__
   65103 __ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
   65104   poly8x8_t __ret;
   65105   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
   65106   return __ret;
   65107 }
   65108 #else
   65109 __ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
   65110   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65111   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65112   poly8x8_t __ret;
   65113   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
   65114   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65115   return __ret;
   65116 }
   65117 #endif
   65118 
   65119 #ifdef __LITTLE_ENDIAN__
   65120 __ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
   65121   poly16x4_t __ret;
   65122   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
   65123   return __ret;
   65124 }
   65125 #else
   65126 __ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
   65127   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65128   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65129   poly16x4_t __ret;
   65130   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
   65131   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65132   return __ret;
   65133 }
   65134 #endif
   65135 
   65136 #ifdef __LITTLE_ENDIAN__
   65137 __ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   65138   poly8x16_t __ret;
   65139   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
   65140   return __ret;
   65141 }
   65142 #else
   65143 __ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   65144   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65145   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65146   poly8x16_t __ret;
   65147   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
   65148   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65149   return __ret;
   65150 }
   65151 #endif
   65152 
   65153 #ifdef __LITTLE_ENDIAN__
   65154 __ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   65155   poly64x2_t __ret;
   65156   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   65157   return __ret;
   65158 }
   65159 #else
   65160 __ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   65161   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65162   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65163   poly64x2_t __ret;
   65164   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   65165   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65166   return __ret;
   65167 }
   65168 #endif
   65169 
   65170 #ifdef __LITTLE_ENDIAN__
   65171 __ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   65172   poly16x8_t __ret;
   65173   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
   65174   return __ret;
   65175 }
   65176 #else
   65177 __ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   65178   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65179   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65180   poly16x8_t __ret;
   65181   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
   65182   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65183   return __ret;
   65184 }
   65185 #endif
   65186 
   65187 #ifdef __LITTLE_ENDIAN__
   65188 __ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   65189   uint8x16_t __ret;
   65190   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
   65191   return __ret;
   65192 }
   65193 #else
   65194 __ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   65195   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65196   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65197   uint8x16_t __ret;
   65198   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
   65199   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65200   return __ret;
   65201 }
   65202 #endif
   65203 
   65204 #ifdef __LITTLE_ENDIAN__
   65205 __ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   65206   uint32x4_t __ret;
   65207   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
   65208   return __ret;
   65209 }
   65210 #else
   65211 __ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   65212   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65213   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65214   uint32x4_t __ret;
   65215   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
   65216   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65217   return __ret;
   65218 }
   65219 #endif
   65220 
   65221 #ifdef __LITTLE_ENDIAN__
   65222 __ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   65223   uint64x2_t __ret;
   65224   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   65225   return __ret;
   65226 }
   65227 #else
   65228 __ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   65229   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65230   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65231   uint64x2_t __ret;
   65232   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   65233   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65234   return __ret;
   65235 }
   65236 #endif
   65237 
   65238 #ifdef __LITTLE_ENDIAN__
   65239 __ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   65240   uint16x8_t __ret;
   65241   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
   65242   return __ret;
   65243 }
   65244 #else
   65245 __ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   65246   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65247   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65248   uint16x8_t __ret;
   65249   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
   65250   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65251   return __ret;
   65252 }
   65253 #endif
   65254 
   65255 #ifdef __LITTLE_ENDIAN__
   65256 __ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
   65257   int8x16_t __ret;
   65258   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
   65259   return __ret;
   65260 }
   65261 #else
   65262 __ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
   65263   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65264   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65265   int8x16_t __ret;
   65266   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
   65267   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65268   return __ret;
   65269 }
   65270 #endif
   65271 
   65272 #ifdef __LITTLE_ENDIAN__
   65273 __ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
   65274   float64x2_t __ret;
   65275   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   65276   return __ret;
   65277 }
   65278 #else
   65279 __ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
   65280   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65281   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65282   float64x2_t __ret;
   65283   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   65284   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65285   return __ret;
   65286 }
   65287 #endif
   65288 
   65289 #ifdef __LITTLE_ENDIAN__
   65290 __ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
   65291   float32x4_t __ret;
   65292   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
   65293   return __ret;
   65294 }
   65295 #else
   65296 __ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
   65297   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65298   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65299   float32x4_t __ret;
   65300   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
   65301   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65302   return __ret;
   65303 }
   65304 #endif
   65305 
   65306 #ifdef __LITTLE_ENDIAN__
   65307 __ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
   65308   int32x4_t __ret;
   65309   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
   65310   return __ret;
   65311 }
   65312 #else
   65313 __ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
   65314   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65315   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65316   int32x4_t __ret;
   65317   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
   65318   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65319   return __ret;
   65320 }
   65321 #endif
   65322 
   65323 #ifdef __LITTLE_ENDIAN__
   65324 __ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
   65325   int64x2_t __ret;
   65326   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   65327   return __ret;
   65328 }
   65329 #else
   65330 __ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
   65331   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65332   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65333   int64x2_t __ret;
   65334   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   65335   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65336   return __ret;
   65337 }
   65338 #endif
   65339 
   65340 #ifdef __LITTLE_ENDIAN__
   65341 __ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
   65342   int16x8_t __ret;
   65343   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
   65344   return __ret;
   65345 }
   65346 #else
   65347 __ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
   65348   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65349   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65350   int16x8_t __ret;
   65351   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
   65352   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65353   return __ret;
   65354 }
   65355 #endif
   65356 
   65357 #ifdef __LITTLE_ENDIAN__
   65358 __ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
   65359   uint8x8_t __ret;
   65360   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
   65361   return __ret;
   65362 }
   65363 #else
   65364 __ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
   65365   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65366   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65367   uint8x8_t __ret;
   65368   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
   65369   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65370   return __ret;
   65371 }
   65372 #endif
   65373 
   65374 #ifdef __LITTLE_ENDIAN__
   65375 __ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
   65376   uint32x2_t __ret;
   65377   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   65378   return __ret;
   65379 }
   65380 #else
   65381 __ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
   65382   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65383   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65384   uint32x2_t __ret;
   65385   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   65386   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65387   return __ret;
   65388 }
   65389 #endif
   65390 
   65391 #ifdef __LITTLE_ENDIAN__
   65392 __ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
   65393   uint16x4_t __ret;
   65394   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
   65395   return __ret;
   65396 }
   65397 #else
   65398 __ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
   65399   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65400   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65401   uint16x4_t __ret;
   65402   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
   65403   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65404   return __ret;
   65405 }
   65406 #endif
   65407 
   65408 #ifdef __LITTLE_ENDIAN__
   65409 __ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
   65410   int8x8_t __ret;
   65411   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
   65412   return __ret;
   65413 }
   65414 #else
   65415 __ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
   65416   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65417   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65418   int8x8_t __ret;
   65419   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
   65420   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65421   return __ret;
   65422 }
   65423 #endif
   65424 
   65425 #ifdef __LITTLE_ENDIAN__
   65426 __ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
   65427   float32x2_t __ret;
   65428   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   65429   return __ret;
   65430 }
   65431 #else
   65432 __ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
   65433   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65434   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65435   float32x2_t __ret;
   65436   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   65437   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65438   return __ret;
   65439 }
   65440 #endif
   65441 
   65442 #ifdef __LITTLE_ENDIAN__
   65443 __ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
   65444   int32x2_t __ret;
   65445   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   65446   return __ret;
   65447 }
   65448 #else
   65449 __ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
   65450   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65451   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65452   int32x2_t __ret;
   65453   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   65454   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65455   return __ret;
   65456 }
   65457 #endif
   65458 
   65459 #ifdef __LITTLE_ENDIAN__
   65460 __ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
   65461   int16x4_t __ret;
   65462   __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
   65463   return __ret;
   65464 }
   65465 #else
   65466 __ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
   65467   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65468   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65469   int16x4_t __ret;
   65470   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
   65471   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65472   return __ret;
   65473 }
   65474 #endif
   65475 
   65476 #ifdef __LITTLE_ENDIAN__
   65477 __ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
   65478   poly8x8_t __ret;
   65479   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
   65480   return __ret;
   65481 }
   65482 #else
   65483 __ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
   65484   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65485   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65486   poly8x8_t __ret;
   65487   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
   65488   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65489   return __ret;
   65490 }
   65491 #endif
   65492 
   65493 #ifdef __LITTLE_ENDIAN__
   65494 __ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
   65495   poly16x4_t __ret;
   65496   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
   65497   return __ret;
   65498 }
   65499 #else
   65500 __ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
   65501   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65502   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65503   poly16x4_t __ret;
   65504   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
   65505   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65506   return __ret;
   65507 }
   65508 #endif
   65509 
   65510 #ifdef __LITTLE_ENDIAN__
   65511 __ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   65512   poly8x16_t __ret;
   65513   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
   65514   return __ret;
   65515 }
   65516 #else
   65517 __ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   65518   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65519   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65520   poly8x16_t __ret;
   65521   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
   65522   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65523   return __ret;
   65524 }
   65525 #endif
   65526 
   65527 #ifdef __LITTLE_ENDIAN__
   65528 __ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   65529   poly64x2_t __ret;
   65530   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   65531   return __ret;
   65532 }
   65533 #else
   65534 __ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   65535   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65536   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65537   poly64x2_t __ret;
   65538   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   65539   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65540   return __ret;
   65541 }
   65542 #endif
   65543 
   65544 #ifdef __LITTLE_ENDIAN__
   65545 __ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   65546   poly16x8_t __ret;
   65547   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
   65548   return __ret;
   65549 }
   65550 #else
   65551 __ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   65552   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65553   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65554   poly16x8_t __ret;
   65555   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
   65556   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65557   return __ret;
   65558 }
   65559 #endif
   65560 
   65561 #ifdef __LITTLE_ENDIAN__
   65562 __ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   65563   uint8x16_t __ret;
   65564   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
   65565   return __ret;
   65566 }
   65567 #else
   65568 __ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   65569   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65570   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65571   uint8x16_t __ret;
   65572   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
   65573   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65574   return __ret;
   65575 }
   65576 #endif
   65577 
   65578 #ifdef __LITTLE_ENDIAN__
   65579 __ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   65580   uint32x4_t __ret;
   65581   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
   65582   return __ret;
   65583 }
   65584 #else
   65585 __ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   65586   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65587   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65588   uint32x4_t __ret;
   65589   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
   65590   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65591   return __ret;
   65592 }
   65593 #endif
   65594 
   65595 #ifdef __LITTLE_ENDIAN__
   65596 __ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   65597   uint64x2_t __ret;
   65598   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   65599   return __ret;
   65600 }
   65601 #else
   65602 __ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   65603   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65604   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65605   uint64x2_t __ret;
   65606   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   65607   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65608   return __ret;
   65609 }
   65610 #endif
   65611 
   65612 #ifdef __LITTLE_ENDIAN__
   65613 __ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   65614   uint16x8_t __ret;
   65615   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
   65616   return __ret;
   65617 }
   65618 #else
   65619 __ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   65620   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65621   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65622   uint16x8_t __ret;
   65623   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
   65624   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65625   return __ret;
   65626 }
   65627 #endif
   65628 
   65629 #ifdef __LITTLE_ENDIAN__
   65630 __ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
   65631   int8x16_t __ret;
   65632   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
   65633   return __ret;
   65634 }
   65635 #else
   65636 __ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
   65637   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65638   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65639   int8x16_t __ret;
   65640   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
   65641   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65642   return __ret;
   65643 }
   65644 #endif
   65645 
   65646 #ifdef __LITTLE_ENDIAN__
   65647 __ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
   65648   float64x2_t __ret;
   65649   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   65650   return __ret;
   65651 }
   65652 #else
   65653 __ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
   65654   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65655   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65656   float64x2_t __ret;
   65657   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   65658   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65659   return __ret;
   65660 }
   65661 #endif
   65662 
   65663 #ifdef __LITTLE_ENDIAN__
   65664 __ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
   65665   float32x4_t __ret;
   65666   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
   65667   return __ret;
   65668 }
   65669 #else
   65670 __ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
   65671   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65672   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65673   float32x4_t __ret;
   65674   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
   65675   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65676   return __ret;
   65677 }
   65678 #endif
   65679 
   65680 #ifdef __LITTLE_ENDIAN__
   65681 __ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
   65682   int32x4_t __ret;
   65683   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
   65684   return __ret;
   65685 }
   65686 #else
   65687 __ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
   65688   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65689   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65690   int32x4_t __ret;
   65691   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
   65692   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65693   return __ret;
   65694 }
   65695 #endif
   65696 
   65697 #ifdef __LITTLE_ENDIAN__
   65698 __ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
   65699   int64x2_t __ret;
   65700   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   65701   return __ret;
   65702 }
   65703 #else
   65704 __ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
   65705   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65706   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65707   int64x2_t __ret;
   65708   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   65709   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65710   return __ret;
   65711 }
   65712 #endif
   65713 
   65714 #ifdef __LITTLE_ENDIAN__
   65715 __ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
   65716   int16x8_t __ret;
   65717   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
   65718   return __ret;
   65719 }
   65720 #else
   65721 __ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
   65722   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65723   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65724   int16x8_t __ret;
   65725   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
   65726   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65727   return __ret;
   65728 }
   65729 #endif
   65730 
   65731 #ifdef __LITTLE_ENDIAN__
   65732 __ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
   65733   uint8x8_t __ret;
   65734   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
   65735   return __ret;
   65736 }
   65737 #else
   65738 __ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
   65739   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65740   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65741   uint8x8_t __ret;
   65742   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
   65743   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65744   return __ret;
   65745 }
   65746 #endif
   65747 
   65748 #ifdef __LITTLE_ENDIAN__
   65749 __ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
   65750   uint32x2_t __ret;
   65751   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   65752   return __ret;
   65753 }
   65754 #else
   65755 __ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
   65756   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65757   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65758   uint32x2_t __ret;
   65759   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   65760   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65761   return __ret;
   65762 }
   65763 #endif
   65764 
   65765 #ifdef __LITTLE_ENDIAN__
   65766 __ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
   65767   uint16x4_t __ret;
   65768   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
   65769   return __ret;
   65770 }
   65771 #else
   65772 __ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
   65773   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65774   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65775   uint16x4_t __ret;
   65776   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
   65777   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65778   return __ret;
   65779 }
   65780 #endif
   65781 
   65782 #ifdef __LITTLE_ENDIAN__
   65783 __ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
   65784   int8x8_t __ret;
   65785   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
   65786   return __ret;
   65787 }
   65788 #else
   65789 __ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
   65790   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65791   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65792   int8x8_t __ret;
   65793   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
   65794   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65795   return __ret;
   65796 }
   65797 #endif
   65798 
   65799 #ifdef __LITTLE_ENDIAN__
   65800 __ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
   65801   float32x2_t __ret;
   65802   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   65803   return __ret;
   65804 }
   65805 #else
   65806 __ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
   65807   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65808   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65809   float32x2_t __ret;
   65810   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   65811   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65812   return __ret;
   65813 }
   65814 #endif
   65815 
   65816 #ifdef __LITTLE_ENDIAN__
   65817 __ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
   65818   int32x2_t __ret;
   65819   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   65820   return __ret;
   65821 }
   65822 #else
   65823 __ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
   65824   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65825   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65826   int32x2_t __ret;
   65827   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   65828   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65829   return __ret;
   65830 }
   65831 #endif
   65832 
   65833 #ifdef __LITTLE_ENDIAN__
   65834 __ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
   65835   int16x4_t __ret;
   65836   __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
   65837   return __ret;
   65838 }
   65839 #else
   65840 __ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
   65841   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65842   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65843   int16x4_t __ret;
   65844   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
   65845   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65846   return __ret;
   65847 }
   65848 #endif
   65849 
   65850 #ifdef __LITTLE_ENDIAN__
   65851 __ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
   65852   poly8x8_t __ret;
   65853   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
   65854   return __ret;
   65855 }
   65856 #else
   65857 __ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
   65858   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65859   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65860   poly8x8_t __ret;
   65861   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
   65862   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65863   return __ret;
   65864 }
   65865 #endif
   65866 
   65867 #ifdef __LITTLE_ENDIAN__
   65868 __ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
   65869   poly16x4_t __ret;
   65870   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
   65871   return __ret;
   65872 }
   65873 #else
   65874 __ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
   65875   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65876   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65877   poly16x4_t __ret;
   65878   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
   65879   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65880   return __ret;
   65881 }
   65882 #endif
   65883 
   65884 #ifdef __LITTLE_ENDIAN__
   65885 __ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   65886   poly8x16_t __ret;
   65887   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
   65888   return __ret;
   65889 }
   65890 #else
   65891 __ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   65892   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65893   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65894   poly8x16_t __ret;
   65895   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
   65896   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65897   return __ret;
   65898 }
   65899 #endif
   65900 
   65901 #ifdef __LITTLE_ENDIAN__
   65902 __ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   65903   poly64x2_t __ret;
   65904   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   65905   return __ret;
   65906 }
   65907 #else
   65908 __ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   65909   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65910   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65911   poly64x2_t __ret;
   65912   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   65913   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65914   return __ret;
   65915 }
   65916 #endif
   65917 
   65918 #ifdef __LITTLE_ENDIAN__
   65919 __ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   65920   poly16x8_t __ret;
   65921   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
   65922   return __ret;
   65923 }
   65924 #else
   65925 __ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   65926   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65927   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65928   poly16x8_t __ret;
   65929   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
   65930   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65931   return __ret;
   65932 }
   65933 #endif
   65934 
   65935 #ifdef __LITTLE_ENDIAN__
   65936 __ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   65937   uint8x16_t __ret;
   65938   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
   65939   return __ret;
   65940 }
   65941 #else
   65942 __ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   65943   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65944   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65945   uint8x16_t __ret;
   65946   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
   65947   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   65948   return __ret;
   65949 }
   65950 #endif
   65951 
   65952 #ifdef __LITTLE_ENDIAN__
   65953 __ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   65954   uint32x4_t __ret;
   65955   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
   65956   return __ret;
   65957 }
   65958 #else
   65959 __ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   65960   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   65961   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   65962   uint32x4_t __ret;
   65963   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
   65964   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   65965   return __ret;
   65966 }
   65967 #endif
   65968 
   65969 #ifdef __LITTLE_ENDIAN__
   65970 __ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   65971   uint64x2_t __ret;
   65972   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   65973   return __ret;
   65974 }
   65975 #else
   65976 __ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   65977   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   65978   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   65979   uint64x2_t __ret;
   65980   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   65981   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   65982   return __ret;
   65983 }
   65984 #endif
   65985 
   65986 #ifdef __LITTLE_ENDIAN__
   65987 __ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   65988   uint16x8_t __ret;
   65989   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
   65990   return __ret;
   65991 }
   65992 #else
   65993 __ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   65994   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   65995   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   65996   uint16x8_t __ret;
   65997   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
   65998   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   65999   return __ret;
   66000 }
   66001 #endif
   66002 
   66003 #ifdef __LITTLE_ENDIAN__
   66004 __ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
   66005   int8x16_t __ret;
   66006   __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
   66007   return __ret;
   66008 }
   66009 #else
   66010 __ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
   66011   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66012   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66013   int8x16_t __ret;
   66014   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
   66015   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66016   return __ret;
   66017 }
   66018 #endif
   66019 
   66020 #ifdef __LITTLE_ENDIAN__
   66021 __ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
   66022   float64x2_t __ret;
   66023   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   66024   return __ret;
   66025 }
   66026 #else
   66027 __ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
   66028   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66029   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66030   float64x2_t __ret;
   66031   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   66032   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66033   return __ret;
   66034 }
   66035 #endif
   66036 
   66037 #ifdef __LITTLE_ENDIAN__
   66038 __ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
   66039   float32x4_t __ret;
   66040   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
   66041   return __ret;
   66042 }
   66043 #else
   66044 __ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
   66045   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66046   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66047   float32x4_t __ret;
   66048   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
   66049   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66050   return __ret;
   66051 }
   66052 #endif
   66053 
   66054 #ifdef __LITTLE_ENDIAN__
   66055 __ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
   66056   int32x4_t __ret;
   66057   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
   66058   return __ret;
   66059 }
   66060 #else
   66061 __ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
   66062   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66063   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66064   int32x4_t __ret;
   66065   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
   66066   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66067   return __ret;
   66068 }
   66069 #endif
   66070 
   66071 #ifdef __LITTLE_ENDIAN__
   66072 __ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
   66073   int64x2_t __ret;
   66074   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   66075   return __ret;
   66076 }
   66077 #else
   66078 __ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
   66079   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66080   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66081   int64x2_t __ret;
   66082   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   66083   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66084   return __ret;
   66085 }
   66086 #endif
   66087 
   66088 #ifdef __LITTLE_ENDIAN__
   66089 __ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
   66090   int16x8_t __ret;
   66091   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
   66092   return __ret;
   66093 }
   66094 #else
   66095 __ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
   66096   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66097   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66098   int16x8_t __ret;
   66099   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
   66100   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66101   return __ret;
   66102 }
   66103 #endif
   66104 
   66105 #ifdef __LITTLE_ENDIAN__
   66106 __ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
   66107   uint8x8_t __ret;
   66108   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
   66109   return __ret;
   66110 }
   66111 #else
   66112 __ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
   66113   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66114   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66115   uint8x8_t __ret;
   66116   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
   66117   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66118   return __ret;
   66119 }
   66120 #endif
   66121 
   66122 #ifdef __LITTLE_ENDIAN__
   66123 __ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
   66124   uint32x2_t __ret;
   66125   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   66126   return __ret;
   66127 }
   66128 #else
   66129 __ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
   66130   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66131   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66132   uint32x2_t __ret;
   66133   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   66134   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66135   return __ret;
   66136 }
   66137 #endif
   66138 
   66139 #ifdef __LITTLE_ENDIAN__
   66140 __ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
   66141   uint16x4_t __ret;
   66142   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
   66143   return __ret;
   66144 }
   66145 #else
   66146 __ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
   66147   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66148   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66149   uint16x4_t __ret;
   66150   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
   66151   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66152   return __ret;
   66153 }
   66154 #endif
   66155 
   66156 #ifdef __LITTLE_ENDIAN__
   66157 __ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
   66158   int8x8_t __ret;
   66159   __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
   66160   return __ret;
   66161 }
   66162 #else
   66163 __ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
   66164   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66165   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66166   int8x8_t __ret;
   66167   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
   66168   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66169   return __ret;
   66170 }
   66171 #endif
   66172 
   66173 #ifdef __LITTLE_ENDIAN__
   66174 __ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
   66175   float32x2_t __ret;
   66176   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   66177   return __ret;
   66178 }
   66179 #else
   66180 __ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
   66181   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66182   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66183   float32x2_t __ret;
   66184   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   66185   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66186   return __ret;
   66187 }
   66188 #endif
   66189 
   66190 #ifdef __LITTLE_ENDIAN__
   66191 __ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
   66192   int32x2_t __ret;
   66193   __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
   66194   return __ret;
   66195 }
   66196 #else
   66197 __ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
   66198   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66199   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66200   int32x2_t __ret;
   66201   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
   66202   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66203   return __ret;
   66204 }
   66205 #endif
   66206 
   66207 #ifdef __LITTLE_ENDIAN__
   66208 __ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
   66209   int16x4_t __ret;
   66210   __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
   66211   return __ret;
   66212 }
   66213 #else
   66214 __ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
   66215   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66216   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66217   int16x4_t __ret;
   66218   __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
   66219   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66220   return __ret;
   66221 }
   66222 #endif
   66223 
   66224 #ifdef __LITTLE_ENDIAN__
   66225 __ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
   66226   poly8x8_t __ret;
   66227   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
   66228   return __ret;
   66229 }
   66230 #else
   66231 __ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
   66232   poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66233   poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66234   poly8x8_t __ret;
   66235   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
   66236   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66237   return __ret;
   66238 }
   66239 #endif
   66240 
   66241 #ifdef __LITTLE_ENDIAN__
   66242 __ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
   66243   poly16x4_t __ret;
   66244   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
   66245   return __ret;
   66246 }
   66247 #else
   66248 __ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
   66249   poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66250   poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66251   poly16x4_t __ret;
   66252   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
   66253   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66254   return __ret;
   66255 }
   66256 #endif
   66257 
   66258 #ifdef __LITTLE_ENDIAN__
   66259 __ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   66260   poly8x16_t __ret;
   66261   __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
   66262   return __ret;
   66263 }
   66264 #else
   66265 __ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
   66266   poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66267   poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66268   poly8x16_t __ret;
   66269   __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
   66270   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66271   return __ret;
   66272 }
   66273 #endif
   66274 
   66275 #ifdef __LITTLE_ENDIAN__
   66276 __ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   66277   poly64x2_t __ret;
   66278   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   66279   return __ret;
   66280 }
   66281 #else
   66282 __ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
   66283   poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66284   poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66285   poly64x2_t __ret;
   66286   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   66287   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66288   return __ret;
   66289 }
   66290 #endif
   66291 
   66292 #ifdef __LITTLE_ENDIAN__
   66293 __ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   66294   poly16x8_t __ret;
   66295   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
   66296   return __ret;
   66297 }
   66298 #else
   66299 __ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
   66300   poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66301   poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66302   poly16x8_t __ret;
   66303   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
   66304   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66305   return __ret;
   66306 }
   66307 #endif
   66308 
   66309 #ifdef __LITTLE_ENDIAN__
   66310 __ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   66311   uint8x16_t __ret;
   66312   __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
   66313   return __ret;
   66314 }
   66315 #else
   66316 __ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
   66317   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66318   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66319   uint8x16_t __ret;
   66320   __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
   66321   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66322   return __ret;
   66323 }
   66324 #endif
   66325 
   66326 #ifdef __LITTLE_ENDIAN__
   66327 __ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   66328   uint32x4_t __ret;
   66329   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
   66330   return __ret;
   66331 }
   66332 #else
   66333 __ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
   66334   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66335   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66336   uint32x4_t __ret;
   66337   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
   66338   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66339   return __ret;
   66340 }
   66341 #endif
   66342 
   66343 #ifdef __LITTLE_ENDIAN__
   66344 __ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   66345   uint64x2_t __ret;
   66346   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   66347   return __ret;
   66348 }
   66349 #else
   66350 __ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
   66351   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66352   uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66353   uint64x2_t __ret;
   66354   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   66355   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66356   return __ret;
   66357 }
   66358 #endif
   66359 
   66360 #ifdef __LITTLE_ENDIAN__
   66361 __ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   66362   uint16x8_t __ret;
   66363   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
   66364   return __ret;
   66365 }
   66366 #else
   66367 __ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
   66368   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66369   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66370   uint16x8_t __ret;
   66371   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
   66372   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66373   return __ret;
   66374 }
   66375 #endif
   66376 
   66377 #ifdef __LITTLE_ENDIAN__
   66378 __ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
   66379   int8x16_t __ret;
   66380   __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
   66381   return __ret;
   66382 }
   66383 #else
   66384 __ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
   66385   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66386   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66387   int8x16_t __ret;
   66388   __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
   66389   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66390   return __ret;
   66391 }
   66392 #endif
   66393 
   66394 #ifdef __LITTLE_ENDIAN__
   66395 __ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
   66396   float64x2_t __ret;
   66397   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   66398   return __ret;
   66399 }
   66400 #else
   66401 __ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
   66402   float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66403   float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66404   float64x2_t __ret;
   66405   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   66406   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66407   return __ret;
   66408 }
   66409 #endif
   66410 
   66411 #ifdef __LITTLE_ENDIAN__
   66412 __ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
   66413   float32x4_t __ret;
   66414   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
   66415   return __ret;
   66416 }
   66417 #else
   66418 __ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
   66419   float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66420   float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66421   float32x4_t __ret;
   66422   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
   66423   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66424   return __ret;
   66425 }
   66426 #endif
   66427 
   66428 #ifdef __LITTLE_ENDIAN__
   66429 __ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
   66430   int32x4_t __ret;
   66431   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
   66432   return __ret;
   66433 }
   66434 #else
   66435 __ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
   66436   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66437   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66438   int32x4_t __ret;
   66439   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
   66440   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66441   return __ret;
   66442 }
   66443 #endif
   66444 
   66445 #ifdef __LITTLE_ENDIAN__
   66446 __ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
   66447   int64x2_t __ret;
   66448   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   66449   return __ret;
   66450 }
   66451 #else
   66452 __ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
   66453   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66454   int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66455   int64x2_t __ret;
   66456   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   66457   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66458   return __ret;
   66459 }
   66460 #endif
   66461 
   66462 #ifdef __LITTLE_ENDIAN__
   66463 __ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
   66464   int16x8_t __ret;
   66465   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
   66466   return __ret;
   66467 }
   66468 #else
   66469 __ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
   66470   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66471   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66472   int16x8_t __ret;
   66473   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
   66474   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66475   return __ret;
   66476 }
   66477 #endif
   66478 
   66479 #ifdef __LITTLE_ENDIAN__
   66480 __ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
   66481   uint8x8_t __ret;
   66482   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
   66483   return __ret;
   66484 }
   66485 #else
   66486 __ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
   66487   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66488   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66489   uint8x8_t __ret;
   66490   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
   66491   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66492   return __ret;
   66493 }
   66494 #endif
   66495 
   66496 #ifdef __LITTLE_ENDIAN__
   66497 __ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
   66498   uint32x2_t __ret;
   66499   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   66500   return __ret;
   66501 }
   66502 #else
   66503 __ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
   66504   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66505   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66506   uint32x2_t __ret;
   66507   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   66508   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66509   return __ret;
   66510 }
   66511 #endif
   66512 
   66513 #ifdef __LITTLE_ENDIAN__
   66514 __ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
   66515   uint16x4_t __ret;
   66516   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
   66517   return __ret;
   66518 }
   66519 #else
   66520 __ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
   66521   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66522   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66523   uint16x4_t __ret;
   66524   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
   66525   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66526   return __ret;
   66527 }
   66528 #endif
   66529 
   66530 #ifdef __LITTLE_ENDIAN__
   66531 __ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
   66532   int8x8_t __ret;
   66533   __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
   66534   return __ret;
   66535 }
   66536 #else
   66537 __ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
   66538   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66539   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66540   int8x8_t __ret;
   66541   __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
   66542   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66543   return __ret;
   66544 }
   66545 #endif
   66546 
   66547 #ifdef __LITTLE_ENDIAN__
   66548 __ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
   66549   float32x2_t __ret;
   66550   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   66551   return __ret;
   66552 }
   66553 #else
   66554 __ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
   66555   float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66556   float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66557   float32x2_t __ret;
   66558   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   66559   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66560   return __ret;
   66561 }
   66562 #endif
   66563 
   66564 #ifdef __LITTLE_ENDIAN__
   66565 __ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
   66566   int32x2_t __ret;
   66567   __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
   66568   return __ret;
   66569 }
   66570 #else
   66571 __ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
   66572   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66573   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66574   int32x2_t __ret;
   66575   __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
   66576   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66577   return __ret;
   66578 }
   66579 #endif
   66580 
   66581 #ifdef __LITTLE_ENDIAN__
   66582 __ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
   66583   int16x4_t __ret;
   66584   __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
   66585   return __ret;
   66586 }
   66587 #else
   66588 __ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
   66589   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66590   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66591   int16x4_t __ret;
   66592   __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
   66593   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66594   return __ret;
   66595 }
   66596 #endif
   66597 
   66598 #endif
   66599 #ifdef __LITTLE_ENDIAN__
   66600 __ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   66601   uint8x16_t __ret;
   66602   __ret = __p0 + vabdq_u8(__p1, __p2);
   66603   return __ret;
   66604 }
   66605 #else
   66606 __ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   66607   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66608   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66609   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66610   uint8x16_t __ret;
   66611   __ret = __rev0 + __noswap_vabdq_u8(__rev1, __rev2);
   66612   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66613   return __ret;
   66614 }
   66615 #endif
   66616 
   66617 #ifdef __LITTLE_ENDIAN__
   66618 __ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   66619   uint32x4_t __ret;
   66620   __ret = __p0 + vabdq_u32(__p1, __p2);
   66621   return __ret;
   66622 }
   66623 #else
   66624 __ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   66625   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66626   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66627   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   66628   uint32x4_t __ret;
   66629   __ret = __rev0 + __noswap_vabdq_u32(__rev1, __rev2);
   66630   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66631   return __ret;
   66632 }
   66633 #endif
   66634 
   66635 #ifdef __LITTLE_ENDIAN__
   66636 __ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   66637   uint16x8_t __ret;
   66638   __ret = __p0 + vabdq_u16(__p1, __p2);
   66639   return __ret;
   66640 }
   66641 #else
   66642 __ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   66643   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66644   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66645   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   66646   uint16x8_t __ret;
   66647   __ret = __rev0 + __noswap_vabdq_u16(__rev1, __rev2);
   66648   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66649   return __ret;
   66650 }
   66651 #endif
   66652 
   66653 #ifdef __LITTLE_ENDIAN__
   66654 __ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   66655   int8x16_t __ret;
   66656   __ret = __p0 + vabdq_s8(__p1, __p2);
   66657   return __ret;
   66658 }
   66659 #else
   66660 __ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
   66661   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66662   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66663   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66664   int8x16_t __ret;
   66665   __ret = __rev0 + __noswap_vabdq_s8(__rev1, __rev2);
   66666   __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   66667   return __ret;
   66668 }
   66669 #endif
   66670 
   66671 #ifdef __LITTLE_ENDIAN__
   66672 __ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   66673   int32x4_t __ret;
   66674   __ret = __p0 + vabdq_s32(__p1, __p2);
   66675   return __ret;
   66676 }
   66677 #else
   66678 __ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
   66679   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66680   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66681   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   66682   int32x4_t __ret;
   66683   __ret = __rev0 + __noswap_vabdq_s32(__rev1, __rev2);
   66684   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66685   return __ret;
   66686 }
   66687 #endif
   66688 
   66689 #ifdef __LITTLE_ENDIAN__
   66690 __ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   66691   int16x8_t __ret;
   66692   __ret = __p0 + vabdq_s16(__p1, __p2);
   66693   return __ret;
   66694 }
   66695 #else
   66696 __ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
   66697   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66698   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66699   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   66700   int16x8_t __ret;
   66701   __ret = __rev0 + __noswap_vabdq_s16(__rev1, __rev2);
   66702   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66703   return __ret;
   66704 }
   66705 #endif
   66706 
   66707 #ifdef __LITTLE_ENDIAN__
   66708 __ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   66709   uint8x8_t __ret;
   66710   __ret = __p0 + vabd_u8(__p1, __p2);
   66711   return __ret;
   66712 }
   66713 #else
   66714 __ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   66715   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66716   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66717   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   66718   uint8x8_t __ret;
   66719   __ret = __rev0 + __noswap_vabd_u8(__rev1, __rev2);
   66720   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66721   return __ret;
   66722 }
   66723 #endif
   66724 
   66725 #ifdef __LITTLE_ENDIAN__
   66726 __ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   66727   uint32x2_t __ret;
   66728   __ret = __p0 + vabd_u32(__p1, __p2);
   66729   return __ret;
   66730 }
   66731 #else
   66732 __ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   66733   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66734   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66735   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   66736   uint32x2_t __ret;
   66737   __ret = __rev0 + __noswap_vabd_u32(__rev1, __rev2);
   66738   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66739   return __ret;
   66740 }
   66741 #endif
   66742 
   66743 #ifdef __LITTLE_ENDIAN__
   66744 __ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   66745   uint16x4_t __ret;
   66746   __ret = __p0 + vabd_u16(__p1, __p2);
   66747   return __ret;
   66748 }
   66749 #else
   66750 __ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   66751   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66752   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66753   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   66754   uint16x4_t __ret;
   66755   __ret = __rev0 + __noswap_vabd_u16(__rev1, __rev2);
   66756   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66757   return __ret;
   66758 }
   66759 #endif
   66760 
   66761 #ifdef __LITTLE_ENDIAN__
   66762 __ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   66763   int8x8_t __ret;
   66764   __ret = __p0 + vabd_s8(__p1, __p2);
   66765   return __ret;
   66766 }
   66767 #else
   66768 __ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   66769   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66770   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66771   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   66772   int8x8_t __ret;
   66773   __ret = __rev0 + __noswap_vabd_s8(__rev1, __rev2);
   66774   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66775   return __ret;
   66776 }
   66777 #endif
   66778 
   66779 #ifdef __LITTLE_ENDIAN__
   66780 __ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   66781   int32x2_t __ret;
   66782   __ret = __p0 + vabd_s32(__p1, __p2);
   66783   return __ret;
   66784 }
   66785 #else
   66786 __ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   66787   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66788   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66789   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   66790   int32x2_t __ret;
   66791   __ret = __rev0 + __noswap_vabd_s32(__rev1, __rev2);
   66792   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66793   return __ret;
   66794 }
   66795 #endif
   66796 
   66797 #ifdef __LITTLE_ENDIAN__
   66798 __ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   66799   int16x4_t __ret;
   66800   __ret = __p0 + vabd_s16(__p1, __p2);
   66801   return __ret;
   66802 }
   66803 #else
   66804 __ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   66805   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66806   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66807   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   66808   int16x4_t __ret;
   66809   __ret = __rev0 + __noswap_vabd_s16(__rev1, __rev2);
   66810   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66811   return __ret;
   66812 }
   66813 #endif
   66814 
   66815 #ifdef __LITTLE_ENDIAN__
   66816 __ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
   66817   uint16x8_t __ret;
   66818   __ret = (uint16x8_t)(vmovl_u8((uint8x8_t)(vabd_u8(__p0, __p1))));
   66819   return __ret;
   66820 }
   66821 #else
   66822 __ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
   66823   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66824   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66825   uint16x8_t __ret;
   66826   __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__rev0, __rev1))));
   66827   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66828   return __ret;
   66829 }
   66830 __ai uint16x8_t __noswap_vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
   66831   uint16x8_t __ret;
   66832   __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__p0, __p1))));
   66833   return __ret;
   66834 }
   66835 #endif
   66836 
   66837 #ifdef __LITTLE_ENDIAN__
   66838 __ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
   66839   uint64x2_t __ret;
   66840   __ret = (uint64x2_t)(vmovl_u32((uint32x2_t)(vabd_u32(__p0, __p1))));
   66841   return __ret;
   66842 }
   66843 #else
   66844 __ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
   66845   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66846   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66847   uint64x2_t __ret;
   66848   __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__rev0, __rev1))));
   66849   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66850   return __ret;
   66851 }
   66852 __ai uint64x2_t __noswap_vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
   66853   uint64x2_t __ret;
   66854   __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__p0, __p1))));
   66855   return __ret;
   66856 }
   66857 #endif
   66858 
   66859 #ifdef __LITTLE_ENDIAN__
   66860 __ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
   66861   uint32x4_t __ret;
   66862   __ret = (uint32x4_t)(vmovl_u16((uint16x4_t)(vabd_u16(__p0, __p1))));
   66863   return __ret;
   66864 }
   66865 #else
   66866 __ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
   66867   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66868   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66869   uint32x4_t __ret;
   66870   __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__rev0, __rev1))));
   66871   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66872   return __ret;
   66873 }
   66874 __ai uint32x4_t __noswap_vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
   66875   uint32x4_t __ret;
   66876   __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__p0, __p1))));
   66877   return __ret;
   66878 }
   66879 #endif
   66880 
   66881 #ifdef __LITTLE_ENDIAN__
   66882 __ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
   66883   int16x8_t __ret;
   66884   __ret = (int16x8_t)(vmovl_u8((uint8x8_t)(vabd_s8(__p0, __p1))));
   66885   return __ret;
   66886 }
   66887 #else
   66888 __ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
   66889   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66890   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66891   int16x8_t __ret;
   66892   __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__rev0, __rev1))));
   66893   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66894   return __ret;
   66895 }
   66896 __ai int16x8_t __noswap_vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
   66897   int16x8_t __ret;
   66898   __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__p0, __p1))));
   66899   return __ret;
   66900 }
   66901 #endif
   66902 
   66903 #ifdef __LITTLE_ENDIAN__
   66904 __ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
   66905   int64x2_t __ret;
   66906   __ret = (int64x2_t)(vmovl_u32((uint32x2_t)(vabd_s32(__p0, __p1))));
   66907   return __ret;
   66908 }
   66909 #else
   66910 __ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
   66911   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66912   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66913   int64x2_t __ret;
   66914   __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__rev0, __rev1))));
   66915   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66916   return __ret;
   66917 }
   66918 __ai int64x2_t __noswap_vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
   66919   int64x2_t __ret;
   66920   __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__p0, __p1))));
   66921   return __ret;
   66922 }
   66923 #endif
   66924 
   66925 #ifdef __LITTLE_ENDIAN__
   66926 __ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
   66927   int32x4_t __ret;
   66928   __ret = (int32x4_t)(vmovl_u16((uint16x4_t)(vabd_s16(__p0, __p1))));
   66929   return __ret;
   66930 }
   66931 #else
   66932 __ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
   66933   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66934   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66935   int32x4_t __ret;
   66936   __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__rev0, __rev1))));
   66937   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66938   return __ret;
   66939 }
   66940 __ai int32x4_t __noswap_vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
   66941   int32x4_t __ret;
   66942   __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__p0, __p1))));
   66943   return __ret;
   66944 }
   66945 #endif
   66946 
   66947 #ifdef __LITTLE_ENDIAN__
   66948 __ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
   66949   uint16x8_t __ret;
   66950   __ret = vmovl_u8(__p0) + vmovl_u8(__p1);
   66951   return __ret;
   66952 }
   66953 #else
   66954 __ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
   66955   uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   66956   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   66957   uint16x8_t __ret;
   66958   __ret = __noswap_vmovl_u8(__rev0) + __noswap_vmovl_u8(__rev1);
   66959   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   66960   return __ret;
   66961 }
   66962 #endif
   66963 
   66964 #ifdef __LITTLE_ENDIAN__
   66965 __ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
   66966   uint64x2_t __ret;
   66967   __ret = vmovl_u32(__p0) + vmovl_u32(__p1);
   66968   return __ret;
   66969 }
   66970 #else
   66971 __ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
   66972   uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   66973   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   66974   uint64x2_t __ret;
   66975   __ret = __noswap_vmovl_u32(__rev0) + __noswap_vmovl_u32(__rev1);
   66976   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   66977   return __ret;
   66978 }
   66979 #endif
   66980 
   66981 #ifdef __LITTLE_ENDIAN__
   66982 __ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
   66983   uint32x4_t __ret;
   66984   __ret = vmovl_u16(__p0) + vmovl_u16(__p1);
   66985   return __ret;
   66986 }
   66987 #else
   66988 __ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
   66989   uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   66990   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   66991   uint32x4_t __ret;
   66992   __ret = __noswap_vmovl_u16(__rev0) + __noswap_vmovl_u16(__rev1);
   66993   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   66994   return __ret;
   66995 }
   66996 #endif
   66997 
   66998 #ifdef __LITTLE_ENDIAN__
   66999 __ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
   67000   int16x8_t __ret;
   67001   __ret = vmovl_s8(__p0) + vmovl_s8(__p1);
   67002   return __ret;
   67003 }
   67004 #else
   67005 __ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
   67006   int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   67007   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   67008   int16x8_t __ret;
   67009   __ret = __noswap_vmovl_s8(__rev0) + __noswap_vmovl_s8(__rev1);
   67010   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   67011   return __ret;
   67012 }
   67013 #endif
   67014 
   67015 #ifdef __LITTLE_ENDIAN__
   67016 __ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
   67017   int64x2_t __ret;
   67018   __ret = vmovl_s32(__p0) + vmovl_s32(__p1);
   67019   return __ret;
   67020 }
   67021 #else
   67022 __ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
   67023   int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67024   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67025   int64x2_t __ret;
   67026   __ret = __noswap_vmovl_s32(__rev0) + __noswap_vmovl_s32(__rev1);
   67027   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67028   return __ret;
   67029 }
   67030 #endif
   67031 
   67032 #ifdef __LITTLE_ENDIAN__
   67033 __ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
   67034   int32x4_t __ret;
   67035   __ret = vmovl_s16(__p0) + vmovl_s16(__p1);
   67036   return __ret;
   67037 }
   67038 #else
   67039 __ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
   67040   int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67041   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67042   int32x4_t __ret;
   67043   __ret = __noswap_vmovl_s16(__rev0) + __noswap_vmovl_s16(__rev1);
   67044   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67045   return __ret;
   67046 }
   67047 #endif
   67048 
   67049 #ifdef __LITTLE_ENDIAN__
   67050 __ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
   67051   uint16x8_t __ret;
   67052   __ret = __p0 + vmovl_u8(__p1);
   67053   return __ret;
   67054 }
   67055 #else
   67056 __ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
   67057   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   67058   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   67059   uint16x8_t __ret;
   67060   __ret = __rev0 + __noswap_vmovl_u8(__rev1);
   67061   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   67062   return __ret;
   67063 }
   67064 #endif
   67065 
   67066 #ifdef __LITTLE_ENDIAN__
   67067 __ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
   67068   uint64x2_t __ret;
   67069   __ret = __p0 + vmovl_u32(__p1);
   67070   return __ret;
   67071 }
   67072 #else
   67073 __ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
   67074   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67075   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67076   uint64x2_t __ret;
   67077   __ret = __rev0 + __noswap_vmovl_u32(__rev1);
   67078   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67079   return __ret;
   67080 }
   67081 #endif
   67082 
   67083 #ifdef __LITTLE_ENDIAN__
   67084 __ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
   67085   uint32x4_t __ret;
   67086   __ret = __p0 + vmovl_u16(__p1);
   67087   return __ret;
   67088 }
   67089 #else
   67090 __ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
   67091   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67092   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67093   uint32x4_t __ret;
   67094   __ret = __rev0 + __noswap_vmovl_u16(__rev1);
   67095   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67096   return __ret;
   67097 }
   67098 #endif
   67099 
   67100 #ifdef __LITTLE_ENDIAN__
   67101 __ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
   67102   int16x8_t __ret;
   67103   __ret = __p0 + vmovl_s8(__p1);
   67104   return __ret;
   67105 }
   67106 #else
   67107 __ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
   67108   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   67109   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   67110   int16x8_t __ret;
   67111   __ret = __rev0 + __noswap_vmovl_s8(__rev1);
   67112   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   67113   return __ret;
   67114 }
   67115 #endif
   67116 
   67117 #ifdef __LITTLE_ENDIAN__
   67118 __ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
   67119   int64x2_t __ret;
   67120   __ret = __p0 + vmovl_s32(__p1);
   67121   return __ret;
   67122 }
   67123 #else
   67124 __ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
   67125   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67126   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67127   int64x2_t __ret;
   67128   __ret = __rev0 + __noswap_vmovl_s32(__rev1);
   67129   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67130   return __ret;
   67131 }
   67132 #endif
   67133 
   67134 #ifdef __LITTLE_ENDIAN__
   67135 __ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
   67136   int32x4_t __ret;
   67137   __ret = __p0 + vmovl_s16(__p1);
   67138   return __ret;
   67139 }
   67140 #else
   67141 __ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
   67142   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67143   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67144   int32x4_t __ret;
   67145   __ret = __rev0 + __noswap_vmovl_s16(__rev1);
   67146   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67147   return __ret;
   67148 }
   67149 #endif
   67150 
   67151 #ifdef __LITTLE_ENDIAN__
   67152 #define vget_lane_f16(__p0_242, __p1_242) __extension__ ({ \
   67153   float16x4_t __s0_242 = __p0_242; \
   67154   float16_t __ret_242; \
   67155 float16x4_t __reint_242 = __s0_242; \
   67156 int16_t __reint1_242 = vget_lane_s16(*(int16x4_t *) &__reint_242, __p1_242); \
   67157   __ret_242 = *(float16_t *) &__reint1_242; \
   67158   __ret_242; \
   67159 })
   67160 #else
   67161 #define vget_lane_f16(__p0_243, __p1_243) __extension__ ({ \
   67162   float16x4_t __s0_243 = __p0_243; \
   67163   float16x4_t __rev0_243;  __rev0_243 = __builtin_shufflevector(__s0_243, __s0_243, 3, 2, 1, 0); \
   67164   float16_t __ret_243; \
   67165 float16x4_t __reint_243 = __rev0_243; \
   67166 int16_t __reint1_243 = __noswap_vget_lane_s16(*(int16x4_t *) &__reint_243, __p1_243); \
   67167   __ret_243 = *(float16_t *) &__reint1_243; \
   67168   __ret_243; \
   67169 })
   67170 #endif
   67171 
   67172 #ifdef __LITTLE_ENDIAN__
   67173 #define vgetq_lane_f16(__p0_244, __p1_244) __extension__ ({ \
   67174   float16x8_t __s0_244 = __p0_244; \
   67175   float16_t __ret_244; \
   67176 float16x8_t __reint_244 = __s0_244; \
   67177 int16_t __reint1_244 = vgetq_lane_s16(*(int16x8_t *) &__reint_244, __p1_244); \
   67178   __ret_244 = *(float16_t *) &__reint1_244; \
   67179   __ret_244; \
   67180 })
   67181 #else
   67182 #define vgetq_lane_f16(__p0_245, __p1_245) __extension__ ({ \
   67183   float16x8_t __s0_245 = __p0_245; \
   67184   float16x8_t __rev0_245;  __rev0_245 = __builtin_shufflevector(__s0_245, __s0_245, 7, 6, 5, 4, 3, 2, 1, 0); \
   67185   float16_t __ret_245; \
   67186 float16x8_t __reint_245 = __rev0_245; \
   67187 int16_t __reint1_245 = __noswap_vgetq_lane_s16(*(int16x8_t *) &__reint_245, __p1_245); \
   67188   __ret_245 = *(float16_t *) &__reint1_245; \
   67189   __ret_245; \
   67190 })
   67191 #endif
   67192 
   67193 #ifdef __LITTLE_ENDIAN__
   67194 __ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   67195   uint16x8_t __ret;
   67196   __ret = __p0 + vmull_u8(__p1, __p2);
   67197   return __ret;
   67198 }
   67199 #else
   67200 __ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   67201   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   67202   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   67203   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   67204   uint16x8_t __ret;
   67205   __ret = __rev0 + __noswap_vmull_u8(__rev1, __rev2);
   67206   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   67207   return __ret;
   67208 }
   67209 __ai uint16x8_t __noswap_vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   67210   uint16x8_t __ret;
   67211   __ret = __p0 + __noswap_vmull_u8(__p1, __p2);
   67212   return __ret;
   67213 }
   67214 #endif
   67215 
   67216 #ifdef __LITTLE_ENDIAN__
   67217 __ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   67218   uint64x2_t __ret;
   67219   __ret = __p0 + vmull_u32(__p1, __p2);
   67220   return __ret;
   67221 }
   67222 #else
   67223 __ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   67224   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67225   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67226   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   67227   uint64x2_t __ret;
   67228   __ret = __rev0 + __noswap_vmull_u32(__rev1, __rev2);
   67229   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67230   return __ret;
   67231 }
   67232 __ai uint64x2_t __noswap_vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   67233   uint64x2_t __ret;
   67234   __ret = __p0 + __noswap_vmull_u32(__p1, __p2);
   67235   return __ret;
   67236 }
   67237 #endif
   67238 
   67239 #ifdef __LITTLE_ENDIAN__
   67240 __ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   67241   uint32x4_t __ret;
   67242   __ret = __p0 + vmull_u16(__p1, __p2);
   67243   return __ret;
   67244 }
   67245 #else
   67246 __ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   67247   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67248   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67249   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   67250   uint32x4_t __ret;
   67251   __ret = __rev0 + __noswap_vmull_u16(__rev1, __rev2);
   67252   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67253   return __ret;
   67254 }
   67255 __ai uint32x4_t __noswap_vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   67256   uint32x4_t __ret;
   67257   __ret = __p0 + __noswap_vmull_u16(__p1, __p2);
   67258   return __ret;
   67259 }
   67260 #endif
   67261 
   67262 #ifdef __LITTLE_ENDIAN__
   67263 __ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   67264   int16x8_t __ret;
   67265   __ret = __p0 + vmull_s8(__p1, __p2);
   67266   return __ret;
   67267 }
   67268 #else
   67269 __ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   67270   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   67271   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   67272   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   67273   int16x8_t __ret;
   67274   __ret = __rev0 + __noswap_vmull_s8(__rev1, __rev2);
   67275   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   67276   return __ret;
   67277 }
   67278 __ai int16x8_t __noswap_vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   67279   int16x8_t __ret;
   67280   __ret = __p0 + __noswap_vmull_s8(__p1, __p2);
   67281   return __ret;
   67282 }
   67283 #endif
   67284 
   67285 #ifdef __LITTLE_ENDIAN__
   67286 __ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   67287   int64x2_t __ret;
   67288   __ret = __p0 + vmull_s32(__p1, __p2);
   67289   return __ret;
   67290 }
   67291 #else
   67292 __ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   67293   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67294   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67295   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   67296   int64x2_t __ret;
   67297   __ret = __rev0 + __noswap_vmull_s32(__rev1, __rev2);
   67298   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67299   return __ret;
   67300 }
   67301 __ai int64x2_t __noswap_vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   67302   int64x2_t __ret;
   67303   __ret = __p0 + __noswap_vmull_s32(__p1, __p2);
   67304   return __ret;
   67305 }
   67306 #endif
   67307 
   67308 #ifdef __LITTLE_ENDIAN__
   67309 __ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   67310   int32x4_t __ret;
   67311   __ret = __p0 + vmull_s16(__p1, __p2);
   67312   return __ret;
   67313 }
   67314 #else
   67315 __ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   67316   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67317   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67318   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   67319   int32x4_t __ret;
   67320   __ret = __rev0 + __noswap_vmull_s16(__rev1, __rev2);
   67321   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67322   return __ret;
   67323 }
   67324 __ai int32x4_t __noswap_vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   67325   int32x4_t __ret;
   67326   __ret = __p0 + __noswap_vmull_s16(__p1, __p2);
   67327   return __ret;
   67328 }
   67329 #endif
   67330 
   67331 #ifdef __LITTLE_ENDIAN__
   67332 #define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   67333   uint64x2_t __s0 = __p0; \
   67334   uint32x2_t __s1 = __p1; \
   67335   uint32x2_t __s2 = __p2; \
   67336   uint64x2_t __ret; \
   67337   __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   67338   __ret; \
   67339 })
   67340 #else
   67341 #define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   67342   uint64x2_t __s0 = __p0; \
   67343   uint32x2_t __s1 = __p1; \
   67344   uint32x2_t __s2 = __p2; \
   67345   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   67346   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   67347   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   67348   uint64x2_t __ret; \
   67349   __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   67350   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   67351   __ret; \
   67352 })
   67353 #endif
   67354 
   67355 #ifdef __LITTLE_ENDIAN__
   67356 #define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   67357   uint32x4_t __s0 = __p0; \
   67358   uint16x4_t __s1 = __p1; \
   67359   uint16x4_t __s2 = __p2; \
   67360   uint32x4_t __ret; \
   67361   __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   67362   __ret; \
   67363 })
   67364 #else
   67365 #define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   67366   uint32x4_t __s0 = __p0; \
   67367   uint16x4_t __s1 = __p1; \
   67368   uint16x4_t __s2 = __p2; \
   67369   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   67370   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   67371   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   67372   uint32x4_t __ret; \
   67373   __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   67374   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   67375   __ret; \
   67376 })
   67377 #endif
   67378 
   67379 #ifdef __LITTLE_ENDIAN__
   67380 #define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   67381   int64x2_t __s0 = __p0; \
   67382   int32x2_t __s1 = __p1; \
   67383   int32x2_t __s2 = __p2; \
   67384   int64x2_t __ret; \
   67385   __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   67386   __ret; \
   67387 })
   67388 #else
   67389 #define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   67390   int64x2_t __s0 = __p0; \
   67391   int32x2_t __s1 = __p1; \
   67392   int32x2_t __s2 = __p2; \
   67393   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   67394   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   67395   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   67396   int64x2_t __ret; \
   67397   __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   67398   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   67399   __ret; \
   67400 })
   67401 #endif
   67402 
   67403 #ifdef __LITTLE_ENDIAN__
   67404 #define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   67405   int32x4_t __s0 = __p0; \
   67406   int16x4_t __s1 = __p1; \
   67407   int16x4_t __s2 = __p2; \
   67408   int32x4_t __ret; \
   67409   __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   67410   __ret; \
   67411 })
   67412 #else
   67413 #define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   67414   int32x4_t __s0 = __p0; \
   67415   int16x4_t __s1 = __p1; \
   67416   int16x4_t __s2 = __p2; \
   67417   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   67418   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   67419   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   67420   int32x4_t __ret; \
   67421   __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   67422   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   67423   __ret; \
   67424 })
   67425 #endif
   67426 
   67427 #ifdef __LITTLE_ENDIAN__
   67428 __ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   67429   uint64x2_t __ret;
   67430   __ret = __p0 + vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
   67431   return __ret;
   67432 }
   67433 #else
   67434 __ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   67435   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67436   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67437   uint64x2_t __ret;
   67438   __ret = __rev0 + __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
   67439   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67440   return __ret;
   67441 }
   67442 __ai uint64x2_t __noswap_vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   67443   uint64x2_t __ret;
   67444   __ret = __p0 + __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
   67445   return __ret;
   67446 }
   67447 #endif
   67448 
   67449 #ifdef __LITTLE_ENDIAN__
   67450 __ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   67451   uint32x4_t __ret;
   67452   __ret = __p0 + vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
   67453   return __ret;
   67454 }
   67455 #else
   67456 __ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   67457   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67458   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67459   uint32x4_t __ret;
   67460   __ret = __rev0 + __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
   67461   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67462   return __ret;
   67463 }
   67464 __ai uint32x4_t __noswap_vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   67465   uint32x4_t __ret;
   67466   __ret = __p0 + __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
   67467   return __ret;
   67468 }
   67469 #endif
   67470 
   67471 #ifdef __LITTLE_ENDIAN__
   67472 __ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   67473   int64x2_t __ret;
   67474   __ret = __p0 + vmull_s32(__p1, (int32x2_t) {__p2, __p2});
   67475   return __ret;
   67476 }
   67477 #else
   67478 __ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   67479   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67480   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67481   int64x2_t __ret;
   67482   __ret = __rev0 + __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
   67483   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67484   return __ret;
   67485 }
   67486 __ai int64x2_t __noswap_vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   67487   int64x2_t __ret;
   67488   __ret = __p0 + __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
   67489   return __ret;
   67490 }
   67491 #endif
   67492 
   67493 #ifdef __LITTLE_ENDIAN__
   67494 __ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   67495   int32x4_t __ret;
   67496   __ret = __p0 + vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
   67497   return __ret;
   67498 }
   67499 #else
   67500 __ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   67501   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67502   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67503   int32x4_t __ret;
   67504   __ret = __rev0 + __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
   67505   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67506   return __ret;
   67507 }
   67508 __ai int32x4_t __noswap_vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   67509   int32x4_t __ret;
   67510   __ret = __p0 + __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
   67511   return __ret;
   67512 }
   67513 #endif
   67514 
   67515 #ifdef __LITTLE_ENDIAN__
   67516 __ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   67517   uint16x8_t __ret;
   67518   __ret = __p0 - vmull_u8(__p1, __p2);
   67519   return __ret;
   67520 }
   67521 #else
   67522 __ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   67523   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   67524   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   67525   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   67526   uint16x8_t __ret;
   67527   __ret = __rev0 - __noswap_vmull_u8(__rev1, __rev2);
   67528   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   67529   return __ret;
   67530 }
   67531 __ai uint16x8_t __noswap_vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   67532   uint16x8_t __ret;
   67533   __ret = __p0 - __noswap_vmull_u8(__p1, __p2);
   67534   return __ret;
   67535 }
   67536 #endif
   67537 
   67538 #ifdef __LITTLE_ENDIAN__
   67539 __ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   67540   uint64x2_t __ret;
   67541   __ret = __p0 - vmull_u32(__p1, __p2);
   67542   return __ret;
   67543 }
   67544 #else
   67545 __ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   67546   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67547   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67548   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   67549   uint64x2_t __ret;
   67550   __ret = __rev0 - __noswap_vmull_u32(__rev1, __rev2);
   67551   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67552   return __ret;
   67553 }
   67554 __ai uint64x2_t __noswap_vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   67555   uint64x2_t __ret;
   67556   __ret = __p0 - __noswap_vmull_u32(__p1, __p2);
   67557   return __ret;
   67558 }
   67559 #endif
   67560 
   67561 #ifdef __LITTLE_ENDIAN__
   67562 __ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   67563   uint32x4_t __ret;
   67564   __ret = __p0 - vmull_u16(__p1, __p2);
   67565   return __ret;
   67566 }
   67567 #else
   67568 __ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   67569   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67570   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67571   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   67572   uint32x4_t __ret;
   67573   __ret = __rev0 - __noswap_vmull_u16(__rev1, __rev2);
   67574   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67575   return __ret;
   67576 }
   67577 __ai uint32x4_t __noswap_vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   67578   uint32x4_t __ret;
   67579   __ret = __p0 - __noswap_vmull_u16(__p1, __p2);
   67580   return __ret;
   67581 }
   67582 #endif
   67583 
   67584 #ifdef __LITTLE_ENDIAN__
   67585 __ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   67586   int16x8_t __ret;
   67587   __ret = __p0 - vmull_s8(__p1, __p2);
   67588   return __ret;
   67589 }
   67590 #else
   67591 __ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   67592   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   67593   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   67594   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   67595   int16x8_t __ret;
   67596   __ret = __rev0 - __noswap_vmull_s8(__rev1, __rev2);
   67597   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   67598   return __ret;
   67599 }
   67600 __ai int16x8_t __noswap_vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   67601   int16x8_t __ret;
   67602   __ret = __p0 - __noswap_vmull_s8(__p1, __p2);
   67603   return __ret;
   67604 }
   67605 #endif
   67606 
   67607 #ifdef __LITTLE_ENDIAN__
   67608 __ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   67609   int64x2_t __ret;
   67610   __ret = __p0 - vmull_s32(__p1, __p2);
   67611   return __ret;
   67612 }
   67613 #else
   67614 __ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   67615   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67616   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67617   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   67618   int64x2_t __ret;
   67619   __ret = __rev0 - __noswap_vmull_s32(__rev1, __rev2);
   67620   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67621   return __ret;
   67622 }
   67623 __ai int64x2_t __noswap_vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   67624   int64x2_t __ret;
   67625   __ret = __p0 - __noswap_vmull_s32(__p1, __p2);
   67626   return __ret;
   67627 }
   67628 #endif
   67629 
   67630 #ifdef __LITTLE_ENDIAN__
   67631 __ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   67632   int32x4_t __ret;
   67633   __ret = __p0 - vmull_s16(__p1, __p2);
   67634   return __ret;
   67635 }
   67636 #else
   67637 __ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   67638   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67639   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67640   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   67641   int32x4_t __ret;
   67642   __ret = __rev0 - __noswap_vmull_s16(__rev1, __rev2);
   67643   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67644   return __ret;
   67645 }
   67646 __ai int32x4_t __noswap_vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   67647   int32x4_t __ret;
   67648   __ret = __p0 - __noswap_vmull_s16(__p1, __p2);
   67649   return __ret;
   67650 }
   67651 #endif
   67652 
   67653 #ifdef __LITTLE_ENDIAN__
   67654 #define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   67655   uint64x2_t __s0 = __p0; \
   67656   uint32x2_t __s1 = __p1; \
   67657   uint32x2_t __s2 = __p2; \
   67658   uint64x2_t __ret; \
   67659   __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   67660   __ret; \
   67661 })
   67662 #else
   67663 #define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
   67664   uint64x2_t __s0 = __p0; \
   67665   uint32x2_t __s1 = __p1; \
   67666   uint32x2_t __s2 = __p2; \
   67667   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   67668   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   67669   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   67670   uint64x2_t __ret; \
   67671   __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   67672   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   67673   __ret; \
   67674 })
   67675 #endif
   67676 
   67677 #ifdef __LITTLE_ENDIAN__
   67678 #define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   67679   uint32x4_t __s0 = __p0; \
   67680   uint16x4_t __s1 = __p1; \
   67681   uint16x4_t __s2 = __p2; \
   67682   uint32x4_t __ret; \
   67683   __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   67684   __ret; \
   67685 })
   67686 #else
   67687 #define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
   67688   uint32x4_t __s0 = __p0; \
   67689   uint16x4_t __s1 = __p1; \
   67690   uint16x4_t __s2 = __p2; \
   67691   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   67692   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   67693   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   67694   uint32x4_t __ret; \
   67695   __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   67696   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   67697   __ret; \
   67698 })
   67699 #endif
   67700 
   67701 #ifdef __LITTLE_ENDIAN__
   67702 #define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   67703   int64x2_t __s0 = __p0; \
   67704   int32x2_t __s1 = __p1; \
   67705   int32x2_t __s2 = __p2; \
   67706   int64x2_t __ret; \
   67707   __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
   67708   __ret; \
   67709 })
   67710 #else
   67711 #define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
   67712   int64x2_t __s0 = __p0; \
   67713   int32x2_t __s1 = __p1; \
   67714   int32x2_t __s2 = __p2; \
   67715   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
   67716   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
   67717   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
   67718   int64x2_t __ret; \
   67719   __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
   67720   __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
   67721   __ret; \
   67722 })
   67723 #endif
   67724 
   67725 #ifdef __LITTLE_ENDIAN__
   67726 #define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   67727   int32x4_t __s0 = __p0; \
   67728   int16x4_t __s1 = __p1; \
   67729   int16x4_t __s2 = __p2; \
   67730   int32x4_t __ret; \
   67731   __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
   67732   __ret; \
   67733 })
   67734 #else
   67735 #define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
   67736   int32x4_t __s0 = __p0; \
   67737   int16x4_t __s1 = __p1; \
   67738   int16x4_t __s2 = __p2; \
   67739   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
   67740   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
   67741   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
   67742   int32x4_t __ret; \
   67743   __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
   67744   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
   67745   __ret; \
   67746 })
   67747 #endif
   67748 
   67749 #ifdef __LITTLE_ENDIAN__
   67750 __ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   67751   uint64x2_t __ret;
   67752   __ret = __p0 - vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
   67753   return __ret;
   67754 }
   67755 #else
   67756 __ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   67757   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67758   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67759   uint64x2_t __ret;
   67760   __ret = __rev0 - __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
   67761   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67762   return __ret;
   67763 }
   67764 __ai uint64x2_t __noswap_vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
   67765   uint64x2_t __ret;
   67766   __ret = __p0 - __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
   67767   return __ret;
   67768 }
   67769 #endif
   67770 
   67771 #ifdef __LITTLE_ENDIAN__
   67772 __ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   67773   uint32x4_t __ret;
   67774   __ret = __p0 - vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
   67775   return __ret;
   67776 }
   67777 #else
   67778 __ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   67779   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67780   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67781   uint32x4_t __ret;
   67782   __ret = __rev0 - __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
   67783   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67784   return __ret;
   67785 }
   67786 __ai uint32x4_t __noswap_vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
   67787   uint32x4_t __ret;
   67788   __ret = __p0 - __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
   67789   return __ret;
   67790 }
   67791 #endif
   67792 
   67793 #ifdef __LITTLE_ENDIAN__
   67794 __ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   67795   int64x2_t __ret;
   67796   __ret = __p0 - vmull_s32(__p1, (int32x2_t) {__p2, __p2});
   67797   return __ret;
   67798 }
   67799 #else
   67800 __ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   67801   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   67802   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   67803   int64x2_t __ret;
   67804   __ret = __rev0 - __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
   67805   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   67806   return __ret;
   67807 }
   67808 __ai int64x2_t __noswap_vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
   67809   int64x2_t __ret;
   67810   __ret = __p0 - __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
   67811   return __ret;
   67812 }
   67813 #endif
   67814 
   67815 #ifdef __LITTLE_ENDIAN__
   67816 __ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   67817   int32x4_t __ret;
   67818   __ret = __p0 - vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
   67819   return __ret;
   67820 }
   67821 #else
   67822 __ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   67823   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   67824   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   67825   int32x4_t __ret;
   67826   __ret = __rev0 - __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
   67827   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   67828   return __ret;
   67829 }
   67830 __ai int32x4_t __noswap_vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
   67831   int32x4_t __ret;
   67832   __ret = __p0 - __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
   67833   return __ret;
   67834 }
   67835 #endif
   67836 
   67837 #ifdef __LITTLE_ENDIAN__
   67838 #define vset_lane_f16(__p0_246, __p1_246, __p2_246) __extension__ ({ \
   67839   float16_t __s0_246 = __p0_246; \
   67840   float16x4_t __s1_246 = __p1_246; \
   67841   float16x4_t __ret_246; \
   67842 float16_t __reint_246 = __s0_246; \
   67843 float16x4_t __reint1_246 = __s1_246; \
   67844 int16x4_t __reint2_246 = vset_lane_s16(*(int16_t *) &__reint_246, *(int16x4_t *) &__reint1_246, __p2_246); \
   67845   __ret_246 = *(float16x4_t *) &__reint2_246; \
   67846   __ret_246; \
   67847 })
   67848 #else
   67849 #define vset_lane_f16(__p0_247, __p1_247, __p2_247) __extension__ ({ \
   67850   float16_t __s0_247 = __p0_247; \
   67851   float16x4_t __s1_247 = __p1_247; \
   67852   float16x4_t __rev1_247;  __rev1_247 = __builtin_shufflevector(__s1_247, __s1_247, 3, 2, 1, 0); \
   67853   float16x4_t __ret_247; \
   67854 float16_t __reint_247 = __s0_247; \
   67855 float16x4_t __reint1_247 = __rev1_247; \
   67856 int16x4_t __reint2_247 = __noswap_vset_lane_s16(*(int16_t *) &__reint_247, *(int16x4_t *) &__reint1_247, __p2_247); \
   67857   __ret_247 = *(float16x4_t *) &__reint2_247; \
   67858   __ret_247 = __builtin_shufflevector(__ret_247, __ret_247, 3, 2, 1, 0); \
   67859   __ret_247; \
   67860 })
   67861 #endif
   67862 
   67863 #ifdef __LITTLE_ENDIAN__
   67864 #define vsetq_lane_f16(__p0_248, __p1_248, __p2_248) __extension__ ({ \
   67865   float16_t __s0_248 = __p0_248; \
   67866   float16x8_t __s1_248 = __p1_248; \
   67867   float16x8_t __ret_248; \
   67868 float16_t __reint_248 = __s0_248; \
   67869 float16x8_t __reint1_248 = __s1_248; \
   67870 int16x8_t __reint2_248 = vsetq_lane_s16(*(int16_t *) &__reint_248, *(int16x8_t *) &__reint1_248, __p2_248); \
   67871   __ret_248 = *(float16x8_t *) &__reint2_248; \
   67872   __ret_248; \
   67873 })
   67874 #else
   67875 #define vsetq_lane_f16(__p0_249, __p1_249, __p2_249) __extension__ ({ \
   67876   float16_t __s0_249 = __p0_249; \
   67877   float16x8_t __s1_249 = __p1_249; \
   67878   float16x8_t __rev1_249;  __rev1_249 = __builtin_shufflevector(__s1_249, __s1_249, 7, 6, 5, 4, 3, 2, 1, 0); \
   67879   float16x8_t __ret_249; \
   67880 float16_t __reint_249 = __s0_249; \
   67881 float16x8_t __reint1_249 = __rev1_249; \
   67882 int16x8_t __reint2_249 = __noswap_vsetq_lane_s16(*(int16_t *) &__reint_249, *(int16x8_t *) &__reint1_249, __p2_249); \
   67883   __ret_249 = *(float16x8_t *) &__reint2_249; \
   67884   __ret_249 = __builtin_shufflevector(__ret_249, __ret_249, 7, 6, 5, 4, 3, 2, 1, 0); \
   67885   __ret_249; \
   67886 })
   67887 #endif
   67888 
   67889 #if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
   67890 #ifdef __LITTLE_ENDIAN__
   67891 __ai int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
   67892   int32_t __ret;
   67893   __ret = vqadds_s32(__p0, vqrdmulhs_s32(__p1, __p2));
   67894   return __ret;
   67895 }
   67896 #else
   67897 __ai int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
   67898   int32_t __ret;
   67899   __ret = __noswap_vqadds_s32(__p0, __noswap_vqrdmulhs_s32(__p1, __p2));
   67900   return __ret;
   67901 }
   67902 #endif
   67903 
   67904 #ifdef __LITTLE_ENDIAN__
   67905 __ai int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
   67906   int16_t __ret;
   67907   __ret = vqaddh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
   67908   return __ret;
   67909 }
   67910 #else
   67911 __ai int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
   67912   int16_t __ret;
   67913   __ret = __noswap_vqaddh_s16(__p0, __noswap_vqrdmulhh_s16(__p1, __p2));
   67914   return __ret;
   67915 }
   67916 #endif
   67917 
   67918 #ifdef __LITTLE_ENDIAN__
   67919 #define vqrdmlahs_lane_s32(__p0_250, __p1_250, __p2_250, __p3_250) __extension__ ({ \
   67920   int32_t __s0_250 = __p0_250; \
   67921   int32_t __s1_250 = __p1_250; \
   67922   int32x2_t __s2_250 = __p2_250; \
   67923   int32_t __ret_250; \
   67924   __ret_250 = vqadds_s32(__s0_250, vqrdmulhs_s32(__s1_250, vget_lane_s32(__s2_250, __p3_250))); \
   67925   __ret_250; \
   67926 })
   67927 #else
   67928 #define vqrdmlahs_lane_s32(__p0_251, __p1_251, __p2_251, __p3_251) __extension__ ({ \
   67929   int32_t __s0_251 = __p0_251; \
   67930   int32_t __s1_251 = __p1_251; \
   67931   int32x2_t __s2_251 = __p2_251; \
   67932   int32x2_t __rev2_251;  __rev2_251 = __builtin_shufflevector(__s2_251, __s2_251, 1, 0); \
   67933   int32_t __ret_251; \
   67934   __ret_251 = __noswap_vqadds_s32(__s0_251, __noswap_vqrdmulhs_s32(__s1_251, __noswap_vget_lane_s32(__rev2_251, __p3_251))); \
   67935   __ret_251; \
   67936 })
   67937 #endif
   67938 
   67939 #ifdef __LITTLE_ENDIAN__
   67940 #define vqrdmlahh_lane_s16(__p0_252, __p1_252, __p2_252, __p3_252) __extension__ ({ \
   67941   int16_t __s0_252 = __p0_252; \
   67942   int16_t __s1_252 = __p1_252; \
   67943   int16x4_t __s2_252 = __p2_252; \
   67944   int16_t __ret_252; \
   67945   __ret_252 = vqaddh_s16(__s0_252, vqrdmulhh_s16(__s1_252, vget_lane_s16(__s2_252, __p3_252))); \
   67946   __ret_252; \
   67947 })
   67948 #else
   67949 #define vqrdmlahh_lane_s16(__p0_253, __p1_253, __p2_253, __p3_253) __extension__ ({ \
   67950   int16_t __s0_253 = __p0_253; \
   67951   int16_t __s1_253 = __p1_253; \
   67952   int16x4_t __s2_253 = __p2_253; \
   67953   int16x4_t __rev2_253;  __rev2_253 = __builtin_shufflevector(__s2_253, __s2_253, 3, 2, 1, 0); \
   67954   int16_t __ret_253; \
   67955   __ret_253 = __noswap_vqaddh_s16(__s0_253, __noswap_vqrdmulhh_s16(__s1_253, __noswap_vget_lane_s16(__rev2_253, __p3_253))); \
   67956   __ret_253; \
   67957 })
   67958 #endif
   67959 
   67960 #ifdef __LITTLE_ENDIAN__
   67961 #define vqrdmlahs_laneq_s32(__p0_254, __p1_254, __p2_254, __p3_254) __extension__ ({ \
   67962   int32_t __s0_254 = __p0_254; \
   67963   int32_t __s1_254 = __p1_254; \
   67964   int32x4_t __s2_254 = __p2_254; \
   67965   int32_t __ret_254; \
   67966   __ret_254 = vqadds_s32(__s0_254, vqrdmulhs_s32(__s1_254, vgetq_lane_s32(__s2_254, __p3_254))); \
   67967   __ret_254; \
   67968 })
   67969 #else
   67970 #define vqrdmlahs_laneq_s32(__p0_255, __p1_255, __p2_255, __p3_255) __extension__ ({ \
   67971   int32_t __s0_255 = __p0_255; \
   67972   int32_t __s1_255 = __p1_255; \
   67973   int32x4_t __s2_255 = __p2_255; \
   67974   int32x4_t __rev2_255;  __rev2_255 = __builtin_shufflevector(__s2_255, __s2_255, 3, 2, 1, 0); \
   67975   int32_t __ret_255; \
   67976   __ret_255 = __noswap_vqadds_s32(__s0_255, __noswap_vqrdmulhs_s32(__s1_255, __noswap_vgetq_lane_s32(__rev2_255, __p3_255))); \
   67977   __ret_255; \
   67978 })
   67979 #endif
   67980 
   67981 #ifdef __LITTLE_ENDIAN__
   67982 #define vqrdmlahh_laneq_s16(__p0_256, __p1_256, __p2_256, __p3_256) __extension__ ({ \
   67983   int16_t __s0_256 = __p0_256; \
   67984   int16_t __s1_256 = __p1_256; \
   67985   int16x8_t __s2_256 = __p2_256; \
   67986   int16_t __ret_256; \
   67987   __ret_256 = vqaddh_s16(__s0_256, vqrdmulhh_s16(__s1_256, vgetq_lane_s16(__s2_256, __p3_256))); \
   67988   __ret_256; \
   67989 })
   67990 #else
   67991 #define vqrdmlahh_laneq_s16(__p0_257, __p1_257, __p2_257, __p3_257) __extension__ ({ \
   67992   int16_t __s0_257 = __p0_257; \
   67993   int16_t __s1_257 = __p1_257; \
   67994   int16x8_t __s2_257 = __p2_257; \
   67995   int16x8_t __rev2_257;  __rev2_257 = __builtin_shufflevector(__s2_257, __s2_257, 7, 6, 5, 4, 3, 2, 1, 0); \
   67996   int16_t __ret_257; \
   67997   __ret_257 = __noswap_vqaddh_s16(__s0_257, __noswap_vqrdmulhh_s16(__s1_257, __noswap_vgetq_lane_s16(__rev2_257, __p3_257))); \
   67998   __ret_257; \
   67999 })
   68000 #endif
   68001 
   68002 #ifdef __LITTLE_ENDIAN__
   68003 __ai int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
   68004   int32_t __ret;
   68005   __ret = vqsubs_s32(__p0, vqrdmulhs_s32(__p1, __p2));
   68006   return __ret;
   68007 }
   68008 #else
   68009 __ai int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
   68010   int32_t __ret;
   68011   __ret = __noswap_vqsubs_s32(__p0, __noswap_vqrdmulhs_s32(__p1, __p2));
   68012   return __ret;
   68013 }
   68014 #endif
   68015 
   68016 #ifdef __LITTLE_ENDIAN__
   68017 __ai int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
   68018   int16_t __ret;
   68019   __ret = vqsubh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
   68020   return __ret;
   68021 }
   68022 #else
   68023 __ai int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
   68024   int16_t __ret;
   68025   __ret = __noswap_vqsubh_s16(__p0, __noswap_vqrdmulhh_s16(__p1, __p2));
   68026   return __ret;
   68027 }
   68028 #endif
   68029 
   68030 #ifdef __LITTLE_ENDIAN__
   68031 #define vqrdmlshs_lane_s32(__p0_258, __p1_258, __p2_258, __p3_258) __extension__ ({ \
   68032   int32_t __s0_258 = __p0_258; \
   68033   int32_t __s1_258 = __p1_258; \
   68034   int32x2_t __s2_258 = __p2_258; \
   68035   int32_t __ret_258; \
   68036   __ret_258 = vqsubs_s32(__s0_258, vqrdmulhs_s32(__s1_258, vget_lane_s32(__s2_258, __p3_258))); \
   68037   __ret_258; \
   68038 })
   68039 #else
   68040 #define vqrdmlshs_lane_s32(__p0_259, __p1_259, __p2_259, __p3_259) __extension__ ({ \
   68041   int32_t __s0_259 = __p0_259; \
   68042   int32_t __s1_259 = __p1_259; \
   68043   int32x2_t __s2_259 = __p2_259; \
   68044   int32x2_t __rev2_259;  __rev2_259 = __builtin_shufflevector(__s2_259, __s2_259, 1, 0); \
   68045   int32_t __ret_259; \
   68046   __ret_259 = __noswap_vqsubs_s32(__s0_259, __noswap_vqrdmulhs_s32(__s1_259, __noswap_vget_lane_s32(__rev2_259, __p3_259))); \
   68047   __ret_259; \
   68048 })
   68049 #endif
   68050 
   68051 #ifdef __LITTLE_ENDIAN__
   68052 #define vqrdmlshh_lane_s16(__p0_260, __p1_260, __p2_260, __p3_260) __extension__ ({ \
   68053   int16_t __s0_260 = __p0_260; \
   68054   int16_t __s1_260 = __p1_260; \
   68055   int16x4_t __s2_260 = __p2_260; \
   68056   int16_t __ret_260; \
   68057   __ret_260 = vqsubh_s16(__s0_260, vqrdmulhh_s16(__s1_260, vget_lane_s16(__s2_260, __p3_260))); \
   68058   __ret_260; \
   68059 })
   68060 #else
   68061 #define vqrdmlshh_lane_s16(__p0_261, __p1_261, __p2_261, __p3_261) __extension__ ({ \
   68062   int16_t __s0_261 = __p0_261; \
   68063   int16_t __s1_261 = __p1_261; \
   68064   int16x4_t __s2_261 = __p2_261; \
   68065   int16x4_t __rev2_261;  __rev2_261 = __builtin_shufflevector(__s2_261, __s2_261, 3, 2, 1, 0); \
   68066   int16_t __ret_261; \
   68067   __ret_261 = __noswap_vqsubh_s16(__s0_261, __noswap_vqrdmulhh_s16(__s1_261, __noswap_vget_lane_s16(__rev2_261, __p3_261))); \
   68068   __ret_261; \
   68069 })
   68070 #endif
   68071 
   68072 #ifdef __LITTLE_ENDIAN__
   68073 #define vqrdmlshs_laneq_s32(__p0_262, __p1_262, __p2_262, __p3_262) __extension__ ({ \
   68074   int32_t __s0_262 = __p0_262; \
   68075   int32_t __s1_262 = __p1_262; \
   68076   int32x4_t __s2_262 = __p2_262; \
   68077   int32_t __ret_262; \
   68078   __ret_262 = vqsubs_s32(__s0_262, vqrdmulhs_s32(__s1_262, vgetq_lane_s32(__s2_262, __p3_262))); \
   68079   __ret_262; \
   68080 })
   68081 #else
   68082 #define vqrdmlshs_laneq_s32(__p0_263, __p1_263, __p2_263, __p3_263) __extension__ ({ \
   68083   int32_t __s0_263 = __p0_263; \
   68084   int32_t __s1_263 = __p1_263; \
   68085   int32x4_t __s2_263 = __p2_263; \
   68086   int32x4_t __rev2_263;  __rev2_263 = __builtin_shufflevector(__s2_263, __s2_263, 3, 2, 1, 0); \
   68087   int32_t __ret_263; \
   68088   __ret_263 = __noswap_vqsubs_s32(__s0_263, __noswap_vqrdmulhs_s32(__s1_263, __noswap_vgetq_lane_s32(__rev2_263, __p3_263))); \
   68089   __ret_263; \
   68090 })
   68091 #endif
   68092 
   68093 #ifdef __LITTLE_ENDIAN__
   68094 #define vqrdmlshh_laneq_s16(__p0_264, __p1_264, __p2_264, __p3_264) __extension__ ({ \
   68095   int16_t __s0_264 = __p0_264; \
   68096   int16_t __s1_264 = __p1_264; \
   68097   int16x8_t __s2_264 = __p2_264; \
   68098   int16_t __ret_264; \
   68099   __ret_264 = vqsubh_s16(__s0_264, vqrdmulhh_s16(__s1_264, vgetq_lane_s16(__s2_264, __p3_264))); \
   68100   __ret_264; \
   68101 })
   68102 #else
   68103 #define vqrdmlshh_laneq_s16(__p0_265, __p1_265, __p2_265, __p3_265) __extension__ ({ \
   68104   int16_t __s0_265 = __p0_265; \
   68105   int16_t __s1_265 = __p1_265; \
   68106   int16x8_t __s2_265 = __p2_265; \
   68107   int16x8_t __rev2_265;  __rev2_265 = __builtin_shufflevector(__s2_265, __s2_265, 7, 6, 5, 4, 3, 2, 1, 0); \
   68108   int16_t __ret_265; \
   68109   __ret_265 = __noswap_vqsubh_s16(__s0_265, __noswap_vqrdmulhh_s16(__s1_265, __noswap_vgetq_lane_s16(__rev2_265, __p3_265))); \
   68110   __ret_265; \
   68111 })
   68112 #endif
   68113 
   68114 #endif
   68115 #if defined(__aarch64__)
   68116 #ifdef __LITTLE_ENDIAN__
   68117 __ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
   68118   uint16x8_t __ret;
   68119   __ret = vabdl_u8(vget_high_u8(__p0), vget_high_u8(__p1));
   68120   return __ret;
   68121 }
   68122 #else
   68123 __ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
   68124   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68125   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68126   uint16x8_t __ret;
   68127   __ret = __noswap_vabdl_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
   68128   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68129   return __ret;
   68130 }
   68131 #endif
   68132 
   68133 #ifdef __LITTLE_ENDIAN__
   68134 __ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
   68135   uint64x2_t __ret;
   68136   __ret = vabdl_u32(vget_high_u32(__p0), vget_high_u32(__p1));
   68137   return __ret;
   68138 }
   68139 #else
   68140 __ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
   68141   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68142   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68143   uint64x2_t __ret;
   68144   __ret = __noswap_vabdl_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
   68145   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68146   return __ret;
   68147 }
   68148 #endif
   68149 
   68150 #ifdef __LITTLE_ENDIAN__
   68151 __ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
   68152   uint32x4_t __ret;
   68153   __ret = vabdl_u16(vget_high_u16(__p0), vget_high_u16(__p1));
   68154   return __ret;
   68155 }
   68156 #else
   68157 __ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
   68158   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68159   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68160   uint32x4_t __ret;
   68161   __ret = __noswap_vabdl_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
   68162   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68163   return __ret;
   68164 }
   68165 #endif
   68166 
   68167 #ifdef __LITTLE_ENDIAN__
   68168 __ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
   68169   int16x8_t __ret;
   68170   __ret = vabdl_s8(vget_high_s8(__p0), vget_high_s8(__p1));
   68171   return __ret;
   68172 }
   68173 #else
   68174 __ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
   68175   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68176   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68177   int16x8_t __ret;
   68178   __ret = __noswap_vabdl_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
   68179   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68180   return __ret;
   68181 }
   68182 #endif
   68183 
   68184 #ifdef __LITTLE_ENDIAN__
   68185 __ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
   68186   int64x2_t __ret;
   68187   __ret = vabdl_s32(vget_high_s32(__p0), vget_high_s32(__p1));
   68188   return __ret;
   68189 }
   68190 #else
   68191 __ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
   68192   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68193   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68194   int64x2_t __ret;
   68195   __ret = __noswap_vabdl_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
   68196   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68197   return __ret;
   68198 }
   68199 #endif
   68200 
   68201 #ifdef __LITTLE_ENDIAN__
   68202 __ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
   68203   int32x4_t __ret;
   68204   __ret = vabdl_s16(vget_high_s16(__p0), vget_high_s16(__p1));
   68205   return __ret;
   68206 }
   68207 #else
   68208 __ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
   68209   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68210   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68211   int32x4_t __ret;
   68212   __ret = __noswap_vabdl_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
   68213   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68214   return __ret;
   68215 }
   68216 #endif
   68217 
   68218 #ifdef __LITTLE_ENDIAN__
   68219 __ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
   68220   uint16x8_t __ret;
   68221   __ret = vmovl_high_u8(__p0) + vmovl_high_u8(__p1);
   68222   return __ret;
   68223 }
   68224 #else
   68225 __ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
   68226   uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68227   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68228   uint16x8_t __ret;
   68229   __ret = __noswap_vmovl_high_u8(__rev0) + __noswap_vmovl_high_u8(__rev1);
   68230   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68231   return __ret;
   68232 }
   68233 #endif
   68234 
   68235 #ifdef __LITTLE_ENDIAN__
   68236 __ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
   68237   uint64x2_t __ret;
   68238   __ret = vmovl_high_u32(__p0) + vmovl_high_u32(__p1);
   68239   return __ret;
   68240 }
   68241 #else
   68242 __ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
   68243   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68244   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68245   uint64x2_t __ret;
   68246   __ret = __noswap_vmovl_high_u32(__rev0) + __noswap_vmovl_high_u32(__rev1);
   68247   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68248   return __ret;
   68249 }
   68250 #endif
   68251 
   68252 #ifdef __LITTLE_ENDIAN__
   68253 __ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
   68254   uint32x4_t __ret;
   68255   __ret = vmovl_high_u16(__p0) + vmovl_high_u16(__p1);
   68256   return __ret;
   68257 }
   68258 #else
   68259 __ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
   68260   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68261   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68262   uint32x4_t __ret;
   68263   __ret = __noswap_vmovl_high_u16(__rev0) + __noswap_vmovl_high_u16(__rev1);
   68264   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68265   return __ret;
   68266 }
   68267 #endif
   68268 
   68269 #ifdef __LITTLE_ENDIAN__
   68270 __ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
   68271   int16x8_t __ret;
   68272   __ret = vmovl_high_s8(__p0) + vmovl_high_s8(__p1);
   68273   return __ret;
   68274 }
   68275 #else
   68276 __ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
   68277   int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68278   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68279   int16x8_t __ret;
   68280   __ret = __noswap_vmovl_high_s8(__rev0) + __noswap_vmovl_high_s8(__rev1);
   68281   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68282   return __ret;
   68283 }
   68284 #endif
   68285 
   68286 #ifdef __LITTLE_ENDIAN__
   68287 __ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
   68288   int64x2_t __ret;
   68289   __ret = vmovl_high_s32(__p0) + vmovl_high_s32(__p1);
   68290   return __ret;
   68291 }
   68292 #else
   68293 __ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
   68294   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68295   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68296   int64x2_t __ret;
   68297   __ret = __noswap_vmovl_high_s32(__rev0) + __noswap_vmovl_high_s32(__rev1);
   68298   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68299   return __ret;
   68300 }
   68301 #endif
   68302 
   68303 #ifdef __LITTLE_ENDIAN__
   68304 __ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
   68305   int32x4_t __ret;
   68306   __ret = vmovl_high_s16(__p0) + vmovl_high_s16(__p1);
   68307   return __ret;
   68308 }
   68309 #else
   68310 __ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
   68311   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68312   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68313   int32x4_t __ret;
   68314   __ret = __noswap_vmovl_high_s16(__rev0) + __noswap_vmovl_high_s16(__rev1);
   68315   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68316   return __ret;
   68317 }
   68318 #endif
   68319 
   68320 #ifdef __LITTLE_ENDIAN__
   68321 __ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
   68322   uint16x8_t __ret;
   68323   __ret = __p0 + vmovl_high_u8(__p1);
   68324   return __ret;
   68325 }
   68326 #else
   68327 __ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
   68328   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68329   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68330   uint16x8_t __ret;
   68331   __ret = __rev0 + __noswap_vmovl_high_u8(__rev1);
   68332   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68333   return __ret;
   68334 }
   68335 #endif
   68336 
   68337 #ifdef __LITTLE_ENDIAN__
   68338 __ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
   68339   uint64x2_t __ret;
   68340   __ret = __p0 + vmovl_high_u32(__p1);
   68341   return __ret;
   68342 }
   68343 #else
   68344 __ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
   68345   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68346   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68347   uint64x2_t __ret;
   68348   __ret = __rev0 + __noswap_vmovl_high_u32(__rev1);
   68349   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68350   return __ret;
   68351 }
   68352 #endif
   68353 
   68354 #ifdef __LITTLE_ENDIAN__
   68355 __ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
   68356   uint32x4_t __ret;
   68357   __ret = __p0 + vmovl_high_u16(__p1);
   68358   return __ret;
   68359 }
   68360 #else
   68361 __ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
   68362   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68363   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68364   uint32x4_t __ret;
   68365   __ret = __rev0 + __noswap_vmovl_high_u16(__rev1);
   68366   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68367   return __ret;
   68368 }
   68369 #endif
   68370 
   68371 #ifdef __LITTLE_ENDIAN__
   68372 __ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
   68373   int16x8_t __ret;
   68374   __ret = __p0 + vmovl_high_s8(__p1);
   68375   return __ret;
   68376 }
   68377 #else
   68378 __ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
   68379   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68380   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68381   int16x8_t __ret;
   68382   __ret = __rev0 + __noswap_vmovl_high_s8(__rev1);
   68383   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68384   return __ret;
   68385 }
   68386 #endif
   68387 
   68388 #ifdef __LITTLE_ENDIAN__
   68389 __ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
   68390   int64x2_t __ret;
   68391   __ret = __p0 + vmovl_high_s32(__p1);
   68392   return __ret;
   68393 }
   68394 #else
   68395 __ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
   68396   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68397   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68398   int64x2_t __ret;
   68399   __ret = __rev0 + __noswap_vmovl_high_s32(__rev1);
   68400   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68401   return __ret;
   68402 }
   68403 #endif
   68404 
   68405 #ifdef __LITTLE_ENDIAN__
   68406 __ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
   68407   int32x4_t __ret;
   68408   __ret = __p0 + vmovl_high_s16(__p1);
   68409   return __ret;
   68410 }
   68411 #else
   68412 __ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
   68413   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68414   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68415   int32x4_t __ret;
   68416   __ret = __rev0 + __noswap_vmovl_high_s16(__rev1);
   68417   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68418   return __ret;
   68419 }
   68420 #endif
   68421 
   68422 #ifdef __LITTLE_ENDIAN__
   68423 #define vcopyq_lane_p64(__p0_266, __p1_266, __p2_266, __p3_266) __extension__ ({ \
   68424   poly64x2_t __s0_266 = __p0_266; \
   68425   poly64x1_t __s2_266 = __p2_266; \
   68426   poly64x2_t __ret_266; \
   68427   __ret_266 = vsetq_lane_p64(vget_lane_p64(__s2_266, __p3_266), __s0_266, __p1_266); \
   68428   __ret_266; \
   68429 })
   68430 #else
   68431 #define vcopyq_lane_p64(__p0_267, __p1_267, __p2_267, __p3_267) __extension__ ({ \
   68432   poly64x2_t __s0_267 = __p0_267; \
   68433   poly64x1_t __s2_267 = __p2_267; \
   68434   poly64x2_t __rev0_267;  __rev0_267 = __builtin_shufflevector(__s0_267, __s0_267, 1, 0); \
   68435   poly64x2_t __ret_267; \
   68436   __ret_267 = __noswap_vsetq_lane_p64(__noswap_vget_lane_p64(__s2_267, __p3_267), __rev0_267, __p1_267); \
   68437   __ret_267 = __builtin_shufflevector(__ret_267, __ret_267, 1, 0); \
   68438   __ret_267; \
   68439 })
   68440 #endif
   68441 
   68442 #ifdef __LITTLE_ENDIAN__
   68443 #define vcopyq_lane_f64(__p0_268, __p1_268, __p2_268, __p3_268) __extension__ ({ \
   68444   float64x2_t __s0_268 = __p0_268; \
   68445   float64x1_t __s2_268 = __p2_268; \
   68446   float64x2_t __ret_268; \
   68447   __ret_268 = vsetq_lane_f64(vget_lane_f64(__s2_268, __p3_268), __s0_268, __p1_268); \
   68448   __ret_268; \
   68449 })
   68450 #else
   68451 #define vcopyq_lane_f64(__p0_269, __p1_269, __p2_269, __p3_269) __extension__ ({ \
   68452   float64x2_t __s0_269 = __p0_269; \
   68453   float64x1_t __s2_269 = __p2_269; \
   68454   float64x2_t __rev0_269;  __rev0_269 = __builtin_shufflevector(__s0_269, __s0_269, 1, 0); \
   68455   float64x2_t __ret_269; \
   68456   __ret_269 = __noswap_vsetq_lane_f64(__noswap_vget_lane_f64(__s2_269, __p3_269), __rev0_269, __p1_269); \
   68457   __ret_269 = __builtin_shufflevector(__ret_269, __ret_269, 1, 0); \
   68458   __ret_269; \
   68459 })
   68460 #endif
   68461 
   68462 #ifdef __LITTLE_ENDIAN__
   68463 #define vcopy_lane_p64(__p0_270, __p1_270, __p2_270, __p3_270) __extension__ ({ \
   68464   poly64x1_t __s0_270 = __p0_270; \
   68465   poly64x1_t __s2_270 = __p2_270; \
   68466   poly64x1_t __ret_270; \
   68467   __ret_270 = vset_lane_p64(vget_lane_p64(__s2_270, __p3_270), __s0_270, __p1_270); \
   68468   __ret_270; \
   68469 })
   68470 #else
   68471 #define vcopy_lane_p64(__p0_271, __p1_271, __p2_271, __p3_271) __extension__ ({ \
   68472   poly64x1_t __s0_271 = __p0_271; \
   68473   poly64x1_t __s2_271 = __p2_271; \
   68474   poly64x1_t __ret_271; \
   68475   __ret_271 = __noswap_vset_lane_p64(__noswap_vget_lane_p64(__s2_271, __p3_271), __s0_271, __p1_271); \
   68476   __ret_271; \
   68477 })
   68478 #endif
   68479 
   68480 #ifdef __LITTLE_ENDIAN__
   68481 #define vcopy_lane_f64(__p0_272, __p1_272, __p2_272, __p3_272) __extension__ ({ \
   68482   float64x1_t __s0_272 = __p0_272; \
   68483   float64x1_t __s2_272 = __p2_272; \
   68484   float64x1_t __ret_272; \
   68485   __ret_272 = vset_lane_f64(vget_lane_f64(__s2_272, __p3_272), __s0_272, __p1_272); \
   68486   __ret_272; \
   68487 })
   68488 #else
   68489 #define vcopy_lane_f64(__p0_273, __p1_273, __p2_273, __p3_273) __extension__ ({ \
   68490   float64x1_t __s0_273 = __p0_273; \
   68491   float64x1_t __s2_273 = __p2_273; \
   68492   float64x1_t __ret_273; \
   68493   __ret_273 = __noswap_vset_lane_f64(__noswap_vget_lane_f64(__s2_273, __p3_273), __s0_273, __p1_273); \
   68494   __ret_273; \
   68495 })
   68496 #endif
   68497 
   68498 #ifdef __LITTLE_ENDIAN__
   68499 #define vcopyq_laneq_p64(__p0_274, __p1_274, __p2_274, __p3_274) __extension__ ({ \
   68500   poly64x2_t __s0_274 = __p0_274; \
   68501   poly64x2_t __s2_274 = __p2_274; \
   68502   poly64x2_t __ret_274; \
   68503   __ret_274 = vsetq_lane_p64(vgetq_lane_p64(__s2_274, __p3_274), __s0_274, __p1_274); \
   68504   __ret_274; \
   68505 })
   68506 #else
   68507 #define vcopyq_laneq_p64(__p0_275, __p1_275, __p2_275, __p3_275) __extension__ ({ \
   68508   poly64x2_t __s0_275 = __p0_275; \
   68509   poly64x2_t __s2_275 = __p2_275; \
   68510   poly64x2_t __rev0_275;  __rev0_275 = __builtin_shufflevector(__s0_275, __s0_275, 1, 0); \
   68511   poly64x2_t __rev2_275;  __rev2_275 = __builtin_shufflevector(__s2_275, __s2_275, 1, 0); \
   68512   poly64x2_t __ret_275; \
   68513   __ret_275 = __noswap_vsetq_lane_p64(__noswap_vgetq_lane_p64(__rev2_275, __p3_275), __rev0_275, __p1_275); \
   68514   __ret_275 = __builtin_shufflevector(__ret_275, __ret_275, 1, 0); \
   68515   __ret_275; \
   68516 })
   68517 #endif
   68518 
   68519 #ifdef __LITTLE_ENDIAN__
   68520 #define vcopyq_laneq_f64(__p0_276, __p1_276, __p2_276, __p3_276) __extension__ ({ \
   68521   float64x2_t __s0_276 = __p0_276; \
   68522   float64x2_t __s2_276 = __p2_276; \
   68523   float64x2_t __ret_276; \
   68524   __ret_276 = vsetq_lane_f64(vgetq_lane_f64(__s2_276, __p3_276), __s0_276, __p1_276); \
   68525   __ret_276; \
   68526 })
   68527 #else
   68528 #define vcopyq_laneq_f64(__p0_277, __p1_277, __p2_277, __p3_277) __extension__ ({ \
   68529   float64x2_t __s0_277 = __p0_277; \
   68530   float64x2_t __s2_277 = __p2_277; \
   68531   float64x2_t __rev0_277;  __rev0_277 = __builtin_shufflevector(__s0_277, __s0_277, 1, 0); \
   68532   float64x2_t __rev2_277;  __rev2_277 = __builtin_shufflevector(__s2_277, __s2_277, 1, 0); \
   68533   float64x2_t __ret_277; \
   68534   __ret_277 = __noswap_vsetq_lane_f64(__noswap_vgetq_lane_f64(__rev2_277, __p3_277), __rev0_277, __p1_277); \
   68535   __ret_277 = __builtin_shufflevector(__ret_277, __ret_277, 1, 0); \
   68536   __ret_277; \
   68537 })
   68538 #endif
   68539 
   68540 #ifdef __LITTLE_ENDIAN__
   68541 #define vcopy_laneq_p64(__p0_278, __p1_278, __p2_278, __p3_278) __extension__ ({ \
   68542   poly64x1_t __s0_278 = __p0_278; \
   68543   poly64x2_t __s2_278 = __p2_278; \
   68544   poly64x1_t __ret_278; \
   68545   __ret_278 = vset_lane_p64(vgetq_lane_p64(__s2_278, __p3_278), __s0_278, __p1_278); \
   68546   __ret_278; \
   68547 })
   68548 #else
   68549 #define vcopy_laneq_p64(__p0_279, __p1_279, __p2_279, __p3_279) __extension__ ({ \
   68550   poly64x1_t __s0_279 = __p0_279; \
   68551   poly64x2_t __s2_279 = __p2_279; \
   68552   poly64x2_t __rev2_279;  __rev2_279 = __builtin_shufflevector(__s2_279, __s2_279, 1, 0); \
   68553   poly64x1_t __ret_279; \
   68554   __ret_279 = __noswap_vset_lane_p64(__noswap_vgetq_lane_p64(__rev2_279, __p3_279), __s0_279, __p1_279); \
   68555   __ret_279; \
   68556 })
   68557 #endif
   68558 
   68559 #ifdef __LITTLE_ENDIAN__
   68560 #define vcopy_laneq_f64(__p0_280, __p1_280, __p2_280, __p3_280) __extension__ ({ \
   68561   float64x1_t __s0_280 = __p0_280; \
   68562   float64x2_t __s2_280 = __p2_280; \
   68563   float64x1_t __ret_280; \
   68564   __ret_280 = vset_lane_f64(vgetq_lane_f64(__s2_280, __p3_280), __s0_280, __p1_280); \
   68565   __ret_280; \
   68566 })
   68567 #else
   68568 #define vcopy_laneq_f64(__p0_281, __p1_281, __p2_281, __p3_281) __extension__ ({ \
   68569   float64x1_t __s0_281 = __p0_281; \
   68570   float64x2_t __s2_281 = __p2_281; \
   68571   float64x2_t __rev2_281;  __rev2_281 = __builtin_shufflevector(__s2_281, __s2_281, 1, 0); \
   68572   float64x1_t __ret_281; \
   68573   __ret_281 = __noswap_vset_lane_f64(__noswap_vgetq_lane_f64(__rev2_281, __p3_281), __s0_281, __p1_281); \
   68574   __ret_281; \
   68575 })
   68576 #endif
   68577 
   68578 #ifdef __LITTLE_ENDIAN__
   68579 __ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   68580   uint16x8_t __ret;
   68581   __ret = vmlal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
   68582   return __ret;
   68583 }
   68584 #else
   68585 __ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   68586   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68587   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68588   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68589   uint16x8_t __ret;
   68590   __ret = __noswap_vmlal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
   68591   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68592   return __ret;
   68593 }
   68594 #endif
   68595 
   68596 #ifdef __LITTLE_ENDIAN__
   68597 __ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   68598   uint64x2_t __ret;
   68599   __ret = vmlal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
   68600   return __ret;
   68601 }
   68602 #else
   68603 __ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   68604   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68605   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68606   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   68607   uint64x2_t __ret;
   68608   __ret = __noswap_vmlal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
   68609   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68610   return __ret;
   68611 }
   68612 #endif
   68613 
   68614 #ifdef __LITTLE_ENDIAN__
   68615 __ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   68616   uint32x4_t __ret;
   68617   __ret = vmlal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
   68618   return __ret;
   68619 }
   68620 #else
   68621 __ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   68622   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68623   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68624   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   68625   uint32x4_t __ret;
   68626   __ret = __noswap_vmlal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
   68627   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68628   return __ret;
   68629 }
   68630 #endif
   68631 
   68632 #ifdef __LITTLE_ENDIAN__
   68633 __ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
   68634   int16x8_t __ret;
   68635   __ret = vmlal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
   68636   return __ret;
   68637 }
   68638 #else
   68639 __ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
   68640   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68641   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68642   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68643   int16x8_t __ret;
   68644   __ret = __noswap_vmlal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
   68645   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68646   return __ret;
   68647 }
   68648 #endif
   68649 
   68650 #ifdef __LITTLE_ENDIAN__
   68651 __ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   68652   int64x2_t __ret;
   68653   __ret = vmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
   68654   return __ret;
   68655 }
   68656 #else
   68657 __ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   68658   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68659   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68660   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   68661   int64x2_t __ret;
   68662   __ret = __noswap_vmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
   68663   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68664   return __ret;
   68665 }
   68666 #endif
   68667 
   68668 #ifdef __LITTLE_ENDIAN__
   68669 __ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   68670   int32x4_t __ret;
   68671   __ret = vmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
   68672   return __ret;
   68673 }
   68674 #else
   68675 __ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   68676   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68677   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68678   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   68679   int32x4_t __ret;
   68680   __ret = __noswap_vmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
   68681   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68682   return __ret;
   68683 }
   68684 #endif
   68685 
   68686 #ifdef __LITTLE_ENDIAN__
   68687 __ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
   68688   uint64x2_t __ret;
   68689   __ret = vmlal_n_u32(__p0, vget_high_u32(__p1), __p2);
   68690   return __ret;
   68691 }
   68692 #else
   68693 __ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
   68694   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68695   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68696   uint64x2_t __ret;
   68697   __ret = __noswap_vmlal_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
   68698   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68699   return __ret;
   68700 }
   68701 #endif
   68702 
   68703 #ifdef __LITTLE_ENDIAN__
   68704 __ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
   68705   uint32x4_t __ret;
   68706   __ret = vmlal_n_u16(__p0, vget_high_u16(__p1), __p2);
   68707   return __ret;
   68708 }
   68709 #else
   68710 __ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
   68711   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68712   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68713   uint32x4_t __ret;
   68714   __ret = __noswap_vmlal_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
   68715   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68716   return __ret;
   68717 }
   68718 #endif
   68719 
   68720 #ifdef __LITTLE_ENDIAN__
   68721 __ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
   68722   int64x2_t __ret;
   68723   __ret = vmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
   68724   return __ret;
   68725 }
   68726 #else
   68727 __ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
   68728   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68729   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68730   int64x2_t __ret;
   68731   __ret = __noswap_vmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
   68732   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68733   return __ret;
   68734 }
   68735 #endif
   68736 
   68737 #ifdef __LITTLE_ENDIAN__
   68738 __ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
   68739   int32x4_t __ret;
   68740   __ret = vmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
   68741   return __ret;
   68742 }
   68743 #else
   68744 __ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
   68745   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68746   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68747   int32x4_t __ret;
   68748   __ret = __noswap_vmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
   68749   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68750   return __ret;
   68751 }
   68752 #endif
   68753 
   68754 #ifdef __LITTLE_ENDIAN__
   68755 __ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   68756   uint16x8_t __ret;
   68757   __ret = vmlsl_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
   68758   return __ret;
   68759 }
   68760 #else
   68761 __ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   68762   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68763   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68764   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68765   uint16x8_t __ret;
   68766   __ret = __noswap_vmlsl_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
   68767   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68768   return __ret;
   68769 }
   68770 #endif
   68771 
   68772 #ifdef __LITTLE_ENDIAN__
   68773 __ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   68774   uint64x2_t __ret;
   68775   __ret = vmlsl_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
   68776   return __ret;
   68777 }
   68778 #else
   68779 __ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   68780   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68781   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68782   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   68783   uint64x2_t __ret;
   68784   __ret = __noswap_vmlsl_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
   68785   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68786   return __ret;
   68787 }
   68788 #endif
   68789 
   68790 #ifdef __LITTLE_ENDIAN__
   68791 __ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   68792   uint32x4_t __ret;
   68793   __ret = vmlsl_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
   68794   return __ret;
   68795 }
   68796 #else
   68797 __ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   68798   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68799   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68800   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   68801   uint32x4_t __ret;
   68802   __ret = __noswap_vmlsl_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
   68803   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68804   return __ret;
   68805 }
   68806 #endif
   68807 
   68808 #ifdef __LITTLE_ENDIAN__
   68809 __ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
   68810   int16x8_t __ret;
   68811   __ret = vmlsl_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
   68812   return __ret;
   68813 }
   68814 #else
   68815 __ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
   68816   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68817   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68818   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   68819   int16x8_t __ret;
   68820   __ret = __noswap_vmlsl_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
   68821   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68822   return __ret;
   68823 }
   68824 #endif
   68825 
   68826 #ifdef __LITTLE_ENDIAN__
   68827 __ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   68828   int64x2_t __ret;
   68829   __ret = vmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
   68830   return __ret;
   68831 }
   68832 #else
   68833 __ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   68834   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68835   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68836   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   68837   int64x2_t __ret;
   68838   __ret = __noswap_vmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
   68839   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68840   return __ret;
   68841 }
   68842 #endif
   68843 
   68844 #ifdef __LITTLE_ENDIAN__
   68845 __ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   68846   int32x4_t __ret;
   68847   __ret = vmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
   68848   return __ret;
   68849 }
   68850 #else
   68851 __ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   68852   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68853   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68854   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   68855   int32x4_t __ret;
   68856   __ret = __noswap_vmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
   68857   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68858   return __ret;
   68859 }
   68860 #endif
   68861 
   68862 #ifdef __LITTLE_ENDIAN__
   68863 __ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
   68864   uint64x2_t __ret;
   68865   __ret = vmlsl_n_u32(__p0, vget_high_u32(__p1), __p2);
   68866   return __ret;
   68867 }
   68868 #else
   68869 __ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
   68870   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68871   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68872   uint64x2_t __ret;
   68873   __ret = __noswap_vmlsl_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
   68874   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68875   return __ret;
   68876 }
   68877 #endif
   68878 
   68879 #ifdef __LITTLE_ENDIAN__
   68880 __ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
   68881   uint32x4_t __ret;
   68882   __ret = vmlsl_n_u16(__p0, vget_high_u16(__p1), __p2);
   68883   return __ret;
   68884 }
   68885 #else
   68886 __ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
   68887   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68888   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68889   uint32x4_t __ret;
   68890   __ret = __noswap_vmlsl_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
   68891   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68892   return __ret;
   68893 }
   68894 #endif
   68895 
   68896 #ifdef __LITTLE_ENDIAN__
   68897 __ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
   68898   int64x2_t __ret;
   68899   __ret = vmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
   68900   return __ret;
   68901 }
   68902 #else
   68903 __ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
   68904   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   68905   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   68906   int64x2_t __ret;
   68907   __ret = __noswap_vmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
   68908   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   68909   return __ret;
   68910 }
   68911 #endif
   68912 
   68913 #ifdef __LITTLE_ENDIAN__
   68914 __ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
   68915   int32x4_t __ret;
   68916   __ret = vmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
   68917   return __ret;
   68918 }
   68919 #else
   68920 __ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
   68921   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   68922   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68923   int32x4_t __ret;
   68924   __ret = __noswap_vmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
   68925   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   68926   return __ret;
   68927 }
   68928 #endif
   68929 
   68930 #ifdef __LITTLE_ENDIAN__
   68931 #define vmulx_lane_f64(__p0_282, __p1_282, __p2_282) __extension__ ({ \
   68932   float64x1_t __s0_282 = __p0_282; \
   68933   float64x1_t __s1_282 = __p1_282; \
   68934   float64x1_t __ret_282; \
   68935   float64_t __x_282 = vget_lane_f64(__s0_282, 0); \
   68936   float64_t __y_282 = vget_lane_f64(__s1_282, __p2_282); \
   68937   float64_t __z_282 = vmulxd_f64(__x_282, __y_282); \
   68938   __ret_282 = vset_lane_f64(__z_282, __s0_282, __p2_282); \
   68939   __ret_282; \
   68940 })
   68941 #else
   68942 #define vmulx_lane_f64(__p0_283, __p1_283, __p2_283) __extension__ ({ \
   68943   float64x1_t __s0_283 = __p0_283; \
   68944   float64x1_t __s1_283 = __p1_283; \
   68945   float64x1_t __ret_283; \
   68946   float64_t __x_283 = __noswap_vget_lane_f64(__s0_283, 0); \
   68947   float64_t __y_283 = __noswap_vget_lane_f64(__s1_283, __p2_283); \
   68948   float64_t __z_283 = __noswap_vmulxd_f64(__x_283, __y_283); \
   68949   __ret_283 = __noswap_vset_lane_f64(__z_283, __s0_283, __p2_283); \
   68950   __ret_283; \
   68951 })
   68952 #endif
   68953 
   68954 #ifdef __LITTLE_ENDIAN__
   68955 #define vmulx_laneq_f64(__p0_284, __p1_284, __p2_284) __extension__ ({ \
   68956   float64x1_t __s0_284 = __p0_284; \
   68957   float64x2_t __s1_284 = __p1_284; \
   68958   float64x1_t __ret_284; \
   68959   float64_t __x_284 = vget_lane_f64(__s0_284, 0); \
   68960   float64_t __y_284 = vgetq_lane_f64(__s1_284, __p2_284); \
   68961   float64_t __z_284 = vmulxd_f64(__x_284, __y_284); \
   68962   __ret_284 = vset_lane_f64(__z_284, __s0_284, 0); \
   68963   __ret_284; \
   68964 })
   68965 #else
   68966 #define vmulx_laneq_f64(__p0_285, __p1_285, __p2_285) __extension__ ({ \
   68967   float64x1_t __s0_285 = __p0_285; \
   68968   float64x2_t __s1_285 = __p1_285; \
   68969   float64x2_t __rev1_285;  __rev1_285 = __builtin_shufflevector(__s1_285, __s1_285, 1, 0); \
   68970   float64x1_t __ret_285; \
   68971   float64_t __x_285 = __noswap_vget_lane_f64(__s0_285, 0); \
   68972   float64_t __y_285 = __noswap_vgetq_lane_f64(__rev1_285, __p2_285); \
   68973   float64_t __z_285 = __noswap_vmulxd_f64(__x_285, __y_285); \
   68974   __ret_285 = __noswap_vset_lane_f64(__z_285, __s0_285, 0); \
   68975   __ret_285; \
   68976 })
   68977 #endif
   68978 
   68979 #endif
   68980 #ifdef __LITTLE_ENDIAN__
   68981 __ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   68982   uint16x8_t __ret;
   68983   __ret = __p0 + vabdl_u8(__p1, __p2);
   68984   return __ret;
   68985 }
   68986 #else
   68987 __ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   68988   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   68989   uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   68990   uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   68991   uint16x8_t __ret;
   68992   __ret = __rev0 + __noswap_vabdl_u8(__rev1, __rev2);
   68993   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   68994   return __ret;
   68995 }
   68996 __ai uint16x8_t __noswap_vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
   68997   uint16x8_t __ret;
   68998   __ret = __p0 + __noswap_vabdl_u8(__p1, __p2);
   68999   return __ret;
   69000 }
   69001 #endif
   69002 
   69003 #ifdef __LITTLE_ENDIAN__
   69004 __ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   69005   uint64x2_t __ret;
   69006   __ret = __p0 + vabdl_u32(__p1, __p2);
   69007   return __ret;
   69008 }
   69009 #else
   69010 __ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   69011   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   69012   uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   69013   uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   69014   uint64x2_t __ret;
   69015   __ret = __rev0 + __noswap_vabdl_u32(__rev1, __rev2);
   69016   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   69017   return __ret;
   69018 }
   69019 __ai uint64x2_t __noswap_vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
   69020   uint64x2_t __ret;
   69021   __ret = __p0 + __noswap_vabdl_u32(__p1, __p2);
   69022   return __ret;
   69023 }
   69024 #endif
   69025 
   69026 #ifdef __LITTLE_ENDIAN__
   69027 __ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   69028   uint32x4_t __ret;
   69029   __ret = __p0 + vabdl_u16(__p1, __p2);
   69030   return __ret;
   69031 }
   69032 #else
   69033 __ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   69034   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   69035   uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   69036   uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   69037   uint32x4_t __ret;
   69038   __ret = __rev0 + __noswap_vabdl_u16(__rev1, __rev2);
   69039   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   69040   return __ret;
   69041 }
   69042 __ai uint32x4_t __noswap_vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
   69043   uint32x4_t __ret;
   69044   __ret = __p0 + __noswap_vabdl_u16(__p1, __p2);
   69045   return __ret;
   69046 }
   69047 #endif
   69048 
   69049 #ifdef __LITTLE_ENDIAN__
   69050 __ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   69051   int16x8_t __ret;
   69052   __ret = __p0 + vabdl_s8(__p1, __p2);
   69053   return __ret;
   69054 }
   69055 #else
   69056 __ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   69057   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   69058   int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   69059   int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   69060   int16x8_t __ret;
   69061   __ret = __rev0 + __noswap_vabdl_s8(__rev1, __rev2);
   69062   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   69063   return __ret;
   69064 }
   69065 __ai int16x8_t __noswap_vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
   69066   int16x8_t __ret;
   69067   __ret = __p0 + __noswap_vabdl_s8(__p1, __p2);
   69068   return __ret;
   69069 }
   69070 #endif
   69071 
   69072 #ifdef __LITTLE_ENDIAN__
   69073 __ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   69074   int64x2_t __ret;
   69075   __ret = __p0 + vabdl_s32(__p1, __p2);
   69076   return __ret;
   69077 }
   69078 #else
   69079 __ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   69080   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   69081   int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
   69082   int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
   69083   int64x2_t __ret;
   69084   __ret = __rev0 + __noswap_vabdl_s32(__rev1, __rev2);
   69085   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   69086   return __ret;
   69087 }
   69088 __ai int64x2_t __noswap_vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
   69089   int64x2_t __ret;
   69090   __ret = __p0 + __noswap_vabdl_s32(__p1, __p2);
   69091   return __ret;
   69092 }
   69093 #endif
   69094 
   69095 #ifdef __LITTLE_ENDIAN__
   69096 __ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   69097   int32x4_t __ret;
   69098   __ret = __p0 + vabdl_s16(__p1, __p2);
   69099   return __ret;
   69100 }
   69101 #else
   69102 __ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   69103   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   69104   int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   69105   int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   69106   int32x4_t __ret;
   69107   __ret = __rev0 + __noswap_vabdl_s16(__rev1, __rev2);
   69108   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   69109   return __ret;
   69110 }
   69111 __ai int32x4_t __noswap_vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
   69112   int32x4_t __ret;
   69113   __ret = __p0 + __noswap_vabdl_s16(__p1, __p2);
   69114   return __ret;
   69115 }
   69116 #endif
   69117 
   69118 #if defined(__aarch64__)
   69119 #ifdef __LITTLE_ENDIAN__
   69120 __ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   69121   uint16x8_t __ret;
   69122   __ret = vabal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
   69123   return __ret;
   69124 }
   69125 #else
   69126 __ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
   69127   uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   69128   uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   69129   uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   69130   uint16x8_t __ret;
   69131   __ret = __noswap_vabal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
   69132   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   69133   return __ret;
   69134 }
   69135 #endif
   69136 
   69137 #ifdef __LITTLE_ENDIAN__
   69138 __ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   69139   uint64x2_t __ret;
   69140   __ret = vabal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
   69141   return __ret;
   69142 }
   69143 #else
   69144 __ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
   69145   uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   69146   uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   69147   uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   69148   uint64x2_t __ret;
   69149   __ret = __noswap_vabal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
   69150   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   69151   return __ret;
   69152 }
   69153 #endif
   69154 
   69155 #ifdef __LITTLE_ENDIAN__
   69156 __ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   69157   uint32x4_t __ret;
   69158   __ret = vabal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
   69159   return __ret;
   69160 }
   69161 #else
   69162 __ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
   69163   uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   69164   uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   69165   uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   69166   uint32x4_t __ret;
   69167   __ret = __noswap_vabal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
   69168   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   69169   return __ret;
   69170 }
   69171 #endif
   69172 
   69173 #ifdef __LITTLE_ENDIAN__
   69174 __ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
   69175   int16x8_t __ret;
   69176   __ret = vabal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
   69177   return __ret;
   69178 }
   69179 #else
   69180 __ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
   69181   int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
   69182   int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   69183   int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
   69184   int16x8_t __ret;
   69185   __ret = __noswap_vabal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
   69186   __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
   69187   return __ret;
   69188 }
   69189 #endif
   69190 
   69191 #ifdef __LITTLE_ENDIAN__
   69192 __ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   69193   int64x2_t __ret;
   69194   __ret = vabal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
   69195   return __ret;
   69196 }
   69197 #else
   69198 __ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
   69199   int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
   69200   int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
   69201   int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
   69202   int64x2_t __ret;
   69203   __ret = __noswap_vabal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
   69204   __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
   69205   return __ret;
   69206 }
   69207 #endif
   69208 
   69209 #ifdef __LITTLE_ENDIAN__
   69210 __ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   69211   int32x4_t __ret;
   69212   __ret = vabal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
   69213   return __ret;
   69214 }
   69215 #else
   69216 __ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
   69217   int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
   69218   int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
   69219   int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
   69220   int32x4_t __ret;
   69221   __ret = __noswap_vabal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
   69222   __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
   69223   return __ret;
   69224 }
   69225 #endif
   69226 
   69227 #endif
   69228 
   69229 #undef __ai
   69230 
   69231 #endif /* __ARM_NEON_H */
   69232